Discussion about this post

User's avatar
Johan Ugander's avatar

I like the idea of first and foremost motivating convex optimization via the "it's where local search finds global optima" argument. Adding to the standard tools here, I have on several occasions found use for the not-widely-known Mountain Pass Theorem: under (commonly satisfied) smoothness conditions, if a function has no local _maxima_ then all the extrema are _minima_. As such, for the purposes of optimization, any local minima is the single global minima.

With some whittling, this can probably be turned into engaging PhD-level lecture material. As an example, see (Fleischer, 1964) [1], which gives an analysis showing that Lloyd's algorithm for k-means finds global optima when used to quantize Gaussian noise. Fleisher states and uses a mountain pass-like lemma and says "A proof of this Lemma will be given in a forthcoming paper" Which never appeared! It's likely somebody clued him in to the fact that his stated lemma was a corollary of a known thing.

[1] https://web.stanford.edu/~jugander/rare/Fleischer1964_SufficientConditionsMinimumDistortion.pdf

Expand full comment
Charlotte LeMay's avatar

This works for the cost function f(x,y)=y, but aren't there types of cost functions which would have local but not global extrema even on convex sets? Something like f(x,y)=cos(x)cos(y)*e^(x^2/10) on the circle of radius 5 for example. Is the circle of radius 5 not considered convex with respect to that cost function? Or are there requirements on cost functions in order for convexity to have the desired property?

Expand full comment
8 more comments...

No posts