Skip to article frontmatterSkip to article content
Site not loading correctly?

This may be due to an incorrect BASE_URL configuration. See the MyST Documentation for reference.

Things on this page are fragmentary and immature notes/thoughts of the author. Please read with your own judgement!

If the objective function has derivative, then an optimization problem is equivalent to finding the root of the derivative of the object function.

Finding Root

Optimization

This is actually an algorithm to find the root of equations. To use this algorithm to do optimization, we can solve for the root of the partial derivates of the objective function. Since the algorithm itself requires the function to be solved to have partial derivatives, the objective function must be twice differtiable. Newton’s algorithm usually requires less steps to converge than the gradient descent algorithm, however, it need to computer the Hessian matrix, which requires more time and sometimes causes numerical problems.

Optimization is really important in science and industry. Many problems need to optimize an objective function. As we discussed before, human beings only knows how to optimize some simple case of objective functions. To make optimization easy, we sometimes use functions that are easier to optimize to approximate the original objective function. Usually a convex function used to approximate the original objective function.

For exponential family, we usually have convext optimization probablems and the gradient descent or Newton’s algorithm is good.

batch gradient descent

coordinate descent (fixed part of the parameter, optimize the other part)