Skip to the content.

iterative algorithm

elements

A useful Algorithm Design Framework

suppose the task is min L, then we design a framework as:

\[\theta_{k+1} = min\{q_{k}+\frac{1}{2u_k}\|\theta-\theta_k\|_2^2\}\]

comment
cannot directly minimize the first approximation term, which is accurate only around $\theta_k$, thus we need the proximal term, which can pinish a large step

almost universal algorithmic design strategy solving the original problem by solving a sequence of simpler subproblems

Choice of search direction

Gradient-based Optimization Algorithm

most learning problems do not have closed form solution, but they are convex problems, which provide great property when using gradient-based methods

ex. Logistic regression, SVM, …,

namely, the point where ${nbala}=0$ is the global optimal point

GD

more info will be shown in optimization part!

Gradient Descent with momentum

convergence

if L is convex and have Lipschitz smooth parameter L, then gradient descent with constant learning rate 1/L satisfies:

\[L(\theta_k)-L(\hat{\theta}) \leq \frac{L\|(\theta_0-\hat{\theta})\|_2^2}{2k}\]

supplementary
两种收敛速度的判别标准:Q-收敛速度 和 R-收敛速度