alexsemendinger / mirror-descent

Learn mirror descent potential with good inductive bias.
0 stars 0 forks source link

Mirror Descent Experiments

Background

Stochastic Gradient / Mirror Descent: Minimax Optimality and Implicit Regularization

Data-Driven Mirror Descent in Input-Convex Neural Networks

Question

Above tells us that mirror descent on data ${x_i, yi}$ with potential $\phi$ converges to $\operatorname*{argmin}{w \in W^} D_{\phi} (w | w_0)$ (where $W^$ is the set of weights that interpolate the data and $w_0$ is the intialization).

Can we learn a useful data-dependent $\phi$ that improves generalization?

Roadmap