google-deepmind / optax

Optax is a gradient processing and optimization library for JAX.
https://optax.readthedocs.io
Apache License 2.0
1.71k stars 194 forks source link

Feature request for the AdeMAMix optimizer. #1058

Open mathDR opened 2 months ago

mathDR commented 2 months ago

The AdeMAMix optimizer is a simple modification of the Adam optimizer with a mixture of two EMAs to better take advantage of past gradients.

The paper has optax skeleton code which I could contribute if the maintainers deem this a good fit for the repo.

clementpoiret commented 1 month ago

Hey, any update on it since the PR has been closed? :) thanks !

mathDR commented 1 month ago

Apologies. I closed this because It was just easier to "begin again". I have a draft PR locally that I was going to push this week.

clementpoiret commented 1 month ago

Awesome, good to know! Thx for your work 👌

mathDR commented 1 month ago

Okay PR here