warner-benjamin / optimi

Fast, Modern, Memory Efficient, and Low Precision PyTorch Optimizers
https://optimi.benjaminwarner.dev
MIT License
58 stars 2 forks source link

[FeatureRequest]gradient-descent-the-ultimate-optimizer #6

Open sdbds opened 4 months ago

sdbds commented 4 months ago

code:https://github.com/kach/gradient-descent-the-ultimate-optimizer paper:https://arxiv.org/abs/1909.13371

Because this method requires modifying the optimizer so it's not as popular as those common versions, but this repository refactors a lot of optimizers so maybe it can be implemented.