Open sxjscience opened 6 years ago
working on FTML
AdaBound The authors call it as fast as Adam and as good as SGD. So it might a good one.
@szha I would like to attempt an implementation of the RAdam optimizer. I have written some code here that probably isn't yet be ready for a PR (might not even be correct), so how would you like me to share it with you for a review so that you can point me in the right direction? Thanks.
@Hunter-Zolomon thanks for offering to contribute.
To implement optimizer in MXNet, the best reference would be the existing optimizers. You can find many in python/mxnet/optimizer
folder.
Once you have the implementation ready, you can refer to the contribution guides and submit a pull request. Make sure to also include tests for the new optimizer following examples here. Feel free to ping me or others for review on the pull request.
Optimizer plays a fundamental role in machine learning. This issue tracks the optimizers that people have requested for support in MXNet. We can also comment below to recommend new optimizers. It's inspired by the thread in https://discuss.gluon.ai/t/topic/3714.
@szha @szhengac