s-nlp / kbqa

0 stars 0 forks source link

Try Amos optimizator for seq2seq model. Compare with AdamW #72

Closed MihailSalnikov closed 1 year ago

MihailSalnikov commented 2 years ago

Try for t5-large Evaluate results and speed on WDSQ and MINTAK

Amos: An Adam-style Optimizer with Adaptive Weight Decay towards Model-Oriented Scale

Amos is a new optimizer that we propose to pre-train large language models. It is more efficient and converges faster than AdamW: ≤ 51% memory for slot variables, and better valid loss within ≤ 70% training time!Amos is a new optimizer that we propose to pre-train large language models. It is more efficient and converges faster than AdamW: ≤ 51% memory for slot variables, and better valid loss within ≤ 70% training time!

ArXiV: https://arxiv.org/abs/2210.11693