Closed janekl closed 1 year ago
MLCommons CLA bot All contributors have signed the MLCommons CLA ✍️ ✅
@samiwilf if you give this an ok I will approve.
LGTM. @erichan1 if you can stamp please
@janekl feel free to land
@janekl feel free to land
Hi @erichan1 thanks. It says You’re not authorized to merge this pull request. So I need to ask either you or @johntran-nv to merge it please.
Author: Jan Lasek, Nvidia (jlasek_at_nvidia.com)
There is parameter mismatch for Adagrad optimizer for embeddings and dense layers in the new DLRMv2 recommender benchmark.
As the dense layers and the embeddings employ PyTorch and FBGEMM Adagrad implementations, respectively, one needs to explicitly pass all relevant optimizer parameters in
apply_optimizer_in_backward
call:eps=1e-8
as default hereeps=1e-10
, see docs here.This causes a mismatch between the two optimizer used that I'm fixing here.