Closed kiranchari closed 11 months ago
Hi - thanks for your interest. It seems you are using only one hparams_seed
, and that's why all the model selection methods give the same results.
In our implementation, we used 16 different hparams_seed
to select the best one. You might want to follow the setting (e.g., see here).
Thanks for your reply @YyzHarry. I think it was just a co-incidence that all model selection methods produced the same results above. Based on your answer here https://github.com/YyzHarry/SubpopBench/issues/4, the best hyperparameter is chosen first. The different model selection strategies are used only after training with the best hparams_seed
. Is my understanding correct?
Could you please share the best hparams_seed
for ERM and other methods? It would be time consuming to run 16 hparams_seeds for all datasets.
Unfortunately we didn't record the best hparams_seed
for every algorithm.
I installed the conda environment using the provided environment.yml. I am using RedHat OS.
When I run ERM on CelebA without training attributes (CelebA_ERM_attrNo) using hparams_seed=0 and seeds={0,1,2} I get the following results, which are quite different from the paper.
I am also not able to reproduce Waterbirds_ERM_attrNo or the other datasets. I did not modify the code in subpopbench.
Appreciate any help in reproducing the results!