drormoran / Equivariant-SFM

MIT License
87 stars 15 forks source link

Which hyper-parameters to use for reproducing the result? #7

Closed lucasbrynte closed 1 year ago

lucasbrynte commented 1 year ago

Greetings, and many thanks for the contribution as well as for sharing the implementation!

I'm trying to make sense of which hyper-parameters have been used for the different settings, and hope you may be able to help me out despite it being a while ago that you ran the experiments. I totally understand if you cannot recall all the details. Any pointers would be greatly appreciated.

In the paper, there is a subsection of Section 4 which mentions experimenting with different hyper-parameters:

  1. Learning rates in {1e−2, 1e−3, 1e−4}
  2. Network width for the encoder E and heads in {128, 256, 512}
  3. Number of layers in these networks {2, 3}
  4. Depth threshold h in {1e-2, 1e-3, 1e-4}
  5. Std normalization for the layer output in Eq. (1)

So here comes some questions :grin:

https://github.com/drormoran/Equivariant-SFM/blob/26658a7452e8de1458a8d9d969a334f965060126/code/confs/Learning_Euc.conf#L22 https://github.com/drormoran/Equivariant-SFM/blob/26658a7452e8de1458a8d9d969a334f965060126/code/confs/Learning_Proj.conf#L19

Best, Lucas

drormoran commented 1 year ago

Hi Lucas,

You should be able to reproduce the paper results using the .conf file in the repository. We will fix the mismatch regarding the number of channels in the paper.

Yes, the "Std normalization for the layer output in Eq. (1)" refers to normalizing the features, as you mention in your question.

I don't remember why the "Learning_Euc" and "Learning_Proj" conf files differ so much. Since a Hyper-parameters search in the learning scheme requires a lot of time and resources, it is possible that we used a "skinnier" HPS in the projective case.