chenxin061 / pdarts

Codes for our paper "Progressive Differentiable Architecture Search:Bridging the Depth Gap between Search and Evaluation"
Other
359 stars 83 forks source link

Question about Dropout Rate in Paper #29

Closed volgachen closed 4 years ago

volgachen commented 4 years ago

Excuse me, In paper you have explained the reason of using a gradually-decayed dropout rate. However, it seems that in your experiments the dropout rate is increasing during searching(0.0, 0.4, 0.7). I wonder if I neglected some important details.

chenxin061 commented 4 years ago

The (0.0, 0.4, 0.7) are the initial dropout rates for stage 1, 2 and 3, respectively. For a specific stage, for example, stage 2, the dropout rate is gradually decayed from 0.4 to around 0. Hope this helps.

volgachen commented 4 years ago

@chenxin061 Thank you! I find the corresponding code now. Is there any explanation for an increasing initial dropout rates?

chenxin061 commented 4 years ago

@chenxin061 Thank you! I find the corresponding code now. Is there any explanation for an increasing initial dropout rates?

Very sorry for the late reply.

The reason for the increasing dropout rates lies in the property that a deeper search network is easier to fall into the trap of skip-connects than a shallow one so stronger regularization is needed.

volgachen commented 4 years ago

@chenxin061 Thank you! I find the corresponding code now. Is there any explanation for an increasing initial dropout rates?

Very sorry for the late reply.

The reason for the increasing dropout rates lies in the property that a deeper search network is easier to fall into the trap of skip-connects than a shallow one so stronger regularization is needed.

Got it! Thank you.