Closed jmduarte closed 3 years ago
Thanks @jmduarte
qdepthwiseconv2d_batchnorm should have the same problem, right? could you also do that in this PR, thanks.
Do you have an example that has the divergence problem? consider to add it into unittest?
@zhuangh thanks for checking this out!
I similarly fixed qdepthwiseconv2d_batchnorm
in this PR. Unfortunately, I couldn't reproduce the original "diverging" loss behavior with the master branch, so it may just have been a bad configuration of my model originally.
In any case, I think propagating the batchnorm parameters as done in this PR still makes sense.
LGTM
This PR does two things
name
unspecified so it can be automatically assigned in case the user forgets to specify (name=None
actually causes an error)