Describe the bug
If you attempt to replace all BatchNormalization layers with AccumBatchNormalization layers it works fine (mechanically).
However, if you attempt to set the weights of the new layer using the old layer, the shapes do not match.
There are also no unit tests that actually verifies that the performance is the same after replacement. This should be added - both in terms of inference and training.
Expected behavior
The AccumBatchNormalization layer should work as a drop-in replacement for the existing BatchNormalization layer.
Describe the bug If you attempt to replace all
BatchNormalization
layers withAccumBatchNormalization
layers it works fine (mechanically).However, if you attempt to set the weights of the new layer using the old layer, the shapes do not match.
There are also no unit tests that actually verifies that the performance is the same after replacement. This should be added - both in terms of inference and training.
Expected behavior The
AccumBatchNormalization
layer should work as a drop-in replacement for the existingBatchNormalization
layer.