I am trying to figure out if it makes any sense for the training loop.
As far as I can see, freeze_all, that sets the trainable property, is only called on convolutional layers, so at least during training the new BatchNormalization has the same behavior as the original BatchNormalization from Keras, because its trainable is actually always True.
I am trying to figure out if it makes any sense for the training loop.
As far as I can see,
freeze_all
, that sets thetrainable
property, is only called on convolutional layers, so at least during training the newBatchNormalization
has the same behavior as the originalBatchNormalization
from Keras, because itstrainable
is actually alwaysTrue
.Can somebody clear my confusion here?