Anyone know the reason for this? Are these updated values that should be used?
Most notably the batch_size would affect the number of training iterations, where the current 200k would only perform half the training with 128 samples per batch, vs 256 samples per batch.
Config.yaml vs Paper
Anyone know the reason for this? Are these updated values that should be used?
Most notably the batch_size would affect the number of training iterations, where the current 200k would only perform half the training with 128 samples per batch, vs 256 samples per batch.