The main change is to replace standard dataloaders (finishing at each epoch) to "infinite" ones - uniformly sampling from the dataset till the program terminates.
This technically might change the order of samples, but:
Everyone is using fairly large batch size (i.e. 56, which is 1/3 of the dataset)
I checked the distribution of samples and throughout the training it is indistinguishable from the previous behavior
The new RCPs are very similar to the old ones - I'll open a PR soon. Indeed some RCPs are a bit faster than previously, not sure if it's related to the new behavior, a bug in the old behavior when we switched to samples, or just due to the variance
The main change is to replace standard dataloaders (finishing at each epoch) to "infinite" ones - uniformly sampling from the dataset till the program terminates. This technically might change the order of samples, but:
There is no need to modify the submission code.