Currently when learning an MRsort model, a population of intermediate models is trained in parallel. After each training iteration, the best half of this population is kept, and the worst half is reinitialized to randomized states. Admittedly, these random states are created using a clever non-uniform distribution, but this seems like a waste of information.
Is it possible to trace the life of a model (with a unique label corresponding to the current run)?
If so, it would be interesting to show whether there is little or no improvement from the random models.
A quick modification might be to simply duplicate the surviving half on the worse half. If the theory is right this should quickly extinguish the diversity and "reduce" the rate of progress (compared to an optimal solution which does not yet exist).
A more complicated modification would be to use the improvement part of the profiles as a mutation point. This is where we have to make meaningful choices based on a random generator (and this seems contradictory to me). Taking advantage of these bifurcation points as diversification tools seems interesting to me (we make several draws at these points). But It comes back to the previous idea (limiting the randomness).
It will thus be necessary to add randomness but perhaps between these two levels of subtleties (full random vs some points of improvement)
Currently when learning an MRsort model, a population of intermediate models is trained in parallel. After each training iteration, the best half of this population is kept, and the worst half is reinitialized to randomized states. Admittedly, these random states are created using a clever non-uniform distribution, but this seems like a waste of information.
We could: