Closed WardLT closed 3 years ago
@WardLT Something to consider going forward. https://github.com/uncertainty-toolbox/uncertainty-toolbox
We could simply wrap our predictions, std, labels around their API and let the tool box figure out what would be best calibration method.
Part of the puzzle. Without bootstrap sampling when updating models and only 4 replicas in the ensemble, our uncertainties are much worse after retraining.
The de-calibration is lessened if we use bootstrap sampling when creating the training set before updating the model.
We see similar, slight degradation with the 16 bootstraped models
Training with more epochs (here, 512) can make the problem worse
Resetting the weights on the optimizer does seem to help. This is back to using 64 epochs to retrain the model.
Using random initial weights seems to work just as well in terms of the uncertainties.
It was a bug 😆 See: 73f0579
We find the performance of our active learning agent gets worse as we retrain the models. The chart below shows how we find fewer high-performing molecules with strategies where we update the model (
update
andtrain
) than we do with a strategy where we never update the MPNNs (no-retrain
)A list of hypotheses:
Potential solutions: