jpata / particleflow

Machine-learned, GPU-accelerated particle flow reconstruction
Apache License 2.0
24 stars 30 forks source link

Learning rate schedules and Mamba layer #282

Closed erwulff closed 11 months ago

erwulff commented 11 months ago

Implement LR schedules in the PyTorch training code. Add MambaLayer and make it configurable through parameter config files and for HPO.

Also contains the following changes.

Screenshot 2023-12-06 at 13 28 12 Learning rate versus training step of a training using the cosinedecay LR schedule and was interrupted halfway through and then continued.

Screenshot 2023-12-06 at 13 27 41 Learning rate versus training step of a training using the onecycle LR schedule and was interrupted halfway through and then continued.