Closed Qmsla1 closed 4 months ago
Please feel free to put a PR with this added lr. You can add it directly from pt scheduler similar to other schedulers as shown here: https://github.com/NVIDIA/NeMo/blob/6ca22351ea7dd231dfb94284d141de938cdb2822/nemo/core/optim/lr_scheduler.py#L963
This issue is stale because it has been open for 30 days with no activity. Remove stale label or comment or this will be closed in 7 days.
This issue was closed because it has been inactive for 7 days since being marked as stale.
The NVIDIA NeMo Framework, a powerful generative AI framework, currently lacks support for the MultiStepLR learning rate scheduler, which is available in PyTorch’s torch.optim.lr_scheduler module. This scheduler is commonly used for adjusting learning rates at specific milestones during training.
Expected Behavior: NeMo should include an implementation or provide guidance on how to use MultiStepLR effectively within its training pipelines.
Steps to Reproduce:
Attempt to use MultiStepLR with NeMo conf yaml. Observe that it is not directly supported. Additional Context: Support for MultiStepLR would enhance the flexibility and adaptability of NeMo’s training process, especially for researchers and developers working on large language models, multimodal models, automatic speech recognition, and text-to-speech synthesis.
how to fix: add MultiStepLR to https://github.com/NVIDIA/NeMo/blob/main/nemo/core/config/schedulers.py