PyTorch Lightning 1.6 is the work of 99 contributors who have worked on features, bug-fixes, and documentation for a total of over 750 commits since 1.5. This is our most active release yet. Here are some highlights:
Introducing Intel's Habana Accelerator
Lightning 1.6 now supports the Habana® framework, which includes Gaudi® AI training processors. Their heterogeneous architecture includes a cluster of fully programmable Tensor Processing Cores (TPC) along with its associated development tools and libraries and a configurable Matrix Math engine.
You can leverage the Habana hardware to accelerate your Deep Learning training workloads simply by passing:
Broadcast the _terminate_gracefully to all processes and add support for DDP (#10638)
Added support for re-instantiation of custom (subclasses of) DataLoaders returned in the *_dataloader() methods, i.e., automatic replacement of samplers now works with custom types of DataLoader (#10680)
Added a function to validate if fault tolerant training is supported. (#10465)
Added a private callback to manage the creation and deletion of fault-tolerance checkpoints (#11862)
Show a better error message when a custom DataLoader implementation is not well implemented and we need to reconstruct it (#10719)
Show a better error message when frozen dataclass is used as a batch (#10927)
Save the Loop's state by default in the checkpoint (#10784)
Added Loop.replace to easily switch one loop for another (#10324)
Added support for --lr_scheduler=ReduceLROnPlateau to the LightningCLI (#10860)
Added LightningCLI.configure_optimizers to override the configure_optimizers return value (#10860)
Added LightningCLI(auto_registry) flag to register all subclasses of the registerable components automatically (#12108)
Added a warning that shows when max_epochs in the Trainer is not set (#10700)
Added support for returning a single Callback from LightningModule.configure_callbacks without wrapping it into a list (#11060)
Added console_kwargs for RichProgressBar to initialize inner Console (#10875)
Added support for shorthand notation to instantiate loggers with the LightningCLI (#11533)
Dependabot will resolve any conflicts with this PR as long as you don't alter it yourself. You can also trigger a rebase manually by commenting @dependabot rebase.
Dependabot commands and options
You can trigger Dependabot actions by commenting on this PR:
- `@dependabot rebase` will rebase this PR
- `@dependabot recreate` will recreate this PR, overwriting any edits that have been made to it
- `@dependabot merge` will merge this PR after your CI passes on it
- `@dependabot squash and merge` will squash and merge this PR after your CI passes on it
- `@dependabot cancel merge` will cancel a previously requested merge and block automerging
- `@dependabot reopen` will reopen this PR if it is closed
- `@dependabot close` will close this PR and stop Dependabot recreating it. You can achieve the same result by closing it manually
- `@dependabot ignore this major version` will close this PR and stop Dependabot creating any more for this major version (unless you reopen the PR or upgrade to it yourself)
- `@dependabot ignore this minor version` will close this PR and stop Dependabot creating any more for this minor version (unless you reopen the PR or upgrade to it yourself)
- `@dependabot ignore this dependency` will close this PR and stop Dependabot creating any more for this dependency (unless you reopen the PR or upgrade to it yourself)
Bumps pytorch-lightning from 1.5.9 to 1.6.1.
Release notes
Sourced from pytorch-lightning's releases.
... (truncated)
Changelog
Sourced from pytorch-lightning's changelog.
... (truncated)
Commits
0b0f1ec
Update CHANGELOG.md2187f30
Support auto_select_gpus with accelerator and devices api (#12608)d95447a
Update deepspeed precision test (#12727)83e0c4a
RaiseMisconfigurationException
when the accelerator is available but… (#12...ba1e869
CI: check docker requires (#12677)dfd9730
fix import failer (#12676)6aca974
Run main progress bar independent of val progress bar inTQDMProgressBar
(#...23d3d46
Fix flaky test, that is not consistent on some configurations (#12707)8bc4a76
Update LightningCLI tests to reflect changes in jsonargparse 4.6.0 (#12704)a4026fe
Fix a typo in warning message inside Trainer.reset_train_dataloader (#12645)Dependabot will resolve any conflicts with this PR as long as you don't alter it yourself. You can also trigger a rebase manually by commenting
@dependabot rebase
.Dependabot commands and options
You can trigger Dependabot actions by commenting on this PR: - `@dependabot rebase` will rebase this PR - `@dependabot recreate` will recreate this PR, overwriting any edits that have been made to it - `@dependabot merge` will merge this PR after your CI passes on it - `@dependabot squash and merge` will squash and merge this PR after your CI passes on it - `@dependabot cancel merge` will cancel a previously requested merge and block automerging - `@dependabot reopen` will reopen this PR if it is closed - `@dependabot close` will close this PR and stop Dependabot recreating it. You can achieve the same result by closing it manually - `@dependabot ignore this major version` will close this PR and stop Dependabot creating any more for this major version (unless you reopen the PR or upgrade to it yourself) - `@dependabot ignore this minor version` will close this PR and stop Dependabot creating any more for this minor version (unless you reopen the PR or upgrade to it yourself) - `@dependabot ignore this dependency` will close this PR and stop Dependabot creating any more for this dependency (unless you reopen the PR or upgrade to it yourself)