🚀 A simple way to launch, train, and use PyTorch models on almost any device and distributed configuration, automatic mixed precision (including fp8), and easy-to-configure FSDP and DeepSpeed support
litepod 8-core single host TPU VM running Ubuntu 22.04
Accelerate 0.31.0
Information
[ ] The official example scripts
[ ] My own modified scripts
Tasks
[ ] One of the scripts in the examples/ folder of Accelerate or an officially supported no_trainer script in the examples folder of the transformers repo (such as run_no_trainer_glue.py)
[ ] My own task or dataset (give details below)
Reproduction
I would like to use the torchxla_trace_once backend as discussed here. However, when I attempt to run the official training scripts for causal language modeling using the following
System Info
Information
Tasks
no_trainer
script in theexamples
folder of thetransformers
repo (such asrun_no_trainer_glue.py
)Reproduction
I would like to use the torchxla_trace_once backend as discussed here. However, when I attempt to run the official training scripts for causal language modeling using the following
it does not list
aot_torchxla_trace_once
as a supported backendI would not mind contributing such a feature if this is something that is desired in accelerate or is on the roadmap.
Expected behavior
Support for the
aot_torchxla_trace_once
XLA backend.