to train data2vec model, distributed_training.distributed_world_size is set to 16, but, the log and the system shows there are only one gpu used in training.
log------------------
[2022-02-18 08:45:58,583][fairseq_cli.train][INFO] - training on 1 devices (GPUs/TPUs)
log------------------
I use this code
code------------------ python fairseq_cli/hydra_train.py -m --config-dir examples/data2vec/config/audio/pretraining \ --config-name base_librispeech task.data=/path/to/manifests common.user_dir=examples/data2vec code------------------
to train data2vec model, distributed_training.distributed_world_size is set to 16, but, the log and the system shows there are only one gpu used in training.
log------------------ [2022-02-18 08:45:58,583][fairseq_cli.train][INFO] - training on 1 devices (GPUs/TPUs) log------------------