Open lipan6461188 opened 1 year ago
I implemented a model.Transformer.cuda function to automatically assign the parameters of each layer to detected CUDA devices. This can help to load the 65B model to ≥ 2 40G A100 GPUs with the following command:
model.Transformer.cuda
CUDA_VISIBLE_DEVICES=0,1 python example.py --ckpt_dir /path/to/model/65B --tokenizer_path /path/to/model/tokenizer.model --max_batch_size=1
I implemented a
model.Transformer.cuda
function to automatically assign the parameters of each layer to detected CUDA devices. This can help to load the 65B model to ≥ 2 40G A100 GPUs with the following command: