Closed daidaiershidi closed 3 weeks ago
Please ensure that this model can be loaded with transformers
successfully.
Please ensure that this model can be loaded with
transformers
successfully.
使用 transformers 加载没有问题。 在opencompass上,我换成只用6张卡加载也没有问题
“--hf-num-gpus 8” indicates the utilization of tensor parallelism for a single model. For a 13B model, “--hf-num-gpus 2” might suffice. In case one desires to employ data parallelism, the “--max-num-worker” parameter can be utilized.
The error File "/home/asfhalshd/miniconda3/envs/opencompass/lib/python3.10/site-packages/transformers/models/llama/modeling_llama.py", line 1190, in <listcomp> logits = [F.linear(hidden_states, lm_head_slices[i]) for i in range(self.config.pretraining_tp)] RuntimeError: Expected all tensors to be on the same device, but found at least two devices, cuda:6 and cuda:7! (when checking argument for argument mat2 in method wrapper_CUDA_mm)
appears to be unrelated to OpenCompass but rather pertains to Transformers.
Feel free to re-open if needed.
先决条件
问题类型
我正在使用官方支持的任务/模型/数据集进行评估。
环境
重现问题 - 代码/配置示例
重现问题 - 命令或脚本
重现问题 - 错误信息
问题:RuntimeError: Expected all tensors to be on the same device, but found at least two devices, cuda:6 and cuda:7! (when checking argument for argument mat2 in method wrapper_CUDA_mm)
报错: