RuntimeError: Expected all tensors to be on the same device, but found at least two devices, cuda:0 and cuda:1!
If I set n_devide=2 in HookedTransformer.from_pretrained(model_name=LLAMA_2_7B_CHAT_PATH, device="cuda",n_devices=2, fold_ln=False, center_writing_weights=False, center_unembed=False) I will get the above errors.
It currently does not support multi-gpu and I will add this feature recently. Will let you know when I tackle it.
You can try to move the tensor to the same device temporarily.
Thanks!
RuntimeError: Expected all tensors to be on the same device, but found at least two devices, cuda:0 and cuda:1!
If I set n_devide=2 in
HookedTransformer.from_pretrained(model_name=LLAMA_2_7B_CHAT_PATH, device="cuda",n_devices=2, fold_ln=False, center_writing_weights=False, center_unembed=False)
I will get the above errors.