Closed keeper-jie closed 1 year ago
Try this : #14 (comment)
Thank you very much for your help, I modified the following code to finish running the program.
model = PeftModel.from_pretrained( model, "tloen/alpaca-lora-7b", device_map={'':0})
Try this : #14 (comment)
I find a more appropriate approach to deploy the model.
tokenizer = LlamaTokenizer.from_pretrained("decapoda-research/llama-7b-hf", device_map={'':0})
model = LlamaForCausalLM.from_pretrained(
"decapoda-research/llama-7b-hf",
load_in_8bit=True,
torch_dtype=torch.float16,
device_map={'':0}
)
# add device_map={'':0} in PeftModel.from_pretrained to confirm 2 2080Ti can work
model = PeftModel.from_pretrained(
model, "tloen/alpaca-lora-7b", torch_dtype=torch.float16, device_map={'':0}
)
When I download the colab code and run it in my GPU server, which is different with git clone the repository to run. I modified the code and tested by my 2 2080Ti GPU server and pulled my code.
generate used to work for me but i rebuilt it today and now i'm getting the same error... Note that I only have a single GPU, so this doesn't have anything to do with having multiple.
generate used to work for me but i rebuilt it today and now i'm getting the same error... Note that I only have a single GPU, so this doesn't have anything to do with having multiple.
I have the same problem, help!!!
I got the error as blow and hope someone can solve it. I have change the device_map(such as "balanced", "balanced_low_0", "sequential") in
but not working.
Error
Enviroment:
3.pip list