Open deltaboukensha opened 1 year ago
I have solved it with a cpu installation by installing this : https://github.com/krychu/llama
instead of https://github.com/facebookresearch/llama
Complete process to install :
https://github.com/facebookresearch/llama
and extract it to a llama-main
folderhttps://github.com/krychu/llama
and extract it and replace files in the llama-main
folderdownload.sh
script in a terminal, passing the URL provided when prompted to start the downloadllama-main
folderpython3 -m venv env
and activate it : source env/bin/activate
python3 -m pip install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/cpu #pour la version cpu
python3 -m pip install -e .
torchrun --nproc_per_node 1 example_text_completion.py \
--ckpt_dir llama-2-7b/ \
--tokenizer_path tokenizer.model \
--max_seq_len 128 --max_batch_size 1 #(instead of 4)
Hi. I'm trying to figure out how to troubleshoot this generic error message i get from running the example locally in my machine.
I suspect either the PyTorch or Cuda version is wrong. Or my hardware is insufficient.
How do I determine what the issue is exactly?
Im running the project from docker with GPU and virtualization enabled Docker Images I've tried: docker pull pytorch/pytorch:2.0.1-cuda11.7-cudnn8-devel docker pull pytorch/pytorch:1.13.1-cuda11.6-cudnn8-devel
64GB RAM OS Windows 11 NVIDIA GeForce RTX 3070 GPU mem 8 GB / 32 GB