Open sido420 opened 9 months ago
Error I am getting after a few minutes:
$ KV_CACHE_IN_GPU=1 python inference.py --ckpt_dir ../codellama/CodeLlama-7b/ --tokenizer_path ../codellama/CodeLlama-7b/tokenizer.model
Prompt:['I believe in ']
Killed
(llama2) xxx@localhost:~/pyllama$ Prompt:['I believe in ']
Prompt:[I believe in ]: command not found
I am new to AI and trying to use
llama2
model locally usingpyllama
.I tried different options, but nothing seems to work. I downloaded llama using https://github.com/facebookresearch/llama.
Here is what I tried (see below for installed packages):
Below seems to work, but I don't get any response whatsoever:
I tried both pytorch cuda and non-cuda packages from https://pytorch.org/get-started/locally/. Example:
conda install pytorch torchvision torchaudio pytorch-cuda=11.8 -c pytorch -c nvidia
but same NCCL error in torchrun and no output frominference.py
I am on an HP workstation running ubuntu (23.04 (Lunar Lobster))