ggerganov / llama.cpp

LLM inference in C/C++
MIT License
65.59k stars 9.41k forks source link

sh: 1: ./llama.cpp/llama-quantize: not found #8107

Closed RakshitAralimatti closed 3 months ago

RakshitAralimatti commented 3 months ago

What happened?

For converting the FP16.gguf to q5_k_m for llama3 i was getting the error sh: 1: ./llama.cpp/llama-quantize: not found previously i used os.system("./llama.cpp/llama-quantize " + gguf_dir + "/" + gguf_F16_name + " " + model_path + " " + m) in my script to convert even that is not working now its giving the same not found issue

Name and Version

Lastest pervsion

What operating system are you seeing the problem on?

Linux

Relevant log output

bash: ./llama.cpp/quantize: No such file or directory
ggerganov commented 3 months ago

The tool has been renamed from quantize to llama-quantize

RakshitAralimatti commented 3 months ago

Thanks for your response. Now its working

elliotthwang commented 2 months ago

@ggerganov llama-quantize: failed to access

showed the message as follows; /bin/bash: line 1: ./llama.cpp/llama-quantize: No such file or directory

ggerganov commented 2 months ago

Run make first

elliotthwang commented 2 months ago

@ggerganov Run make as follows:

Install llama.cpp

!git clone https://github.com/ggerganov/llama.cpp !cd llama.cpp && git pull && make clean && LLAMA_CUBLAS=1 make !pip install -r llama.cpp/requirements.txt

Is it coming correctly?