Open prathameshza opened 7 months ago
@prathameshza is there a curl request to this same URL that you can get to succeed? And you are running llama.cpp's non-OpenAI ./server example, right?
@prathameshza is there a curl request to this same URL that you can get to succeed? And you are running llama.cpp's non-OpenAI ./server example, right?
I tried to ping the address which gives error. Also I missed to configure the model, so i followed this steps:
But I am getting this error:
llm_load_print_meta: EOS token = 2 '</s>'
llm_load_print_meta: UNK token = 0 '<unk>'
llm_load_print_meta: LF token = 13 '<0x0A>'
llm_load_tensors: ggml ctx size = 0.00 MiB
llama_model_load: error loading model: create_tensor: tensor 'token_embd.weight' not found
llama_load_model_from_file: failed to load model
llama_init_from_gpt_params: error: failed to load model 'models/ggml-vocab-llama.gguf'
{"timestamp":1708505822,"level":"ERROR","function":"load_model","line":380,"message":"unable to load model","model":"models/ggml-vocab-llama.gguf"}
terminate called without an active exception
Aborted (core dumped)
After this command:
./server -c 4096 --host 0.0.0.0 -t 16 --mlock -m models/ggml-vocab-llama.gguf
Before submitting your bug report
Relevant environment info
Description
Getting FetchError 2 in vscode when trying to run WizardCoder-7b (llama.cpp) Bellow is the screenshot:
To reproduce
Add model to config.js
I have generated the certificates like this:
my folder dir looks like this
Log output