simonw / llm-llama-cpp

LLM plugin for running models using llama.cpp
Apache License 2.0
139 stars 20 forks source link

Error running with `-m gguf` option #32

Closed nickagliano-st closed 9 months ago

nickagliano-st commented 9 months ago

Hello!

I am trying to run the following command—

llm -m gguf \
  -o path mixtral-8x7b-instruct-v0.1.Q6_K.gguf \
  '[INST] Write a Python function that downloads a file from a URL[/INST]'

—referenced from the tutorial here.

I'm getting this output: Error: 'gguf' is not a known model, which I assume is coming from cli.py.

Some more details:

Hopefully I'm not missing something obvious. Let me know if I can be of help! I clicked around a bit hoping it might be an easy fix but couldn't find anything.

– NA

garthk commented 9 months ago

27 workaround so -m gguf -o file "$GGUF_MODEL_FILE" will work:

llm llama-cpp models
nickagliano-st commented 9 months ago

Oh, I see. Easy enough. Thanks @garthk!