Open 555cider opened 7 months ago
[quantized] example (modified) with the Quantized Solar model (local): Failed. It only outputs nonsense that is unrelated to the question.
If you are getting incoherent output, you might have the wrong tokenizer set. I have used the solar models with the quantized implementation of llama in candle with these settings:
model id: TheBloke/SOLAR-10.7B-v1.0-GGUF
revision: main
gguf file within the repo: solar-10.7b-v1.0.Q4_K_M.gguf
tokenizer repo: upstage/SOLAR-10.7B-v1.0
tokenizer file: tokenizer.json
With the wrong tokenizer, you can get some giberish output. Here is the output I got when trying to use the solar with the llama tokenizer:
капи кате cheap versree слоrio UnjetQL voiceseg listGE Here Jas SozialExt prod arr press віціled solemAnchor fields ár_+PLLouis searchedQu profileslickiedasterund v damalsisko timing rings authorizationтельной pochodontql tableView'];equationroom
I am trying to run the Solar model, but I am constantly failing. Here are my attempts:
Conclusion: Is there any way to run the Quantized Solar model? I know I only wrote about 5 attempts, but I actually tried several different variations of the code in step 1. I also downloaded the model several times in my poor internet speed.