Closed netandreus closed 7 months ago
Hi! I'm a bot running with LocalAI ( a crazy experiment of @mudler ) - please beware that I might hallucinate sometimes!
_but.... I can also be funny or helpful :smilecat: and I can provide generally speaking good tips or places where to look after in the documentation or in the code based on what you wrote in the issue.
Don't engage in conversation with me, I don't support (yet) replying!
ERROR: The prompt size exceeds the context window size and cannot be processed.
Sources:
@mudler do you have any ideas, what is wrong?
@netandreus did you tried with falcon/gguf files? GGML are quite outdated and old now. That should be working with the default llama-cpp backend as for now. Also, which version of LocalAI are you trying this with?
Is there any way to run Falcon 40B model with LocalAI? I'm trying this models:
and this backends:
falcon
falcon-ggml
without any success.
My model config.yml file:
Model = wizardlm-uncensored-falcon-40b.ggccv1.q4_0.bin, backend = falcon
Model = wizardlm-uncensored-falcon-40b.ggccv1.q4_0.bin, backend = falcon-ggml
Model = wizardlm-uncensored-falcon-40b.ggccv1.q4_0.bin, backend line commented
Model = falcon-40b-instruct.ggccv1.q4_0.bin, backend = falcon
Model = falcon-40b-instruct.ggccv1.q4_0.bin, backend = falcon-ggml
Model = falcon-40b-instruct.ggccv1.q4_0.bin, without backend in config
LocalAI version:
Environment, CPU architecture, OS, and Version:
Describe the bug
To Reproduce curl:
Expected behavior Some response.
Logs Full logs:
falcon_40b_local_ai_full_logs.txt