Open savchenko opened 3 weeks ago
Hi @savchenko. We really want to reproduce the issue. Could you tell us the context length and NGL you have set in the model settings? (Top-right corner of the screen, Model tab). It would be great if you could upload the log file here.
So far as I know, Ollama is limited to a 2048 context length, requiring less RAM/VRAM to run. You can configure the same parameters on Jan to fit your device's capability, I guess it's currently 4096 by default.
Jan version
0.5.7
Describe the Bug
codestral:22b-v0.1-q3_K_M
works perfectly fine viaollama
/ Hollama, but returns "Failed to start" via Jan.aiSteps to Reproduce
Screenshots / Logs
What is your OS?