Open sgdesmet opened 3 days ago
Related issue #3760
Jan works well with small models, but larger models are unpredictable. With 16Gb VRAM and 32Gb RAM, sometimes a 34B model loads, sometimes it doesn't. Sometimes my whole system freezes and I have to restart, sometimes my GPU driver silently resets. Sometimes, everything works normally. Koboldcpp, Msty, LM Studio, etc. load the same model fine every time.
Jan version
0.5.5
Describe the Bug
I'm running Fedora 40 on a laptop with a GTX 1050 Ti with 4Gb RAM. When enabling GPU acceleration and attempting to run models that are marked as 'Slow on your device' (such as
Llama 3.2 3B Instruct Q8
) , they fail to start without any visible error message. At first glance, the logs show what appears to be a memory issue:Is it correct that my device is unable to run this particular model? If so, a 'Not enough VRAM' indicator when downloading the model and an explicit error message when starting the model would be my expectation.
Steps to Reproduce
Screenshots / Logs
app.log
What is your OS?