psugihara / FreeChat

llama.cpp based AI chat app for macOS
https://www.freechat.run
MIT License
438 stars 41 forks source link

Crash on sending prompt #66

Closed sussyboiiii closed 7 months ago

sussyboiiii commented 7 months ago

Im in the Beta, macOS Sonoma, M1 Pro 16GB, loading any model crashed even small ones which worked beforehand.

psugihara commented 7 months ago

Hm, i wonder if the models you're using aren't compatible with the latest llama.cpp anymore (llama.cpp periodically breaks compatibility). That's the only substantive change in the latest beta.

psugihara commented 7 months ago

If you have time, can you download a recently quantized model and let me know if it works?

sussyboiiii commented 7 months ago

Hello, I used the newest llama.cpp version and the same model

Screenshot 2024-04-25 at 19 47 33
sussyboiiii commented 7 months ago

Tested two models both crashed in FreeChat and worked in the terminal.

psugihara commented 7 months ago

dang, no good! thanks for testing.

i'm doing a refactor at the moment to simplify a few things and I'll try to figure out what's causing it.

sussyboiiii commented 7 months ago

After updating FreeChat it now doesn't crash anymore.

sussyboiiii commented 7 months ago

Where did the ability to change the system prompt go? Can't find it anymore.

psugihara commented 7 months ago

that's great news! thanks for testing again.

Where did the ability to change the system prompt go? Can't find it anymore.

I still see it in the FreeChat > Settings panel:

Screenshot 2024-04-26 at 8 19 27 AM Screenshot 2024-04-26 at 8 19 24 AM
sussyboiiii commented 7 months ago

Oh well, I ment the tab to choose the template or whatever it's called where I can choose between Vicuna or e.g. ChatML.

psugihara commented 7 months ago

Yes, I have removed that and now defer to llama.cpp's conversation templating. The prompt formats are now embedded in GGUFs so that inference servers (like the local llama.cpp server we use) can do it correctly themselves. This makes freechat a lot easier to maintain because I don't have to hardcode template formats (and model -> template mappings) myself as new models are released.

sussyboiiii commented 7 months ago

Ah ok nice! Then this seems to be working as intended. Thank you for your work, have a nice day!