undreamai / LLMUnity

Create characters in Unity with LLMs!
https://undream.ai
MIT License
713 stars 74 forks source link

Need to add CFG Scale and CFG and context free grammars #30

Open ArEnSc opened 10 months ago

amakropoulos commented 10 months ago

Thanks for the issue! I think you mean adding the following options from llama.cpp, is it correct? --grammar / --grammar-file --cfg-negative-prompt / --cfg-negative-prompt-file --cfg-scale

If yes, sure I can do that. I'm trying to think of a good way to add different options from llama.cpp without blowing up the UI. Maybe I would separate some options that are probably less frequently in a usual chat setting. and provide a field that the used can add arguments to the llama.cpp server command. What do you think?

ArEnSc commented 10 months ago

yeah I wonder if this is supported onthe llamafile server, I checked I don't see those last 3 options

amakropoulos commented 10 months ago

I'm upgrading to llamafile 0.6, that contains those :)

ArEnSc commented 10 months ago

I'm upgrading to llamafile 0.6, that contains those :)

amazing! =D

amakropoulos commented 10 months ago

I have released v1.0.5 that allows grammars along with other parameters. Unfortunately I can't set the negative-prompt and the scale because they are not exposed as arguments from the llama.cpp server :/

ArEnSc commented 10 months ago

I have released v1.0.5 that allows grammars along with other parameters. Unfortunately I can't set the negative-prompt and the scale because they are not exposed as arguments from the llama.cpp server :/

yeah it was confusing since I saw args code that implied that it was availible from llamafile, I think ill or someone will need to go in there and make the changes to the server and understand how the CFG scale works once my project is off the ground, amazing work though btw! =) from my understanding CFG scale requires that we run inference twice and the server did drift away from the original PR made and it was never merged

amakropoulos commented 9 months ago

yes exactly, I saw the llamafile args and confused it 🙈

amakropoulos commented 4 months ago

Old topic, but I revisited it :). The llama.cpp server does not support CFG at the moment, only the main example does. There is a WIP PR here: https://github.com/ggerganov/llama.cpp/pull/2217