KoljaB / LocalAIVoiceChat

Local AI talk with a custom voice based on Zephyr 7B model. Uses RealtimeSTT with faster_whisper for transcription and RealtimeTTS with Coqui XTTS for synthesis.
Other
509 stars 55 forks source link

Switch model to Llama 3 #17

Open mirix opened 2 weeks ago

mirix commented 2 weeks ago

Thanks for this great project. It works quite well.

However, I would like to change the model to:

https://huggingface.co/mlabonne/Meta-Llama-3.1-8B-Instruct-abliterated-GGUF

I have been trying to adapt the prompt template. It kind of works, even without changing the template, but eventually it goes into an endless response loop. I guess that the issue are the stop tokens or something like that.

Has anyone tried this?

KoljaB commented 2 weeks ago

Quite sure you are right, the reason will most probably be the different prompt formats ("chat templates") the models use. This project is a bit outdated and therefore still uses the raw template format.

Meanwhile most of this can be abstracted away. Today most LLM providers offer servers with chat endpoints, like these:

Or you can directly use huggingface transformers.pipeline, which abstracts the model chat template away.

If you want to use the chat endpoints, the LocalEmotionalAIVoiceChat project has example code that does it this way.

Alternatively you can also use then python OpenAI library to wrap the base endpoints (http://localhost:1234/v1). For example code, Linguflex project does some LLM requests in this way.