ggerganov / llama.cpp

LLM inference in C/C++
MIT License
64.58k stars 9.24k forks source link

Feature Request: Add support for Lite-Mistral-Instruct chat template #8529

Closed AmgadHasan closed 6 days ago

AmgadHasan commented 1 month ago

Prerequisites

Feature Description

OuteAI has released a new small model that is very coherent for its size.

I am requesting the addition of this model's chat template to llama.cpp's list of supported templates

Motivation

The model is already supported by llama.cpp. However, it's using a new chat template that isn't in the list of supported templates. As a result, llama.cpp assumes ChatML for this model. Due to the model's size, it's very sensitive to the prompt template and gives terrible results with wrong formats and users have terrible experience running it.

Possible Implementation

The model's template is very similar to ChatML so we can just copy-paste the implementation for ChatML and modify it.

AmgadHasan commented 1 month ago

I've opened a PR to add support for this template: https://github.com/ggerganov/llama.cpp/pull/8522

github-actions[bot] commented 6 days ago

This issue was closed because it has been inactive for 14 days since being marked as stale.