Lightning-AI / lit-llama

Implementation of the LLaMA language model based on nanoGPT. Supports flash attention, Int8 and GPTQ 4bit quantization, LoRA and LLaMA-Adapter fine-tuning, pre-training. Apache 2.0-licensed.
Apache License 2.0
6k stars 520 forks source link

Mistral Model #458

Open PierreColombo opened 1 year ago

PierreColombo commented 1 year ago

Hello ! Is there any plan to incorporate Mistral ai Model? Cheers,

lantiga commented 1 year ago

Hey, lit-GPT supports it https://github.com/Lightning-AI/lit-gpt/blob/main/tutorials/download_mistral.md with capped context length (which should be enough for most practical uses right now)