Open charles-cai opened 7 months ago
OpenChat-Cuda works great with my 3080 10GB GPU under Unraid Docker environment.
Could someone explain how to swap the LLM model to https://huggingface.co/TheBloke/Mistral-7B-Code-16K-qlora-GGUF ?
I want to test code generation comparison between Llama 2 7B vs Mistral 7B Code.
Many thanks. Charles
OpenChat-Cuda works great with my 3080 10GB GPU under Unraid Docker environment.
Could someone explain how to swap the LLM model to https://huggingface.co/TheBloke/Mistral-7B-Code-16K-qlora-GGUF ?
I want to test code generation comparison between Llama 2 7B vs Mistral 7B Code.
Many thanks. Charles