Open PlanetMacro opened 10 months ago
Would it be possible to reformat the Quantized models of MIXTRAL 8x7B to run in sparse /base mode in LLaMA2-Accessory?
Theoretically all mixtral implementations are interchangeable. However, since I don't know what your Quantized model looks like, I am not able to suggest the specific pipeline
https://huggingface.co/TheBloke/Mixtral-8x7B-v0.1-GGUF
Would it be possible to reformat the Quantized models of MIXTRAL 8x7B to run in sparse /base mode in LLaMA2-Accessory?