Closed Sneakr closed 4 months ago
I pushed some updates that should make this work but the resulting file sizes seem off, so I probably also need to update llama-cpp-conv too.
Mixtral tested out OK with an updated llama-cpp-conv, so I am closing this.
Seems not work for mixtral.
File "/home/eyera/miniconda3/envs/unsloth_env/lib/python3.10/site-packages/quantkit/convert.py", line 1417, in do_gguf_conversion
model = convert_model_names(model, params, False) #skip_unknown
File "/home/eyera/miniconda3/envs/unsloth_env/lib/python3.10/site-packages/quantkit/convert.py", line 1207, in convert_model_names
raise Exception(f"Unexpected tensor name: {name}. Use --skip-unknown to ignore it (e.g. LLaVA)")
Exception: Unexpected tensor name: model.layers.0.block_sparse_moe.gate.weight. Use --skip-unknown to ignore it (e.g. LLaVA)