Closed AtlasPilotPuppy closed 10 months ago
I apologize. You now need an additional argument "quantized" and it should work. README is updated
However for some reason that particular repo is producing bad output for me. Need to check why.
okay, fixed it. Needed 0.3 version of auto-gpt.
I exported the following model:
gsaivinay/Llama-2-7b-Chat-GPTQ
on the main branch.using the export script.
The model is a 7B model with a group size of 128.
I compiled the rust binary with the following appropriate parameters
cargo build --release --features 7B,group_128,python
and then ran
pip install .
I get the following error trying to import the model:
I have tried this with a few different models, group sizes and parameter sizes and still get the same error. Any guidance on debugging this will be quite helpful.