Closed freckletonj closed 7 months ago
I found some discussions, but have not dug in sufficiently yet.
https://huggingface.co/TheBloke/deepseek-coder-33B-instruct-GGUF/discussions/2
https://github.com/oobabooga/text-generation-webui/discussions/4495
ctransformers hasn't been updated since early September and doesn't support models that use a BPE vocab, like Deepseek Coder
I'd recommend using llama-cpp-python until ctransformers has been updated - we're not sure if/when that will happen as marella, the developer, is not available at the moment.
Thanks @TheBloke, I'll switch over. Also, geez you have incredible work ethic and a positive impact on this ecosystem. Thank you!
I tried out
deepseek-coder-X-.gguf
withctransformers
. https://huggingface.co/TheBloke/deepseek-coder-33B-instruct-GGUFUpon loading it, ie
model = AutoModelForCausalLM.from_pretrained(...)
, I get the following error:For reference, this model works with
llama.cpp
.