Open sammcj opened 2 months ago
+1, need to try this with ollama :)
yes
won't have bandwidth to do this, but if anyone is interested, that'd be amazing!
Yeah, this is hard to work with if it isn't in GGUF format to run locally, or available from Ollama directly. I'm looking into how to serve from Safetensors but not a lot of servers support that.
also cc @2015aroras
It still isn't merged 😞
It would be great to see OLMoE/OlmoeForCausalLM Llama.cpp/GGUF support.
Really neat project!