Closed sierra-alpha closed 3 weeks ago
You don't need to add chat buffer to context explicitly. You can just call ellama-chat
multiple times and it should work. Maybe zephyr model just not smart enough. Try newer models from ollama like gemma2 9b, Llama 3.1 8b, Qwen2.5 7b or others.
Ah okay, was just following the defaults. I'll try with a different model and report back.
In the LLM space, things have been improving very fast. Zephyr was a good model when it was released, but for now it is very bad compared to newer models of the same size. And I can't change defaults without breaking the setup for people who are already using it.
Sorry, it took me a while to get around to testing this out. Yep I can confirm that a later model works as expected, great stuff thanks for that!
Wise call on the defaults, thanks for an awesome package!
I'll close this now.
Similar to this issue: https://github.com/s-kostyaev/ellama/issues/5
I'm using ellama defaults as best as I can tell (using
llm-client
layer in spacemacs, but I did try anemacs -q
and package installellama
and had the same issue).I'd expect to be able to ask and have the LLM answer multiple questions in the same context in the org buffer started. But it seems that's not working. I use
m-x ellama-chat
and ask a question in the minibuffer, it then opens up an org mode buffer with my question and populates the answer, if I then usem-x ellama-chat
again I'd expect it to have the context of the previous question too, an example (the second and third Q&A I gave the whole buffer as context also, but I don't think that should be neccesary):