Open Ligerbot opened 2 weeks ago
it runs slow because llama3 is selfhosted ai, you need better hardware to run atleast 32gb of ram for better performance
But llama3 runs fast when I am not using mindcraft. It usually just responds in a few seconds.
That ollama config that works the best at lease for me is.
"model": "dolphin-llama3", "embedding": "all-minilm",
This ran better than anything I could run on ollama since its self hosted. Here is my pc specs, CPU: AMD Ryzen 5 3600, Memory: 48gb, GPU: Gigabyte Eagle 8GB Nvidia GeForce RTX 3060 OS: Win 11
That ollama config that works the best at lease for me is.
"model": "dolphin-llama3", "embedding": "all-minilm",
This ran better than anything I could run on ollama since its self hosted. Here is my pc specs, CPU: AMD Ryzen 5 3600, Memory: 48gb, GPU: Gigabyte Eagle 8GB Nvidia GeForce RTX 3060 OS: Win 11
I get error "Unknown embedding: all-minilm . Using word overlap."
Whenever I give the llama3 model a prompt it takes about 5 minutes to actually respond and sometimes is just times out. When I use a smaller model like Tinyllama it goes faster but tinyllama has no idea how to use the mindcraft tools. My computer is powerful enough to run the model pretty fast and llama3 can answer a prompt in a couple seconds on my computer, but when I use the mindcraft thing it is super laggy. My laptop specs are
And I am running Debian 12.