TransformerOptimus / SuperAGI

<⚡️> SuperAGI - A dev-first open source autonomous AI agent framework. Enabling developers to build, manage & run useful autonomous agents quickly and reliably.
https://superagi.com/
MIT License
14.94k stars 1.78k forks source link

Using Mixtral as Local LLM Fails #1401

Open CharlesMod opened 4 months ago

CharlesMod commented 4 months ago

⚠️ Check for existing issues before proceeding. ⚠️

Where are you using SuperAGI?

Linux

Which branch of SuperAGI are you using?

Main

Do you use OpenAI GPT-3.5 or GPT-4?

GPT-3.5

Which area covers your issue best?

Agents

Describe your issue.

Attempt to use nous-hermes-2-mixtral-8x7b-sft.Q4_K_M.gguf from TheBloke using the standard Local LLM loader shown in the youtube video released this January.

How to replicate your Issue?

Edit docker-compose-gpu.yml to mount the volume containing the local llm model.

Then, attempt to run the model with a new agent. This will result in "no model found". (The docker log in the CLI will give more information, the error occurs right after loading the model after running the agent.)

Upload Error Log Content

backend-1 | error loading model: create_tensor: tensor 'blk.0.ffn_gate.weight' not found backend-1 | llama_load_model_from_file: failed to load model backend-1 | 2024-02-04 22:11:06 UTC - Super AGI - ERROR - [/app/superagi/helper/llm_loader.py:27] - backend-1 | from_string grammar: backend-1 | backend-1 | 2024-02-04 22:11:06 UTC - Super AGI - ERROR - [/app/superagi/controllers/models_controller.py:185] - Model not found. backend-1 | 2024-02-04 22:11:06 UTC - Super AGI - INFO - [/app/superagi/controllers/models_controller.py:203] - Error: backend-1 | 2024-02-04 22:11:06 UTC - Super AGI - INFO - [/app/superagi/controllers/models_controller.py:203] -

CharlesMod commented 4 months ago

Here is a more complete error log: errorLogsDocker.txt

memamun commented 3 months ago

I am also facing the same issue. Have you able to get it fixed?

rounak610 commented 2 months ago

@memamun @CharlesMod could you try running mixtral with the "fixes_for_mixtral" branch instead of the main branch and let me know if you face any error

zero-stroke commented 3 weeks ago

Were you able to get this working, either with mixtral or with any local model?