logancyang / obsidian-copilot

THE Copilot in Obsidian
https://www.obsidiancopilot.com/
GNU Affero General Public License v3.0
2.92k stars 201 forks source link

Unable to use Vault QA - "An error occurred while re-indexing vault to vector store" #340

Closed Welding-Torch closed 7 months ago

Welding-Torch commented 7 months ago

Describe the bug A clear and concise description of what the bug is.

Hi there. I'm trying to set up Vault QA. But after following the instructions in the Local Copilot Setup Guide with LM Studio, I'm still not able to start chatting. When I open the chat and set it to Vault QA (BETA), it gives me the message in IMAGE 1. So I click the Refresh Index button and nothing happens. So I try to refresh it from the command palette as seen in IMAGE 2. Running either of the two commands gives me the following error: "An error occurred while re-indexing vault to vector store" as seen in IMAGE 3.

Screenshots image 1 _IMAGE 1 - What I see after following these instructions._

image 2 IMAGE 2 - Command Palette.

image 3 IMAGE 3 - The error.

I am unable to chat with my vault.

Additional context

Also, I read in a similar issue (#335) about something called embeddings?

"Indexing relies on embeddings" - @logancyang

I'm a noob but, what is that? Do I need to pay for an OpenAI API key for it? Or should I abandon using LM Studio and use only Ollama (for the local embeddings) for chatting with my vault?

Thanks.

logancyang commented 7 months ago

Hi @Welding-Torch, thanks for trying the plugin, could you check the new video? There's a setup walkthrough and explanation.

Welding-Torch commented 7 months ago

Can I use Ollama embedding even if I don't have Ollama installed?

Also, I'm getting stuck on this message. It isn't moving on from '4/70': image

logancyang commented 7 months ago

you need ollama installed to use it. Could you share a screenshot of your console with debug on, as the troubleshoot section of the video showed?

Welding-Torch commented 7 months ago

I see, so the LM Studio method isn't a complete solution? I followed the LM Studio method because it seemed easier than Ollama and also it was the first thing in the guide.

logancyang commented 7 months ago

Right, LM Studio is chat model only, QA needs an embedding model as well.

pascalandy commented 7 months ago

I just tried v2.5 with Ollama, nomic-embed-text:latest + dolphin-mixtral:8x7b-v2.7-q6_K. Works perfectly here!

Welding-Torch commented 7 months ago

Okay I set it up with Ollama. The speed is... really slow. image

Right, LM Studio is chat model only, QA needs an embedding model as well.

I think this should be clarified in the setup guide and Ollama should be treated as the default way to set this up. This can't be done entirely using LM Studio (yet).


Also, do you have any tips to speed it up? My specs are i5-9300H, 16GB Ram, GTX 1650 with 4GB VRAM.

logancyang commented 7 months ago

Okay I set it up with Ollama. The speed is... really slow. image

Right, LM Studio is chat model only, QA needs an embedding model as well.

I think this should be clarified in the setup guide and Ollama should be treated as the default way to set this up. This can't be done entirely using LM Studio (yet).

Also, do you have any tips to speed it up? My specs are i5-9300H, 16GB Ram, GTX 1650 with 4GB VRAM.

Your specs are a bit low for running local LLMs tbh. Your best bet for Ollama is to deploy it on a remote server with a lot more VRAM, and use that url as an override.

pascalandy commented 7 months ago
SCR-20240307-oqai

This is an heavy RAM setup running LLM and vector db locally. @logancyang you should maybe run it via openai