Closed Welding-Torch closed 7 months ago
Hi @Welding-Torch, thanks for trying the plugin, could you check the new video? There's a setup walkthrough and explanation.
Can I use Ollama embedding even if I don't have Ollama installed?
Also, I'm getting stuck on this message. It isn't moving on from '4/70':
you need ollama installed to use it. Could you share a screenshot of your console with debug on, as the troubleshoot section of the video showed?
I see, so the LM Studio method isn't a complete solution? I followed the LM Studio method because it seemed easier than Ollama and also it was the first thing in the guide.
Right, LM Studio is chat model only, QA needs an embedding model as well.
I just tried v2.5 with Ollama, nomic-embed-text:latest + dolphin-mixtral:8x7b-v2.7-q6_K. Works perfectly here!
Okay I set it up with Ollama. The speed is... really slow.
Right, LM Studio is chat model only, QA needs an embedding model as well.
I think this should be clarified in the setup guide and Ollama should be treated as the default way to set this up. This can't be done entirely using LM Studio (yet).
Also, do you have any tips to speed it up? My specs are i5-9300H, 16GB Ram, GTX 1650 with 4GB VRAM.
Okay I set it up with Ollama. The speed is... really slow.
Right, LM Studio is chat model only, QA needs an embedding model as well.
I think this should be clarified in the setup guide and Ollama should be treated as the default way to set this up. This can't be done entirely using LM Studio (yet).
Also, do you have any tips to speed it up? My specs are i5-9300H, 16GB Ram, GTX 1650 with 4GB VRAM.
Your specs are a bit low for running local LLMs tbh. Your best bet for Ollama is to deploy it on a remote server with a lot more VRAM, and use that url as an override.
This is an heavy RAM setup running LLM and vector db locally. @logancyang you should maybe run it via openai
Describe the bug A clear and concise description of what the bug is.
Hi there. I'm trying to set up Vault QA. But after following the instructions in the Local Copilot Setup Guide with LM Studio, I'm still not able to start chatting. When I open the chat and set it to
Vault QA (BETA)
, it gives me the message in IMAGE 1. So I click the Refresh Index button and nothing happens. So I try to refresh it from the command palette as seen in IMAGE 2. Running either of the two commands gives me the following error: "An error occurred while re-indexing vault to vector store" as seen in IMAGE 3.Screenshots _IMAGE 1 - What I see after following these instructions._
IMAGE 2 - Command Palette.
IMAGE 3 - The error.
I am unable to chat with my vault.
Additional context
Also, I read in a similar issue (#335) about something called embeddings?
I'm a noob but, what is that? Do I need to pay for an OpenAI API key for it? Or should I abandon using LM Studio and use only Ollama (for the local embeddings) for chatting with my vault?
Thanks.