-
“使用本地Ollama模型
如果你使用本地Ollama模型,需要配置环境变量:OLLAMA_ORIGINS=chrome-extension://bciglihaegkdhoogebcdblfhppoilclp,否则访问会出现403错误。
然后在插件配置里,apiKey随便填一个,服务器地址填http://localhost:11434,模型选自定义,然后填入自定义模型名如llama2。
…
-
In the documentation the Bionic GPT it is mentioned that it works with ollama and OpenAPI compatible backends and it is demonstrated running a local gemma model. I could not find information on how to…
-
### What is the issue?
Hello Team,
Below is the attached server log; I am trying to run llama3.1 70B on
5700x, 23GB RAM and p100 16GB,
the model loads successfully, but as soon as the promp…
-
[ ] I checked the [documentation](https://docs.ragas.io/) and related resources and couldn't find an answer to my question.
**Your Question**
Hi, I am trying to generate test samples using ragas…
-
### Feature request
Recently, we have added the ability to load `gguf` files within [transformers](https://huggingface.co/docs/hub/en/gguf).
The goal was to offer the possibility to users …
-
`gemma-2-27b-it` and `gemma-2-9b-it` were the highest ranked models under 70B in 2024-07-26.
However they disappeared from the latest release of 2024-08-31.
Even weirder, they're not on the [homepag…
-
Version: llama-cpp-python==0.2.82
Model: "bartowski/gemma-2-9b-it-GGUF/gemma-2-9b-it-Q8_0.gguf"
When I load the gemma2 model with temperature=0, and run a simple prompt, it always gives the same o…
-
### Question Validation
- [X] I have searched both the documentation and discord for an answer.
### Question
Hello everyone,
I've developed two RAG (Naive RAG and Advance RAG) applications using…
-
![image](https://github.com/user-attachments/assets/24824fc2-b8b5-4c4e-84ab-eb0331a3d1cb)
I had to create a screenshot because it's not selectable text and I can't copy&paste it.
It was running f…
-
Hello! I had a thought. To minimize constant load for tasks that occur infrequently, is there a way to keep the Docker container running with the HTTP server, but only load the model when a query is m…