-
Hi,
I have tried adding phi3-3.8b, as an ollama model, hosted on my own prem ollama server.
I have basically copied the prompt template and parameters from microsoft/Phi-3-mini-4k-instruct used in h…
-
# Problem
* Until we can solve the problem of the 403 access #676, there's no way to pull models from the Ollama server
* At the time I'm writing this, I don't think the Ollama registry (Docker …
-
How to solve the problem in the picture below
![Snipaste_2024-10-22_21-42-05](https://github.com/user-attachments/assets/a6967a17-ad9e-4ccd-bb13-06ee2d7747da)
-
I am opening a report identical to one that had been closed previously for lack of more information: #8869
Open question from @chrmarti was:
> The boostrap container mounts the /var/run/docker.…
-
### Motivation
LMDeploy's 4-bit quantized prefix cache (along with 4-bit AWQ for weights) allows running ~70B models on 48GB of RAM with good performance for many-user scenarios. The prefix cache c…
-
Can we pull llama3 using this genai-stack? I was not sure, if the docker files are configured to pull llama3......Have pulled and used llama2 before.....
-
Currently memgraph supports ollama for AI chat however it hardcodes llama2:latest and no other options are available. This enhancement would allow the memgraph-lab user to select the model to use with…
-
I run the docker compose up command and all was installed correctly.
I entered the ollama docker container and installed llama2 but when I run devika, no LLM is found for ollama.
Should I configure …
suoko updated
6 months ago
-
While trying to get GenAI stack up (docker compost up --build) I am getting the error:
**genai-stack-pull-model-1 | pulling ollama model llama2 #or any Ollama model tag, gpt-4, gpt-3.5, or claudev…
-
running:
`cm run script --tags=run-mlperf,inference,_find-performance,_full,_r4.1 --model=llama2-70b-99 --implementation=reference --framework=pytorch --category=datacenter --scenari…