-
This issue contains the test results for the upstream sync, develop PR, and release testing branches. Comment 'proceed with rebase' to approve. Close when maintenance is complete or there will be prob…
-
### Prerequisites
- [X] I am running the latest code. Mention the version if possible as well.
- [X] I carefully followed the [README.md](https://github.com/Mozilla-Ocho/llamafile/blob/master/README.…
-
### Your current environment
```text
Collecting environment information...
PyTorch version: 2.3.1+cu121
Is debug build: False
CUDA used to build PyTorch: 12.1
ROCM used to build PyTorch: N/A
…
-
### Describe the issue
Hi all I'm trying to replicate the code in this tutorial, with a local llm using LM Studio endpoint
https://microsoft.github.io/autogen/docs/tutorial/tool-use
When I initiate…
-
### What is the issue?
环境:
直接下的release 0.2.5 的ollama,显卡rx 570 gfx803 win10 64位
运行 ollama run qwen2:1.5b 或者 ollama run phi3 到报错, 请问是需要自己重新编译么 或者我的环境缺失什么依赖
日志
```
2024/07/19 14:10:24 routes…
-
### Your current environment
```text
The output of `python collect_env.py`
```
### 🐛 Describe the bug
Recently, we have seen reports of `AsyncEngineDeadError`, including:
- [ ] #5060
…
-
When I run Ollama on my local PC with model gemma:2b I get a response.
My rest call works, below is a print screen:
![image](https://github.com/OpenDevin/OpenDevin/assets/19372922/307fbce0-9599-48…
-
Hi. I am using exactly the same code as yours in run_sft.sh:
```
#!/bin/bash
CUR_DIR=`pwd`
ROOT=${CUR_DIR}
export PYTHONPATH=${ROOT}:${PYTHONPATH}
VISION_MODEL=openai/clip-vit-large-pa…
-
is there a working example for using Ollama? Or is it not supposed to work? Did try, but without any success.
Thanks in advance
-
I'm working with a long context model (gradientai/Llama-3-8B-Instruct-262k) that exceeds the memory of a single A100 GPU. While the model weights are loaded, when I try to run inference, I get CUDA Ou…