-
### System Info / 系統信息
python 3.11.8
### Running Xinference with Docker? / 是否使用 Docker 运行 Xinfernece?
- [ ] docker / docker
- [X] pip install / 通过 pip install 安装
- [ ] installation from source / 从源…
-
### Feature request / 功能建议
2024-09-18 02:22:06,994 xinference.core.worker 68 INFO [request 690d9782-759f-11ef-af77-0242ac110002] Leave launch_builtin_model, elapsed time: 27 s
2024-09-18 02:22:4…
-
### Your current environment
vllm docker image: vllm/vllm-openai:latest
### 🐛 Describe the bug
It works for the first time then stops generating responses, as shown below.
ChatCompletion(id='c…
-
[lalrpop](https://github.com/nikomatsakis/lalrpop/issues/156)
gwenn updated
6 years ago
-
Hi, is there any plan to add a requirement.txt that allows us to install needed packages with pip? Thanks.
-
It would be helpful to have a hook to allow custom attribute filtering. I propose something much simpler than #24 that would integrate with the existing builder syntax:
```go
// AttrTransform is …
-
### Checked other resources
- [X] I added a very descriptive title to this issue.
- [X] I searched the LangChain.js documentation with the integrated search.
- [X] I used the GitHub search to find a …
-
**Is your feature request related to a problem? Please describe.**
VRAM is a major limitation for running most models locally, and guidance by design requires to run models locally to get the most va…
-
It would be great to get the instructions to run the 3B model locally on a gaming GPU (e.g. 3090/4090 with 24GB VRAM).
### Confirmed GPUs
From this thread
| GPU Model | VRAM (GB) | Tuned-3b | T…
-
I am trying to fine tune bigcode/starcoderbase model on compute A100 with 8 GPUs 80Gb VRAM.
My initial steps are to adjust parameters.
I get some impression that it becomes slow if I increase batch …