-
### What behavior of the library made you think about the improvement?
I have just started to use Outlines, and my use case is that I am hosting a local model on a server using [Serve with vLLM](http…
-
**The bug**
I've been unable to use models from together.ai due to a 401 error relating to huggingface, despite passing in a valid API key, both as a `api_key` init variable and environment variable …
-
### Your current environment
The output of `python collect_env.py`
```sh
docker run --runtime nvidia --gpus all \
-v ~/.cache/huggingface:/root/.cache/huggingface \
--ipc=host -p 80…
-
## Prerequisites
- [X] I read the [Deployment and Setup](https://docs.opencti.io/latest/deployment/overview) section of the OpenCTI documentation as well as the [Troubleshooting](https://docs.openc…
-
- [x] This is actually a bug report.
- [ ] I am not getting good LLM Results
- [ ] I have tried asking for help in the community on discord or discussions and have not received a response.
- [ ] I …
-
> electron-app@0.1.0 dev
> cross-env NODE_ENV=development concurrently -k "cross-env BROWSER=none npm run next:dev" "npm run electron:dev"
[1]
[1] > electron-app@0.1.0 electron:dev
[1] > npm ru…
-
Both the OpenAI and MistralAI chat completion clients are respecting AutoFunctionInvocationFilters. There's no code in Gemini implementation that pays attention to it at all.
-
### How are you running AnythingLLM?
Docker (remote machine)
### What happened?
I've embedded the chat widget in a web page. When asking a question via the chat embed, the response is truncated. Wh…
-
I didn't see any documentation that mentions that.
-
### System Info
- `transformers` version: 4.38.1
- Platform: Linux-6.2.0-1019-azure-x86_64-with-glibc2.35
- Python version: 3.10.12
- Huggingface_hub version: 0.21.1
- Safetensors version: 0.4.…
dfdx updated
6 months ago