-
### Checked other resources
- [X] I added a very descriptive title to this issue.
- [X] I searched the LangChain documentation with the integrated search.
- [X] I used the GitHub search to find a…
-
### What happened?
Commit 1553f7fa4844ea4d4117c7a75d165ca2e747b81a introduced some incompatibilities and causes `AttributeError: 'CompletionUsage' object has no attribute 'get'` thrown by https://git…
-
If you are submitting a bug report, please fill in the following details and use the tag [bug].
**Describe the bug**
Gemma-2-{size} is not loadable using from_pretrained. I checked OFFICIAL_MODEL_…
-
### System Info
```Shell
colab
t4
https://huggingface.co/docs/accelerate/concept_guides/
https://huggingface.co/docs/accelerate/concept_guides/big_model_inference
If I have a single 16 G…
-
### My current environment
```text
PyTorch version: 2.3.0+cu121
Is debug build: False
CUDA used to build PyTorch: 12.1
ROCM used to build PyTorch: N/A
OS: Ubuntu 22.04.3 LTS (x86_64)
GCC vers…
-
Blog
https://medium.com/@obrienlabs/running-the-70b-llama-2-llm-locally-on-metal-via-llama-cpp-on-mac-studio-m2-ultra-32b3179e9cbe
https://www.linkedin.com/posts/michaelobrien-developer_running-70b-…
-
I can see others having success running mistral on a 3090. Am I doing something wrong?
```
Request at 2024-03-29 20:33:44.758325187 -07:00: {"messages":[{"content":"What is the capital of France?"…
-
**Describe the bug**
LLM tool reports error code: 422 when using Mistral-large based serverless connection.
**How To Reproduce the bug**
1. deploy Mistral-large from model catalog as serverless e…
-
I am using the default set-up with default sequence length, i tried pushing the q4 quant but it fails .
I also tried pushing the 16 bit merged aswell , i tried pushing the model itself without finet…
-
Research task leads to this error in the terminal and the UI hangs without any information.
```
n\nWrite 2 paragraphs completely covering the main topics based on what you know and use the web sni…