-
- [ ] [README.md · defog/sqlcoder-7b-2 at main](https://huggingface.co/defog/sqlcoder-7b-2/blob/main/README.md?code=true)
# README.md · defog/sqlcoder-7b-2 at main
**DESCRIPTION:**
```yaml
license:…
-
### Issue description
It seems to me that parameter `threads` doesn't work as expected
### Expected Behavior
If I have 24 CPUs and pass `threads:24` then all CPUs should be utilized. II tried call …
-
Hi, script is becoming a separate valuable system now. Great!! I just found issue. I have Ollama server as separate instance. Its accessible in default port but during crew creation it turns out its l…
-
Add a `/stats` REST endpoint that explains how many items are currently in the Redis cache.
I want to be able to evaluate cache performance.
@autopilot Please do this.
-
Hi I finally got it working and I'm going to share my step by step to make this work.
#### My system:
RTX 3060 12GB
CUDA 12.1
Windows 10
PHPSTORM 2023.2.4
## Step 1 - Install TGI
Follow the…
-
### Describe the bug
When i use xinference to run [codellama-70b-instruct](https://huggingface.co/codellama/CodeLlama-70b-hf/tree/main). It output a set of unrelated text.
Just like below:
![im…
-
### System Info
peft: 0.12.1.dev0
accelerate: 0.33.0.dev0
transformers: 4.45.0.dev0
platform: ubuntu22.04 LTS
python 3.10.12
hardward: NVIDIA RTX2080TI * 4
### Who can help?
_No response_
###…
-
Notes on running llamas and other self-hosted llm models on multiple GPUs
-
It looks like the Llama models Code Llama and Llama Guard 2 are missing from this library. Since this library seems to be for using Llama models going forward and these are the only two models left, c…
xNul updated
2 months ago
-
### Checked other resources
- [X] I added a very descriptive title to this issue.
- [X] I searched the LangChain documentation with the integrated search.
- [X] I used the GitHub search to find a…