-
Hi,
I'm trying to reproduce the section "[How do I run it locally?](https://huggingface.co/blog/personal-copilot#how-do-i-run-it-locally)" from this blog post:
[Personal Copilot: Train Your Own …
-
**Is your feature request related to a problem? Please describe.**
Code models tend to generate long responses, but without token streaming the chat interface ends up feeling unresponsive.
…
-
Is it possible to integrate converter scripts for the GPTBigCodeForCausalLM architecture from the transformers libary?
This would enable integration of models like Starcoder / Santacoder.
With t…
-
Hello,
I have been trying to use the finetune.py script with my own dataset on a single H100 GPU with CUDA 11.8
I have been getting the following error.
The same script and dataset are working wh…
-
**Describe the bug**
StartCoder inference with AutoTP doesn't work.
I get the following error:
```
File "[...]/venv38/lib64/python3.8/site-packages/transformers/models/gpt_bigcode/modeling_gpt_b…
Epliz updated
10 months ago
-
### What happened?
Running the following code results in an error. How can it be fixed?
code:
```
import litellm
from litellm import CustomLLM, completion, get_llm_provider
class Qwen(Cust…
-
Hi! I am trying to use the tool but somehow the code completion is not working. The chat functionality works just fine so I am quite sure I configured the connectors properly. Unfortunately, I couldn…
-
Hello,
I am running inferences on StarCoder on a 112GB RAM CPU cluster. While asking StarCoder to help find some issues in my code, it highlights possible errors but it also generates some junk out…
-
I am trying to run Starcoder locally through Ollama. And I want to get code auto-completion like in the README gif.
But I keep getting the following error after every debounce: `[LLM] inference api…
V4G4X updated
6 months ago
-
First of all, thank you for your work! I used ggml to quantize the starcoder model to 8bit (4bit), but I encountered difficulties when using GPU for inference. If you can provide me with an example, I…