-
what was the maximum sequence length used for finetuning starcoder to produce star chat alpha? Was it done on a single GPU card or multiple cards? Please provide insights on the memory requirement…
-
hello after loading the model i asked it what are you able to generate and it responded with a question mark then I asked what project we were working on and it gave me a link to someone's google driv…
-
![image](https://github.com/vllm-project/vllm/assets/26181650/70c44266-fbb4-4945-a567-78e9eb4c9b06)
-
support More model like: https://github.com/THUDM/ChatGLM-6B,https://github.com/openai/whisper
-
Hi,
What's the process in finetuning BLOOM?
Did anyone succeed and willing to share the code?
Thanks!
-
Consider switching to [langchain](https://github.com/hwchase17/langchain). Need to consider the positives and negatives. This requires some replacement of the backend, specifically in the `BaseChatInt…
-
Hi Experts,
Recently some of the emerging models use MQA (Multi-Query Attention) or GQA (Grouped-Query Attention), From issues list, I noticed that some users have already mentioned about the suppo…
ljayx updated
10 months ago
-
**Describe the bug**
Can get work on tabby 0.13.1 or 0.14.0 follow by the quick-start guide, it's just start process with the a embed model
/opt/tabby/bin/llama-server -m /data/models/TabbyML/Nom…
-
How can I explain code in this model?I know that code interpretation can be done through the application of chatCould. But could you please give us some api or code example to explain code?
-
Hi @loubnabnl, thanks for this great repo.
I've seen a blog from the VMware OCTO, which described their works on fine-tuning **star-coder**, but `modified the code provided by the [SantaCoder](http…