-
I'm looking through the code to see how different dataset formats besides the default LLAVA dataset format could be supported e.g. [{"question": "...", "answer": "..."}]
Doing so I wonder if the in…
-
## ❓ General Questions
Hi guys,
I modified Llava so that the modified model uses GPT2 tokenizer instead of Llama tokenzier.
I trained the model, and
1. convert the weights,
2. compiled the…
-
This thread is dedicated to discussing the setup of the webui on AMD GPUs.
You are welcome to ask questions as well as share your experiences, tips, and insights to make the process easier for all…
-
## ⚙️ Request New Models
- Link to an existing implementation (e.g. Hugging Face/Github):
- Is this model architecture supported by MLC-LLM? (the list of [supported models](https://llm.mlc.ai/do…
-
### Your current environment
```
The output of `python collect_env.py`
Collecting environment information...
/usr/local/cuda-11.0/targets/x86_64-linux/lib/libcudnn_ops_train.so.8
/usr/local…
-
### Describe the issue
Issue:
When I try EP_SIZE > 1 (like EP_SIZE = 2), the program hangs at beginning. It hasn't even finished the first batch yet.
May I ask if you have encountered this issue be…
-
Thank you for your excellent job.
I followed your work and download the released dataset from your link.
Since you have kindly provided an end-to-end script and processed dataset file. I thought we…
-
### What happened?
I get a CUDA out of memory error when sending large prompt (about 20k+ tokens) to Phi-3 Mini 128k model on laptop with Nvidia A2000 4GB RAM. At first about 3.3GB GPU RAM and 8GB CP…
-
### Checklist
- [ ] 1. I have searched related issues but cannot get the expected help.
- [ ] 2. The bug has not been fixed in the latest version.
### Describe the bug
```bash
(lmdeploy042) yuzail…
-
## 🐛 Bug
I am trying to work with StableLM 1.6b model.But getting error in model compilation step.
## To Reproduce
Steps to reproduce the behavior:
1. Library Installation:
```
!python3 …