-
I wanted to ask, if anyone has used the `intfloat/e5-mistral-7b-instruct` as base model for a multiclass classification task. I am trying to use but I have a problem with the GPU memory and I don't kn…
-
### Question
Many thanks to the authors for this very good work.
I'm trying to run the instruct tuning part on one 3090 GPU, using the int-8 mode:
```
python llava/train/train.py \
--…
-
Hi,
How many epoch did your checkpoint train ?Thanks.
-
### Your current environment
Collecting environment information.
PyTorch version: 2.3.0+cu121
Is debug build: False
CUDA used to build PyTorch: 12.1
ROCM used to build PyTorch: N/A
0S: Cent0S Li…
-
It looks like the tokenizer patching breaks. Here's the log:
```
ValueError Traceback (most recent call last)
Cell In[1], line 20
7 # 4bit pre quantized models…
rwl4 updated
3 months ago
-
I started a new installation of `GPT-2` using this fork of the project, following the directions in https://www.youtube.com/watch?v=4iK-IuvatxI (for training) and https://lambdalabs.com/blog/run-opena…
-
### Prerequisites
- [X] I am running the latest code. Mention the version if possible as well.
- [X] I carefully followed the [README.md](https://github.com/ggerganov/llama.cpp/blob/master/README.md)…
-
Hi,
I am trying to reproduce the results of llama-adapter v2. I am finetuning the model with "alpaca_gpt4_data" and "llava_instruct_150k" datasets and using the settings from https://github.com/Ope…
dmlpt updated
10 months ago
-
When i trained llava-llama3 use your code, the log print tokenization mismatch as below.
how to fix it?
thanks!
WARNING: tokenization mismatch: 55 vs. 54. (ignored)
WARNING: tokenization m…
-
I can load the model using below code:
```
from transformers import AutoModelForCausalLM, AutoTokenizer
model_id = "/root/private_data/models/Meta-Llama-3.1-70B-Instruct"
model = AutoModelForC…