-
I am trying to fine tune Mistral 7B model and want to use PEFT setting. but getting following error :
File [~/.conda/envs/env/lib/python3.10/site-packages/dspy/teleprompt/finetune.py:166](http://lo…
-
### Python Version
```shell
Python 3.10.12
```
### Pip Freeze
```shell
absl-py==2.1.0
annotated-types==0.7.0
astunparse==1.6.3
attrs==24.2.0
beautifulsoup4==4.12.3
blis==0.7.11
bs4…
-
Dear authors,
in addition to bge-series, I would like to see how other embedding model perform on my own custom dataset.
I was wondering if I can use the following fine-tuning script for fine-tu…
-
### Python Version
```shell
Appreciate any help solving the issue...
(I've seeing in other threads people blaming this type of crashes on the CPU memory, but a g4dn.12xlarge has 192Gb RAM. So unless…
-
Hello, thanks a lot for sharing your training code!
In the training script (https://github.com/LLaVA-VL/LLaVA-NeXT/blob/main/scripts/train/finetune_onevision.sh), is it possible to use the trained L…
-
Hi,
I recently fine-tuned the phi-3.5-moe-instruct model and phi-3.5-mini-instruct model using PEFT LORA. It seems the Moe model is performing way worse than 3.5 Mini Are there any specific things …
-
Simply adding a newline to the default template in `chat_templates.construct_chat_template` causes a RuntimeError:
The template:
```
system
{SYSTEM}user
{INPUT}assistant
{OUTPUT}user
…
-
# Common Issue
More Questions will be added......
## Training Related
Q : Can not finetune the existing LLaVA-Onevision checkpoints
A : We edit our model's config so that it is able to be se…
-
Hi, Thanks for your wonderful work.
I am struggling using my lora tuned model.
I conducted following steps
1. finetuning with lora
- Undi95/Meta-Llama-3-8B-Instruct-hf model base
- llama3 …
-
Following up on a longer internal discussion we had (cc @carmocca @lantiga @awaelchli ), we want to support the following user-friendly API in LitGPT:
```bash
# ligpt [action] [model]
litgpt dow…
rasbt updated
4 months ago