-
- Paper name: From Quantity to Quality: Boosting LLM Performance with Self-Guided Data Selection for Instruction Tuning
- ArXiv Link: https://arxiv.org/abs/2308.12032
To close this issue open a …
-
I tried fine-tuning **Llama 2**, **Llama 3** & even **LLama 3.1** but my loss is decreasing/increasing. I can't figure out.
I have my dataset in alpaca format like this:
```
[
{
…
-
Like what I said, does it support the title? does it multimodal-in, multimodal-out(with multi images)?
-
### Question
Hi~
100 hours was needed for instruction tuning in my experiment.....is this nomal?
2machines×4A100-40G(total 8A100)was used in our fintuning,same dataset with paper.
Due to the 40G…
-
Hi,
Thanks a lot for open source the code to fetch the FLAN data set.
I noticed in the paper: The Flan Collection: Designing Data and Methods for Effective Instruction Tuning. (https://arxiv.or…
quq99 updated
7 months ago
-
The fine-tuning code runs when I replace the base model with LLAMA-2.
I am aware that LLAMA and LLAMA-2 share the same configuration files and other associated components.
However, I would sti…
-
I am performing a Mega Merge using LLaMA 3.2 3B, both the base model and fine-tuning/instruction tuning, with the DARE linear method. Following the successful completion of the initial merge, I encoun…
-
非常感谢您很有意义的工作,想请教一下所使用到的instruction-tuning的数据量。
另外,想再请教一下是否有探究多大的instruction-tuning数据量就够用了呢?
非常感谢
-
## Overview
Our previous stable release was cut on 2024-11-04: https://github.com/iree-org/iree/releases/tag/candidate-20241104.1068. We aim for roughly one stable release every 6 weeks, though in th…
-
### Question
I wonder about the performance when using Qwen2 as the LLM. Does it outperform the original LLaVA-v1.5?
By the way, are there any scripts for instruction tuning? I only found the scri…