-
Hi,
I see that the Llava model is trained with full finetuning of the LLM.
Did you do the ablation using LoRAs?
If so do you have a version with LoRA instead of full finetuning?
Thanks!
-
It appears that models will be loaded on different gpus when setting `num_processes` to more than one, which will cause error:
`RuntimeError: Expected all tensors to be on the same device, but found …
-
````
koboldcpp_163.exe --model "knowledge/Meta-Llama-3-8B.Q8_0.gguf" --threads 12 --noshift --smartcontext --contextsize 16384 --usemlock --blasbatchsize 2048 --useclblast 0 0 --gpulayers 2 --bantoke…
-
我自己尝试复现, 在英文领域表现更低了, 感觉不正常.
我在原版llava上进行的改动, 我只改了对输入进行preprocess的整体逻辑,包括mask targets 这部分和 conversation部分的内容.
请问其他还有啥需要改动的地方嘛?
-
### What is the issue?
Hi, Downloaded latest llama3 model after installing ollama for **Windows** from https://www.ollama.com
I have downloaded llama3 latest model.
```
C:\Windows\System32>oll…
-
### Checklist
- [X] 1. I have searched related issues but cannot get the expected help.
- [X] 2. The bug has not been fixed in the latest version.
### Describe the bug
The latest lmdeploy 0.4.1
Th…
-
请问使用llava进行SFT时,如何添加special tokens呢?主要是tokenizer.add_tokens后,模型应该如何操作
-
您好,我在指定本地路径后,仍然需要到 HuggingFace 下载,结果报错。我希望知道哪里设置错了。我想训练 llava-llama3-8b。
- 命令:`NPROC_PER_NODE=1 xtuner train llava_llama3_8b_instruct_quant_clip_vit_large_p14_336_e1_gpu1_pretrain --deepspeed deepspe…
-
It would be good to replace llama 2 with llama 3 as 2 is a very old model now.
-
> ```shell
> pip's dependency
> ```
Hi @Luo-Z13,
- The error related to `pip's dependency` can be ignored.
- The error `TypeError: pad_sequence(): argument 'padding_value' (posi…