-
错误日志
```
/root/anaconda3/envs/cosyvoice/lib/python3.8/site-packages/torch/_jit_internal.py:726: FutureWarning: ignore(True) has been deprecated. TorchScript will now drop the function call on compil…
-
### System Info
Hello I am trying to load Mistral-Nemo Instruct-2407 in bnb 4bit on 4 A10 gpus on ec2 instance.
I upgraded all the packages.
Still I face cuda memory out of error when train batc…
-
Hi,
I successfully ran the finetuning code using config/pretrain/saprot.py and config/Thermostability/saprot.py
Then I newly got these questions
I would really appreciate it if you could answer…
-
I want to run [sft](https://github.com/huggingface/peft/tree/main/examples/sft) example and I get some erros, Can you help me to find the problem?
I run [run_peft_fsdp.sh](https://github.com/huggin…
-
Loading PhotoMaker components [1] id_encoder from [/root/.cache/huggingface/hub/models--TencentARC--PhotoMaker/snapshots/d7ec3fc17290263135825194aeb3bc456da67cc5]...
Loading PhotoMaker components [2]…
-
Hi!👋
I'd like to translate the following files into Korean.
1. [docs/source/en/tutorials/autopipeline.md](https://github.com/huggingface/diffusers/blob/main/docs/source/en/tutorials/autopipeline.m…
-
I am trying to finetune Qwen-2.5 Coder-7B-Instruct on my custom dataset but am getting the following error:
``
ValueError: Unsloth: Untrained tokens of [[]] found, but embed_tokens & lm_head not t…
-
Trace -
` File "/home/ec2-user/SageMaker/mistral-finetune-unsloth/multi-run-compare/run_model_qwen.py", line 1, in
from unsloth import FastLanguageModel
File "/home/ec2-user/anaconda3/en…
-
Getting the following error when using gradient checkpointing with PEFT LoRA training.
> NotImplementedError
> self.get_input_embeddings()
```
Traceback (most recent call last):
File "/home…
-
I am trying to run single GPU to multinode distributed fine tuning for Llama3-70B and Llama3 8B Models.
Below is my training configuration:
SFT (Llama3 8B & 70B)
Epochs: 3
Gradient Accumulatio…