Closed imrankh46 closed 1 year ago
If you want to reproduce the full example, then yes. However, we also provide the adapter weights for each step on the Hub: https://huggingface.co/trl-lib so you could reuse them e.g. if you just want to run the last step.
If you want to reproduce the full example, then yes. However, we also provide the adapter weights for each step on the Hub: https://huggingface.co/trl-lib so you could reuse them e.g. if you just want to run the last step.
My question is, should I need to run each steps individually or not . I have my dataset like stack_exchange one . 🤗
Yes, in that case you would need to run each step individually!
Yes, in that case you would need to run each step individually!
Thank you for the response. 2nd question is, should I need to follow instructions base model ? Or I can train small model also.
I don't think I understand your question. Fine-tuning on StackExchange data should be ok in this example. If you already have a strong instruction model maybe fine-tuning step might not be necessary.
If you want to reproduce the full example, then yes. However, we also provide the adapter weights for each step on the Hub: https://huggingface.co/trl-lib so you could reuse them e.g. if you just want to run the last step.
Here you mean that's, if I have a data set so I don't need to follow step one and Step two. Just run the last step.
Iam I correct?
No, I meant that if you want to reproduce the results on our dataset you can use the checkpoints. If you have your own dataset then you need to run all the steps.
No, I meant that if you want to reproduce the results on our dataset you can use the checkpoints. If you have your own dataset then you need to run all the steps.
Thank you so much.
No, I meant that if you want to reproduce the results on our dataset you can use the checkpoints. If you have your own dataset then you need to run all the steps.
i try to run the following code but they not working. i think you need to refresh the model page.
!torchrun --nnodes 1 --nproc_per_node 1 /content/trl/examples/stack_llama/scripts/supervised_finetuning.py --model_path=trl-lib/llama-7b-se-peft --streaming --no_gradient_checkpointing --learning_rate 1e-5 --max_steps 5000 --output_dir ./llama-se
OSError: trl-lib/llama-7b-se-peft does not appear to have a file named
config.json. Checkout 'https://huggingface.co/trl-lib/llama-7b-se-peft/main' for
available files.
ERROR:torch.distributed.elastic.multiprocessing.api:failed (exitcode: 1) local_rank: 0 (pid: 4607) of binary: /usr/bin/python3
Traceback (most recent call last):
File "/usr/local/bin/torchrun", line 8, in <module>
sys.exit(main())
File "/usr/local/lib/python3.9/dist-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py", line 346, in wrapper
return f(*args, **kwargs)
File "/usr/local/lib/python3.9/dist-packages/torch/distributed/run.py", line 794, in main
run(args)
File "/usr/local/lib/python3.9/dist-packages/torch/distributed/run.py", line 785, in run
elastic_launch(
File "/usr/local/lib/python3.9/dist-packages/torch/distributed/launcher/api.py", line 134, in __call__
return launch_agent(self._config, self._entrypoint, list(args))
File "/usr/local/lib/python3.9/dist-packages/torch/distributed/launcher/api.py", line 250, in launch_agent
raise ChildFailedError(
torch.distributed.elastic.multiprocessing.errors.ChildFailedError:
============================================================
/content/trl/examples/stack_llama/scripts/supervised_finetuning.py FAILED
------------------------------------------------------------
Failures:
<NO_OTHER_FAILURES>
------------------------------------------------------------
Root Cause (first observed failure):
[0]:
time : 2023-04-15_12:30:39
host : 5b5507893892
rank : 0 (local_rank: 0)
exitcode : 1 (pid: 4607)
error_file: <N/A>
traceback : To enable traceback see: https://pytorch.org/docs/stable/elastic/errors.html
I think for the supervised finetuning you need the full model, not just the adapters. You could adapt the code to load the pretrained model and load the adapters on top of it.
I think for the supervised finetuning you need the full model, not just the adapters. You could adapt the code to load the pretrained model and load the adapters on top of it.
How I can do this ?
This issue has been automatically marked as stale because it has not had recent activity. If you think this still needs to be addressed please comment on this thread.
Hi, everyone. My question is, the steps mentioned for fine tune Stack_llama model. So should I need to run all the step in once. ?