Closed avishniakov closed 1 month ago
[!IMPORTANT]
Review skipped
Auto reviews are disabled on this repository.
Please check the settings in the CodeRabbit UI or the
.coderabbit.yaml
file in this repository. To trigger a single review, invoke the@coderabbitai review
command.You can disable this status message by setting the
reviews.review_status
tofalse
in the CodeRabbit configuration file.
Thank you for using CodeRabbit. We offer it for free to the OSS community and would appreciate your support in helping us grow. If you find it useful, would you consider giving us a shout-out on your favorite social media?
LLM Finetuning template updates in examples/llm_finetuning
have been pushed.
@strickvl , I can see that the docs for the finetuning-llms
are not quite ready. Shall I wait for you or how to proceed best to document these changes?
@strickvl , I can see that the docs for the
finetuning-llms
are not quite ready. Shall I wait for you or how to proceed best to document these changes?
Could you maybe make changes directly on the feature/gro-1047-docs
branch directly @avishniakov and I'll touch them up. You can see I added a short section at the bottom of docs/book/how-to/training-with-gpus/training-with-gpus.md
page. Or just send something to me in Discord and I'll add it. I couldn't find/think of the good way to document this with code examples, but probably you're in a better place to do so.
LLM Finetuning template updates in examples/llm_finetuning
have been pushed.
Quickstart template updates in examples/quickstart
have been pushed.
E2E template updates in examples/e2e
have been pushed.
NLP template updates in examples/e2e_nlp
have been pushed.
@bcdurak, as discussed this one is ready to be merged, but can give more work to the pydantic merge. LMK, when can I merge it?
Describe changes
Key changes
zenml.integrations.huggingface.steps.run_with_accelerate
to be able to run any step using Accelerate (obviously, the step should be written in a way that it makes sense to run it with Accelerate). This function is supported by a utility, which wraps any function into a Click CLI script (which is needed by most of the Distributed train tools). CLI is quite limited, so it accepts only(str, int, float, bool, Path, tuple)
as params. This can be maybe extended in the future. How it works?@pipeline def llm_peft_full_finetune(): ... ft_model_dir = run_with_accelerate(finetune)( base_model_id=base_model_id, dataset_dir=datasets_dir, load_in_8bit=load_in_8bit, ) ...
cleanup_gpu_memory
to clean up GPU memory at the step start. It has a side effect on the environment, so there is a mustforce
param to be passed to protect that.Companion PR: https://github.com/zenml-io/zenml-projects/pull/102
Sample run:
runs/5080958e-31f5-46b0-8e0b-fbf7086f1ff4
in DemoPre-requisites
Please ensure you have done the following:
develop
and the open PR is targetingdevelop
. If your branch wasn't based on develop read Contribution guide on rebasing branch to develop.Types of changes