huggingface / transformers

🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
https://huggingface.co/transformers
Apache License 2.0
135.4k stars 27.09k forks source link

Enhancing Hugging Face Models with Tensor Parallelism for Large-Scale Model Support 🚀 #32470

Open SeungyounShin opened 3 months ago

SeungyounShin commented 3 months ago

Feature request

Description

This feature proposal aims to update Hugging Face's support for tensor parallelism (TP) to accommodate the increasing size and complexity of models such as LLaMA 3.1, Nemotron-4-340B-Instruct, and others, which have surpassed the capabilities of current training frameworks like TRL + DeepSpeed.

Currently, the Hugging Face codebase is outdated concerning these advancements. Although parallelism requires careful customization based on hardware setup, dataset size, sequence length, and model size, implementing TP across many Hugging Face models is crucial.

Proposal

With the introduction of tensor parallelism in PyTorch 2.0, the previous method of creating processes per device and model in the Megatron style is no longer efficient.

Key Changes:

  1. Refactoring Code for TP:

    • Remove the use of kwargs in favor of more straightforward TP implementations, as PyTorch parallel plans do not accommodate kwargs.
    • Refactor PyTorch models to incorporate TP effectively.
  2. Current Limitations:

    • Existing implementations, such as in modeling_llama, are not trainable and are incompatible with torch.compile for inference optimization.
  3. Future Integration:

    • As models scale to large sizes, 8-way Tensor Parallel is becoming standard.
    • This change would enable Accelerate to later support TP + FSDP (Fully Sharded Data Parallel), which many users could benefit from.

Personal Contribution

I have personally developed code that allows LLaMA to run entirely with TP, observing that it handles longer token sequences with less memory than FSDP. However, I have not submitted a pull request due to the need for comprehensive code refactoring.

Call to Action

If Hugging Face acknowledges this need, I am willing to contribute further if there is an overarching plan for abstraction and integration.

Motivation

Motivation

The motivation behind this proposal is to address the limitations and frustrations experienced when using Hugging Face with the current parallelism approaches, especially for large-scale models like LLaMA 3.1 and Nemotron-4-340B-Instruct. As models grow in complexity, existing frameworks struggle to support efficient training and inference.

Current Issues with Existing Solutions:

Implementing tensor parallelism (TP) in Hugging Face models is crucial to keep up with the trend towards larger models and to enhance compatibility with modern optimization techniques like torch.compile.

Your contribution

Contribution

I am willing to contribute to implementing tensor parallelism (TP) within the Hugging Face ecosystem. To facilitate this, I would appreciate guidance on the following aspects:

  1. Integration Approach: Clarification on whether TP should be applied during model initialization, such as with AutoModelForCausalLM, or if it should be managed externally using torchrun.

  2. Automatic Initialization: Decide if the implementation should automatically initialize torch.distributed without requiring explicit commands from users.

With a defined plan or abstraction level, I can work on refactoring the necessary code and submit a pull request to integrate TP effectively. My experience with TP, particularly with LLaMA, has demonstrated its efficiency in handling large models with reduced memory usage compared to current methods.

amyeroberts commented 3 months ago

Hi @SeungyounShin, thanks for opening up this feature request and for writing up such a detailed report!

TP is definitely something we'd like to support. How easy that is will depend a bit on the necessary changes to modeling code as we need to be able to maintain backwards compatibility with our models and certain standards of readability etc. From the description above it sounds do-able!

Would you like to open a PR showing the updates for one model, and we can iterate from there?

cc @ArthurZucker @LysandreJik @muellerzr @SunMarc

SeungyounShin commented 3 months ago

Sounds good. However, as I'm currently busy with other commitments, I'll be progressing slowly. My goal is to wrap the Hugging Face models using PyTorch's core features to ensure compatibility with FSDP and torch.compile, allowing lowering at the compiler level while maintaining backward compatibility. I will start with LLaMA 3.1 first.

Additionally, I plan to allow users to easily apply tensor parallelism with a pre-configured plan. For example:

# Apply pre-configured tensor parallel plan
model = AutoModelForCausalLM.from_pretrained("model name", tensor_parallel='auto')

# Compiler-level optimization (I noticed that this saves almost 10GB for `LLaMA 3.1 8B`)
model = torch.compile(model)

# Backward compatibility (enables full fine-tuning of 16K context for `LLaMA 3.1 8B`)
output = model(prompt_ids)
loss_fn(output, label).backward()

# Generation (Utilizing available GPU)
model.generate(prompts)

# Saving (Priority: low)
## Since the model weight type is DTensor (Distributed), we need an all-gather operation before saving to disk
model.save_pretrained(...)

Both training and generation should work seamlessly with this setup.

cc @amyeroberts

muellerzr commented 3 months ago

Looking forward to it @SeungyounShin !

ArthurZucker commented 3 months ago

btw modeling llama does support compile. Regarding tensor Parallel, a first thought is to potentially store a dictionannary that maps the layer's name with "row" and "column", which will be used on the fly when loading the checkpoint for TP.

Abstracting a little bit to prevent us from having to add code for new models, appart from the mapping!

{ "model.layers.1.mlp.down_proj": "column", ....}

something like that, which could go in the config.json directly

skyshine102 commented 2 months ago

Looking forward to this PR too! TP style training requires seeing the same input data across TP rank while seeing different data in DP rank. Does datasets & accelerator currently support this out of the box?

ArthurZucker commented 2 months ago

BTW there is some work here: https://github.com/huggingface/optimum/pull/2011 . I am not sure it's out of the box, you probably need a custom collator

zhenglongjiepheonix commented 2 months ago

I have been doing stuffs around automatic TP from fx graph modules traced by torch compile in the last three months, see https://github.com/huggingface/optimum/tree/main/optimum/fx/parallelization for details, basically the goal is to start from a transformers model like LlamaForCausalLM, trace it using torch dynamo, then analyze the traced graph to generate a TP plan, then replace the layers to their parallelized counterparts correspondingly. I think it's close to what you have proposed here, contributions are also welcome!

kmehant commented 2 months ago

I am thinking, the use of device mesh has to be adopted and exposed to some level to extend support for 2D (TP + DP/FSDP) and 3D (TP + FSDP + DP) parallelism with control.

Attaching an old discussion thread for similar support - https://github.com/huggingface/transformers/issues/13690#issuecomment-1732633955 https://github.com/huggingface/transformers/issues/10321

ZeroToAGI commented 1 month ago

I am looking forward to using TP in the transformers library. Is this task currently in progress?

ArthurZucker commented 1 month ago

No PRs are linked, we are internally debating as well given tools like nanotron, accelerate, deepspeed etc that work great. Our aim would probably be MP / TP at inference time. Debate is about whether or not we should have this live in transformers! 🤗 feedback from the community is most welcomed!