Open KaiLv69 opened 1 week ago
Hi, when training big model like llama2-70b with lora, it will run into oom due to the unsharded model.
It could help a lot if lora supported with GeminiPlugin or HybridParallelPlugin. Wonder if any plan to support that?
GeminiPlugin
HybridParallelPlugin
Describe the feature
Hi, when training big model like llama2-70b with lora, it will run into oom due to the unsharded model.
It could help a lot if lora supported with
GeminiPlugin
orHybridParallelPlugin
. Wonder if any plan to support that?