DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
33.6k
stars
3.94k
forks
source link
Inference with the MoE based GPT model trained by ds_pretrain_gpt_345M_MoE128.sh [BUG] #5647
Open
haoranlll opened 2 weeks ago
The existing scripts failed to inference with the MoE based GPT model. The MoE policy provided in the source code https://github.com/microsoft/DeepSpeed/blob/master/deepspeed/module_inject/containers/megatron_gpt_moe.py can't handle the MoE model at all. Is there a quick way to fix this problem?