Open joeljang opened 2 years ago
@joeljang, thanks for reporting this issue. No, the library does not currently support MoE layers. We have not previously tested that usage.
Do you have future plans to provide this feature?
@joeljang, yes we plan to provide support for converting checkpoints with MoE layers using convert_to_fp32.py
. In the meantime, can you please share stack trace of the errors you encountered? Thanks!
I replaced one of the layers of gpt2 model with a moe layer training with deepspeed_stage_2. However when trying to run convert_to_fp32.py, I run into all sorts of errors. Does the library currently support converting MoE layers trained with deepspeed into fp32 ?