Since some transformer structure is too huge to be loaded on a single GPU. Does mmselfsup support API that load such big models with more than one GPUs?
What is the feature?
An user-friendly api to load different layers of a huge model to several devices. Just like huggingface models' model.parallelize()
What is the problem this feature will solve?
Since some transformer structure is too huge to be loaded on a single GPU. Does mmselfsup support API that load such big models with more than one GPUs?
What is the feature?
An user-friendly api to load different layers of a huge model to several devices. Just like huggingface models' model.parallelize()
What alternatives have you considered?
No response