Open BigShuiTai opened 2 months ago
Hey @BigShuiTai! Thanks for opening an issue.
I'm guessing that you want to split a single version of the model across multiple GPUs, possibly to be able to run the model on GPU with less memory.
Unfortunately, this is not supported by the version of Aurora in this repository. Aurora, however, is just a plain PyTorch model, so model parallelism (I believe the kind that you're referring to) would be possible to implement.
yes - lets make this happen!
@wesselb is this something you are looking into in a near-future? Was this not implemented for the original training in the paper (i.e., was Aurora simply trained with Data parallelism + activation checkpointing for the Swin3D backbone)?
Hello, is there any way to run a inference with 2 or more GPUs?