Open milad1378yz opened 2 years ago
Hi, I have a sizeable pre-trained model and I want to get inference on multiple GPU from it(I don't want to train it).so is there any way for that? In fact, I want model-parallelism. and if there is a way, how is it done?
Hi, I have a sizeable pre-trained model and I want to get inference on multiple GPU from it(I don't want to train it).so is there any way for that? In fact, I want model-parallelism. and if there is a way, how is it done?