Closed yw155 closed 5 years ago
Hi yw155,
Thanks for reaching out! Regarding the multi-GPU training:
You can use torch.distributed
do perform the distributed training and use the barrier()
function for GPUs 1-7 to wait for GPU 0 to perform quantization as well as the broadcast()
function to broadcast the centroids and assignments obtained by GPU 0 to GPUs 1-7.
Hope this helps,
Pierre
Thank you very much.
@yw155 Hi, did you implement multi-gpu training?
Hi @pierrestock, I would like to ask you how to implement the codes of multi-gnu training. Which parts could be run on multi-gnu, like quantization, fine-tune and global fine-tune? Thank you.