openvinotoolkit / nncf

Neural Network Compression Framework for enhanced OpenVINO™ inference
Apache License 2.0
910 stars 227 forks source link

Synchronize quantizer setups for DistributedDataParallel cases #431

Open vshampor opened 3 years ago

vshampor commented 3 years ago

Now that the quantizer setup is being decided during create_compressed_model, and for precision init cases the resulting setup is dependent on the data loaders used for initialization, there is a possibility for DDP that each process may receive significantly different data values, and then compute a different quantizer setup each; since the entire quantizer setup is not technically a torch.Tensor, it cannot be broadcasted to all processes using PyTorch facilities. A special tensor-only synchronization object is required so that the precision init (determining the quantizer setup) only happens in one process of the DDP group, and then the resulting quantizer setup is broadcasted to other processes in the group.

fxmarty commented 1 year ago

Hi @vshampor , was this implemented in nncf?