Open wuguangbin1230 opened 2 years ago
@wuguangbin1230 Sorry for the delayed response. Regarding 1) For Conv layer, since weight shape is [out_channels, in_channels, kernel_size[0], kernel_size[1]], channel_axis 0 should be selected. For ConvTranspose layer, weight shape is [in_channels, out_channels, kernel_size[0], kernel_size[1]], channel_axis 1 is used. But, this should be determined internally. Can you please use the latest release and let us know if you run into any errors.
Regarding 2) and 3) Could you share minimum script to reproduce?
for me, bug1 solved by above modification, bug2 solved by directly "param.grad = torch.squeeze(grad,0)" without if-else
Hi All, There are three questions for your help! Thank you very much.
`
(2) During QAT with adaround quantization method, there are two bugs listing as follows: (a) bug 1
(b) bug 2
and modefied by (a) debug 1 in _../aimet_torch/quantsim_straight_throughgrad.py `def compute_dloss_by_dx(x, grad, encoding_min, encoding_max, ch_axis=0):
` (b) debug 2 in _../aimet_torch/qc_quantizeop.py
Modified to be:
(3) When I modefied codes with above methods, I trained the CNN model with 3090 GPU. However the training speed is very slow. How can I resovle this bug?