666DZY666 / micronet

micronet, a model compression and deploy lib. compression: 1、quantization: quantization-aware-training(QAT), High-Bit(>2b)(DoReFa/Quantization and Training of Neural Networks for Efficient Integer-Arithmetic-Only Inference)、Low-Bit(≤2b)/Ternary and Binary(TWN/BNN/XNOR-Net); post-training-quantization(PTQ), 8-bit(tensorrt); 2、 pruning: normal、regular and group convolutional channel pruning; 3、 group convolution structure; 4、batch-normalization fuse for quantization. deploy: tensorrt, fp32/fp16/int8(ptq-calibration)、op-adapt(upsample)、dynamic_shape
MIT License
2.21k stars 478 forks source link

入门小白求问XNORnet #43

Open b-bread opened 4 years ago

b-bread commented 4 years ago

你好,请问下大大,关于代码,我在代码结构中看到quantization,WbWtAb下看到有bnn与xnornet分类,但由于刚入门,分不太清楚,哪个是bnn代码哪个是xnor代码,xnort论文提及是引入特征因子a进行二值量化,而代码中的A特征值是否就是尼?如果是,,我可否取消A的二值,先只做W的二值呢,因为我试过只输入--w 2,--A还是自动为2了

666DZY666 commented 3 years ago

“分不太清楚,哪个是bnn代码哪个是xnor代码”,融在一起了,都在 util_wbwtab.py(https://github.com/666DZY666/Model-Compression-Deploy/blob/master/compression/quantization/WbWtAb/util_wbwtab.py) 里。 “xnort论文提及是引入特征因子”,缩放因子在这里: https://github.com/666DZY666/Model-Compression-Deploy/blob/3959f194033a520d40fca4c2758874681981ea3c/compression/quantization/WbWtAb/util_wbwtab.py#L107 “取消A的二值,先只做W的二值”,--W 2 --A 32,这样设置。

helloyongyang commented 2 years ago

您好,请问在xnornet代码中,看到了alpha,没有看到特征对应的K,请问这个在哪里实现的?