-
I using alexnet_bn_wrpn in the imagenet dataset. But the result in epoch 36 is lower than Dorefa or PACT or fp32.
This is my config file:
quantizers:
wrpn_quantizer:
class: WRPNQuantizer
…
-
Bolt supports both XNOR-style and DoReFa-style BNN networks. Just save the binary weights as FP32 in an Onnx model, and X2bolt will automatically convert the storage to 1-bit representations. So far, …
-
I was wondering if there is a way to perform power of 2 quantization with Larq. Maybe a specific quantizer is needed? Any suggestions?
-
I haven’t found the ADQ method related code in the project. Haven’t I uploaded it yet?
-
I do think this repository is an elegant and faithful reimplementation of the DoReFa paper. Belowing are some minor concerns to me:
1. Please try to list some numeracal results and compare them wit…
-
Hello!
I found that without weight normalization, the network ceases to learn, and the loss is equal to nan. Could you please explain why this is happening and how it can be fixed?
-
why there is no extended experiments on LLMs or large vision transformers?
-
* [deep compression](https://arxiv.org/pdf/1605.07678.pdf): pruning and quantization , 35x reduction
* [squeezenet](https://arxiv.org/pdf/1602.07360.pdf): its [openreview](https://openreview.net/foru…
-
大佬你好,请问做8bit量化的时候,转成量化模型quant_model的时候,为什么权值是小数吖,比如是0.4256之类的,为什不是0-255之间的整数呢?还是说我哪儿操作步骤有问题
恳请大佬教做人,刚入门量化的小白,感激不尽!
-
**Describe the issue**:
AssertionError, while speeding up Convnext model.
**Environment**:
- NNI version: 2.10
- Training service (local|remote|pai|aml|etc): remote
- Server OS (for remote m…