-
**Describe the bug**
I have created a model in which I have used the annotate layer API to quantize some of the layers. The weights/model are saved in h5 format. I want the model and weights to be …
anidh updated
3 years ago
-
TorchApprox has been extended to use PyTorch's internal affine quantization: https://github.com/etrommer/torch-approx/tree/feature/torchquant
It needs to be benchmarked in order to assess whether t…
-
I didn't find anything except CTC
[recognition.py ](https://github.com/JaidedAI/EasyOCR/blob/master/easyocr/recognition.py)
method get_recognizer get_recouncer method is not implemented for Attn
…
-
I'm quantizing the [MI-GAN](https://github.com/Picsart-AI-Research/MI-GAN/) network that I have previously converted to tflite successfully with your help (thank-you!)
I'm basing my conversion off …
-
Review major architectures and techniques that are hardware friendly:
- quantization
- binarization (full and partial)
- XNOR networks
**DOD**
- [x] chose network to implement on Zybo-Z7-20
-
I've trained an OCR model on a specialized dataset by following the methodology outlined in the README of the **deep-text-recognition-benchmark** repository. My setup includes the model's architecture…
-
# Open Grant Proposal: PrivyML as a ZKML solution on the Swan network, aims to facilitate the implementation and execution of the ZKML concept on the Swan network.
**Project Name:**`PrivyML`
**P…
-
Question for you guys: as best I can tell, there is no support at present for keeping activations in fp8 between the "output" matmul (of either an attention block or MLP block) and the next norm (laye…
-
Hi,
Has anyone tried deploying a low-precision quantized network (int4, int5, etc.) on NVDLA?
If so, please let me know the steps and if you are able to successfully generate the calibration tab…
-
I am using TensorRT 5 and trying to add the code for Int8 Quantization. I tried adding the following lines in baseEngine.cpp but it is giving me an error.
builder->setInt8Mode(true);
IInt8Calibrat…