-
I am doing quantization aware training for Mobilenetv2 using TF Slim (TF - 1.15).
For the full model (without quantization) I get an accuracy of 80% with 1200 steps while the quantization aware tra…
kg512 updated
3 years ago
-
Hi,
Trying TTQ on RESNET18 but getting a runtime error. Can't seem to find what the issue is:
/home/user2/Desktop/pttq/resnet_caltech/trained-ternary-quantization-master/utils/training.pyc in t…
-
Hello,
I am having problems with compiling a U-Net tensorflow 2.2 model for the coral tpu using post-training int8 quantization on windows 10 (wsl with Ubuntu 18.04 for compiler). According to the …
-
**System information**
- Have I written custom code (as opposed to using a stock example script provided in TensorFlow):
No
- OS Platform and Distribution (e.g., Linux Ubuntu 16.04):
Linux Ubuntu …
-
### What is the issue?
I downloaded llama3.1 8b quantized to 8 bits from huggingface. It appears to have a default context size of 132k. Looking at numerous sources on the internet it seemed reasona…
-
Hello,
I would like to train my model in a QAT scenario.
But from what I understand, during QAT, only the Forward pass calculations are done in quantized mode, whereas the weights that are saved are…
-
### Search before asking
- [X] I have searched the YOLOv8 [issues](https://github.com/ultralytics/ultralytics/issues) and [discussions](https://github.com/ultralytics/ultralytics/discussions) and fou…
-
## Prerequisites
Please make sure to check off these prerequisites before submitting a bug report.
- [+] Test that the bug appears on the current version of the dev-branch. Make sure to include the …
-
From the tutorials and recipes it looks like you can only do dynamic Int8 Int4? Also I cannot export the trained model to onnx?
```
import torch
from torchao.quantization.prototype.qat import I…
-
mlc-ai-nightly-cu122 0.15.dev404
mlc-llm-nightly-cu122 0.1.dev1355
transformers 4.41.2
git clone https://huggingface.co/THUDM/glm-4-9b-chat
mlc_llm convert_we…