-
1> python quant_ptq.py
2> error log:
Traceback (most recent call last):
File "quant_ptq.py", line 27, in
import quant_utils as quant
File "/home/work/YOLOv8_PTQ/quant_utils.py", line…
-
https://github.com/intel/neural-compressor/blob/master/docs/source/quantization_weight_only.md#examples
how to set eval_func?
https://github.com/intel/neural-compressor/blob/master/examples/3…
-
### 🐛 Describe the bug
This is my model
```
class ConvNet(nn.Module):
def __init__(self, num_classes=10):
super(ConvNet, self).__init__()
self.layer1 = nn.Sequential(
…
-
A lot of code for tensor subclasses can likely be conslidated together into a base class that other classes can utilize
_get_to_kwargs:
https://github.com/pytorch/ao/blob/main/torchao/dtypes/affin…
-
when I want to pip install pytorch-quantization:
************************************************************************
Collecting pytorch-quantization
Downloading https://pypi.tuna.tsinghua.e…
-
I found a statement which said **3. Better support for vision transformers.** in link https://nvidia.github.io/TensorRT-Model-Optimizer/guides/_onnx_quantization.html.
I'm working on quantizing VIT n…
-
I am trying to run quantization for int4 examples from `examples/3.x_api/pytorch/nlp/huggingface_models/language-modeling/quantization/weight_only` but there is a package missing in the requirement.tx…
-
Status: Draft
Updated: 09/18/2024
# Objective
In this doc we’ll talk about how different optimization techniques are structured in torchao and how to contribute to torchao.
# torchao Stack Ove…
-
Had this idea and discussed briefly with @andrewor14.
Conceptually the current QAT + FSDP looks like this
- sharded FP32 weight -> all-gather in BF16 -> fake quantize
However, we can do low-…
-
### Issue Type
Bug
### Source
pip (mct-nightly)
### MCT Version
PR #1186
### OS Platform and Distribution
Linux Ubuntu 22.04
### Python version
3.10
### Describe the issu…