-
Hey there!
Is it possible to do post-training quantitzation with Parseq? I'm looking for ways to speed up inference time. I tried training a parseq-tiny model but lost about 13% absolute val accur…
-
## Formal verification
1. [Modular, Compositional, and Executable Formal Semantics for LLVM IR](https://dl.acm.org/doi/pdf/10.1145/3473572)
2. [Alive2: Bounded Translation Validation for LLVM](https…
-
Hii
I am getting a `KeyError` while the quantization process of OpenVINO model
My openVINO model is made of IR of a tensorflow model (pre-trained from tensorflow-hub)
ERROR
```
[/usr/local/lib…
-
### Documentation issue/request
0 useful information on Quantization. How do i perform it, what settings should i choose for different Quantization types Q8, Q5, (and what would be the difference i…
-
This repo. has supported fp16 training and inference for some time.
Running inference at fp16 currently offers the best AP/FPS trade-off but I have found that training at fp16 results in a relative…
-
when trained a model,the accuracy is fairly good,but when convert to pt with torchscript_int8,the Accuracy becomes very poor, I found Quantization aware training ,but now document about it ,can any…
-
## When
- 11.15
## Who :
- 권세중 (sejung.kwon@navercorp.com)
## What
- Title: AlphaTuning: Quantization-Aware Parameter-Efficient Adaptation of Large-Scale Pre-Trained Language Models
- Links: ht…
-
Hi,
I wonder if alpha and delta are layer based (as clained in Trained Ternary Quantization paper) or filter based (like XNOR-Net). I assumed they are filter based because your paper mention n as …
-
### 1. System information
Colab , as of 2023-10-23
### 2. Code
Please see the attached colab notebook here
https://colab.research.google.com/drive/1yUD0nDu8oeeDtQBa7xCbQWx_w8PxS4UC?usp=sharin…
-
Hi, I'm trying to replace incremental BoW with fixed vocabulary trained offline on subset of mapping data. Is it possible to train it correctly offline?
I'm currently using the following code for tra…