-
Thank you for submitting a good paper. I have three questions regarding gpu
Is there a way to make the code work using multiple GPUs?
What GPU was it trained on?
Did you use a model that …
-
![image](https://github.com/user-attachments/assets/d8ea00ec-7106-4e30-bec3-02273b322218)
Hello, thank you very much for your work. …
-
## Where are we?
Exporting pytorch model for ExecuTorch runtime goes through multiple AoT (Ahead of Time) stages.
At high level there are 3 stages.
1. `exir.capture`: This captures model’s graph …
-
### System Info
```shell
The examples provided do not work correctly, I think there has been updates in the intel neural compressor toolkit, which is now 3.0. and the habana quantization toolkit, and…
-
The project is so cool. Using TensorRT or OpenVINO to optimize the model to a lower precision could increase the performance of the edge inference. BTW is the project accepting any pull requests?
-
****
When running the command `tune run generate ./custom_quantization_generation_config.yaml`, I encountered the following error:
`AttributeError: module 'torchtune.utils' has no attribute 'gen…
-
## ❓ Questions and Help
Hello,
Great paper! kudos!
After reading I was wondering if it is possible to use these quantization methods on trained model using one of huggingface transformers or shal…
-
### Your current environment
PyTorch version: 2.4.1+cu121
Is debug build: False
CUDA used to build PyTorch: 12.1
ROCM used to build PyTorch: N/A
OS: Ubuntu 22.04.3 LTS (x86_64)
GCC version: (U…
-
I have my quantized YoloV8m and I'm trying make some inference but I'm facing some errors...
When I run those lines of code:
"
ov_model = YOLO("YOLO8_quantization/quantization_OpenVino/quantized_re…
-
According to [this Refact blog post](https://refact.ai/blog/2023/self-hosted-15b-code-model/):
> Check out the [docs on self-hosting](https://github.com/smallcloudai/refact-self-hosting) to get you…