-
I've written a function to pack a 2D int4 tensor into a `int32` tensor. I want use `torch.compile` to speed up it, but the compilation process takes over 200 seconds.
It's not a bug, but I wonder …
-
Status: Draft
Updated: 06/17/2024
# Objective
In this doc we’ll talk about Tensor subclass based quantization API for modeling users and developers.
# Modeling User API
Modeling users refer t…
-
### OpenVINO Version
2024.2.0
### Operating System
Ubuntu 20.04 (LTS)
### Device used for inference
GPU
### Framework
PyTorch
### Model used
VGG16
### Issue description
Hiya,
My goal is …
-
### Describe the issue
I am trying to replicate the following : [https://intel.github.io/intel-extension-for-pytorch/llm/llama3/xpu/](url) . While running the `python run_generation_gpu_woq_for_llama…
-
### Description
```python
import jax
import jax.numpy as jnp
def f(x):
return jax.lax.clz(x)
x = jnp.array([0, 1 tensor
"mhlo.return"(%3) : (tensor) -> ()
}) {base_dilations = dense…
-
### Context
OpenVINO component responsible for support of TensorFlow models is called as TensorFlow Frontend (TF FE). TF FE converts a model represented in [TensorFlow opset](https://www.tensorflow.o…
-
### Context
OpenVINO component responsible for support of TensorFlow models is called as TensorFlow Frontend (TF FE). TF FE converts a model represented in [TensorFlow opset](https://www.tensorflow.o…
-
### Context
OpenVINO component responsible for support of TensorFlow models is called as TensorFlow Frontend (TF FE). TF FE converts a model represented in [TensorFlow opset](https://www.tensorflow.o…
-
### Context
OpenVINO component responsible for support of TensorFlow models is called as TensorFlow Frontend (TF FE). TF FE converts a model represented in [TensorFlow opset](https://www.tensorflow.o…
-
### Context
OpenVINO component responsible for support of TensorFlow models is called as TensorFlow Frontend (TF FE). TF FE converts a model represented in [TensorFlow opset](https://www.tensorflow.o…