-
### Issue type
Documentation Feature Request
### Have you reproduced the bug with TensorFlow Nightly?
Yes
### Source
source
### TensorFlow version
2.10.1
### Custom code
Yes
### OS platform …
-
## Bug Report
Does Tensorflow Serving support XLA compiled SavedModels ? or am I doing something wrong ?
### System information
- **OS Platform and Distribution (e.g., Linux Ubuntu 16.04)**: [D…
-
### Feature description
This is opened as a feature just to keep track of things.
The recently-released ControlNet Omni model may require a specific Diffusers pipeline (or similar approaches) f…
-
### OS Platform and Distribution
Ubuntu 22.04, Android 14
### Compiler version
Build failure with CLANG 9.0.0
### Programming Language and version
C++, Java, Python 3.10
### Installed using virt…
-
### Describe the issue
After quantization, the output ONNX model had faster inference speed and smaller model size, but why are the input and output tensors still float32?
I thought it should be u…
-
### Search before asking
- [X] I have searched the YOLOv8 [issues](https://github.com/ultralytics/ultralytics/issues) and [discussions](https://github.com/ultralytics/ultralytics/discussions) and fou…
Avv22 updated
2 weeks ago
-
I have a few questions about the inference efficiency of deepseek v2
1.
> In order to efficiently deploy DeepSeek-V2 for service, we first convert its parameters into the precision of FP8.
Ar…
-
### Describe the issue
I exported my medium Whisper model correctly. It could run the inference with the correct answer. After that, I optimized my model. I ran the command line: `python -m onnxrunti…
-
### Describe the issue
I have an OCR model with the following architecture ResNet-BiLSTM-CTC
OS environment:
+ cuda:11.6.2
+ python 3.7
+ onnxruntime-gpu==1.14.1
+ torch 1.10.0 cpu
cuda_pro…
-
### 🐛 Describe the bug
Calling `torch.autograd.functional.jacobian` inside inference mode silently returns all zeros. I'm not sure if this is the intended behavior - the documentation states:
'E…