-
Platforms: rocm
This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/flakytest?name=test_comprehensive_std_unbiased_cuda_float16&suite=TestInductorOpInfoC…
-
### Feature description
I would like to be able to store and load float16 values in datasets. Many dataset formats support this, and many modern compilers support this as well. I do need to store flo…
-
This is on mac-arm64 running macOS 15.0 24A335 with Xcode 16.0 16A242d. I expect that this would also occur targeting mac-arm64 on macOS 14 as long as Xcode 16 is used.
While bootstrapping gcc-12.4…
-
I got the following error. Is there a way to fix it? My machine runs whisper without problem. So I think whisperx should also be adapted to machines without fp16.
```
Traceback (most recent call l…
-
### 🐛 Describe the bug
Hi everyone, thanks for your effort on this issue.
I'm trying to build executorch on my OrangePi 5 Pro board equipped with an 8 core ARMv8 CPU, But I encountered a compile …
-
Platforms: linux
This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/flakytest?name=test_comprehensive_transpose_copy_cuda_float16&suite=TestInductorOpIn…
-
Platforms: inductor
This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/flakytest?name=test_comprehensive_randint_like_cuda_float16&suite=TestInductorOpI…
-
Hello,
I tried to run a fast tuning of GEMM with float16:
```python
from bitblas.base.roller.policy import TensorCorePolicy, DefaultPolicy
from bitblas.base.arch import CUDA
from bitblas.base.uti…
-
I have posted this issue on AMD's side but no response there...
https://github.com/ROCm/flash-attention/issues/73
As the same issue happens in main branch as well, I thought I'd try here...
Step…
-
Platforms: inductor
This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/flakytest?name=test_comprehensive_special_i0e_cuda_float16&suite=TestInductorOpIn…