-
Great work. Thanks for sharing.
In the paper it is said that the shape of the adjacancy matrix is (n+1)*(n+1), which sould be 22,22 (21 for joints and 1 for food contact). However, in your implementa…
-
### Checklist
- [x] 1. I have searched related issues but cannot get the expected help.
- [ ] 2. The bug has not been fixed in the latest version.
- [ ] 3. Please note that if the bug-related issue y…
-
We recently released [FlexAttention](https://pytorch.org/blog/flexattention/), which automatically generates fused flashattention kernels for a diverse range of attention variants.
For example, the…
-
Ce soir y a le swatt chez toi sale batard ont va te retrouver j'vais te faire bouffer le béton sale pute
-
Attention Markers can be incorrect at times because the prediction is not simulated out the same way it actually occurs in real time. Things like PlanningView and OnDrawSelected tend to find targets b…
-
My device is a 4090 in hopper architecture, consistent with the h100 architecture. But on the homepage it says “Requirements: H100 / H800 GPU, CUDA >= 12.3.”
I would like to know if flash attentio…
-
we aim to get 80%+ of XeTLA
use `python/tutorials/06-fused-attention.py` as the test case.
- #912
- #913
- #914
- #915
- #916
- #917
- #1102
- #1103
- #1192
(batch head n_ctx d…
-
### Request description
E2E tests suite for the Attention that has reference implementation in it.
### What component(s) does this issue relate to?
Compiler
### Additional context
I ra…
-
The attention mask is not set and cannot be inferred from input because pad token is same as eos token. As a consequence, you may observe unexpected behavior. Please pass your input's `attention_mask`…
-
Hi @cubiq,
Since the `SD3 Attention Seeker L/G` node adjusts Clip L and Clip G, does that mean it could also work with SDXL?
I tried it and it does something, but I don't know if it's working p…