-
I have seen the deploy link the Js code is not working in any of the page . Resolve all the contributor those who are intreseted and can resolve mention your that you can easily handle without creatin…
-
**The person who made this package is called "BestBurning" and is a complete idiot because he simply abandoned the package and no longer maintains it.
Google, on the other hand, does not take any act…
-
Your work is important to me, can you provide the textual attention assessment code? Thank you very much for inspiring me!
-
The software doesn't implement attention masking correctly. Simply zeroing out the text embedding does _not_ result in stopping the model from using registers, but does result in skewing the input dis…
-
Do you have plans to enable distributed attention methods (Ring Attention, Striped Attention, Tree Attention, etc.) via a FlexAttention-like interface? Most of the functional implementations today are…
-
I tried to imitate your educational coding style hehe
Here's a pure Pytorch implementation of Flash Attention, hope you like it @karpathy
```
def flash_attention(Q, K, V, is_causal=True, BLOCK_S…
ghost updated
2 weeks ago
-
# Table of Contents
1. [Intuition and underlying principles ](#Intuitionandunderlyingprinciples)
2. [Original Paper](#Paperreading)
3. [Implementation details](#Implementation-details)
4. [Fourth …
-
Hi, I am very interested in the way you visualize the attention maps? Is there any code or links to your visualization method? Thanks a lot
## Upvote & Fund
- We're using [Polar.sh](https://polar.s…
-
Great Great work.
I'm very interested in **Omni-Attention**, but I couldn't find it in the code. Could you please provide a hint or some explanation about Omni-Attention?
-
Would be good to have a meeting to discuss the roadmap and see if we can push anything forward or prioritise things. Could anyone with open pet PRs comment here with a link and brief summary to one or…