-
I am currently conducting research on applying Transformer models to EEG signals. In your paper, I learned that by combining CNN and Transformer, it is possible to learn both global and local features…
-
Thank you very much for your work!
Where can I find more details about your implementation of positional encoding in the model?
-
Hello,
First of all, thank you for the code you developed. I see that some sections in the article are missing. For example, I could not find the Fuzzy Positional Encoding section in the code.
A…
-
Check AttentionStore paper and see if the performance would be good or not.
- AttentionStore: Cost-effective Attention Reuse across Multi-turn Conversations in Large Language Model Serving https://…
-
Does the mamba model need any kind of positional encodings? My understanding based on the code and paper is that no position encoding is needed due to the recurrent nature. However I tried adding posi…
-
hi authors, thanks a lot for releasing the code!
i noticed that position ids for text embeddings is set to zeros at this line: https://github.com/black-forest-labs/flux/blob/c23ae247225daba30fbd56…
-
Hello,
Thank you for your excellent work on this project!
While reviewing the code, I noticed a few discrepancies between the implementation and the manuscript's description, specifically in the…
-
Hi, thanks for the great work and congrats for CVPR acceptance. I saw in the code that ob_dist and ob_view are directly concatenated with anchor feature and input to the MLPs. Would using positional e…
-
Hi, in your PTv3 paper, a variety of positional encoding method are discussed, with CPE being chosen and implemented using SpConv.
Considering hardware constraints, storing grid positions and perfo…
-
I notice that Metaformer has no positional encoding(PE) either in the attention layers or at the model input, does this affect the performance? Is the positional encoding not necessary? What if metafo…