-
Thanks for sharing your excellent work as open source. I have a question about your code. I couldn't find the specific parameters for the temporal span **I1** and **I2** that you mentioned in your pap…
-
# Description
Current challenges in using Neural Operators are: irregular meshes, multiple inputs, multiple inputs on different meshes, or multi-scale problems. [1] The Attention mechanism is promi…
-
I want to add attention mechanism in the maddpg network, please tell me which .py file to modify? This question has been bothering me for a long time and I would appreciate it if you could solve the…
-
## 🚀 Feature
Currently, nn.Transformer and related modules return only outputs. I suggest returning attention weights as well.
## Motivation
For all purposes -- demos, tutorials, and practica…
-
I think there is a mechanism where you
- use the server in Docker
- install plugins with dependencies that the server already has, so they are not installed
- pull a new server version that no long…
-
Does the author have a cfg file with attention mechanism Thanks
-
Thank you very much for your great work !
I encountered a problem while reading the source code: what is the role of num_tokens?
I found the `num_tokens` parameter in the source code of `IPAttnPr…
-
Hi,
I'd like to know how did you visualize the 2D and 3D heatmaps in "Figure.8 Motion-word cross-attention visualization" in your paper.
The attention matrix in [CrossAttention module](https://githu…
-
Hi,
First of all, great work. I am big proponent of FLan-t5 and use it in my projects. For multilingual, mT5 and bigscience/mt0 models provide a good baseline and are truly multilingual. Does Flash…
-
Hi, is there any update on implementing the Generative attention masking?
Could you please also provide some explanation in https://github.com/bowang-lab/scGPT/blob/dev-temp/examples/pretrain.py reg…