-
您好!请问我在运行save_attention.py该文件时,在line84行出现“RuntimeError: requested resize to 13x20 (260 elements in total), but the given tensor has a size of 512 (512 elements). autograd's resize can only change the …
-
How to visualize the attention diagram? Is there any relevant code?
-
hello, i run the pip install flash-attn, but it fails. the gpu i used is tesla m40, can the flash attn be installed on m40?
-
This comes up in NeMo / NeVA:
https://github.com/NVIDIA/NeMo/blob/32503fd946cedc41152152837c01f95ae4bc6dc6/nemo/collections/nlp/modules/common/megatron/attention.py#L973-L973
cc @tfogal
-
### Your current environment
code review
### 🐛 Describe the bug
flash_attn.py
forward func
code:
else:
# prefix-enabled attention
assert prefill_m…
-
We have a ShoutOuts channel in Slack, where contributors give kudos to others who have done things to make the project better. While we used to read out the kudos at the Community Meeting, that does…
-
### 是否已有关于该错误的issue或讨论? | Is there an existing issue / discussion for this?
- [X] 我已经搜索过已有的issues和讨论 | I have searched the existing issues / discussions
### 该问题是否在FAQ中有解答? | Is there an existing ans…
-
Have not from transformers.modeling_attn_mask_utils import **_prepare_4d_causal_attention_mask_for_sdpa**
flash_decoding_chunkllama.py:
- Ln 510: attention_mask = _prepare_4d_causal_attention_mas…
-
Hi @fkodom,
I really like your implementation and I wanted to use dilated attention into a vanilla transformer model to try how things work.
Right now, I am facing a problem during the attenti…
-
The crashes are frequent and intermittent, sometimes only steam is open, no game at all, and it crashes.
terminal output
```
[gamescope] [Info] console: gamescope version undefined
ATTENTION:…