-
I have a cluster with 3 machines:
* Ubuntu Linux 22.04 with 32GB RAM + Quadro RTX 4000
* Ubuntu Linux 22.04 with 64GB RAM + Quadro RTX 5000
* M2 MacOS 14.7 with 32GB of RAM
_I still couldn…
-
SNMPv2-SMI::enterprises.232.6.2.6.7.1.6.0.1 = INTEGER: 12
SNMPv2-SMI::enterprises.232.6.2.6.7.1.6.0.2 = INTEGER: 12
SNMPv2-SMI::enterprises.232.6.2.6.7.1.6.0.3 = INTEGER: 12
SNMPv2-SMI::enterprises…
-
### Describe the bug
Tried to run the THUDM/CogVideoX1.5-5B model using Diffusers from git (20th Nov, approx 8:30am GMT)
The script failed with
```
hidden_states = F.scaled_dot_product_attent…
-
hi, I am trying to implement a streaming WavTokenizer. I set `causal = True` in encoder without other modification, and replace all nn.Conv1d in the decoder with SConv1d. For example, in ` WavTokenize…
-
### What happened?
I am using SD15. When the batch size on "Empty Latent Image" is set to 2, I get a CUDA error with `torch.nn.functional.scaled_dot_product_attention`from attention_sharing.py and …
Lia-C updated
1 month ago
-
Hi, I face different errors under different gpus to run get_emb.py with generated personal dataset.
If I run it based on gtx 1080 ti, the error is:
```
File /gpfs/gibbs/project/zhao/tl688/con…
-
I want to add a easy to medium question on NLP category related to Transformers positional encoding upon the input embeddings. The question will be done using only `tensorflow as tf, tf.cast(), tf.con…
-
### 🚀 The feature, motivation and pitch
I am unable to find the clean implementation of local multi-headed self-attention in pytorch geometric. I found three types of multi-head attention, one Transf…
-
老师您好,您的CNN LSTM self attention写得非常棒。我想请问下这个 self attention的模块可以用来进行timeseries的每一个时间步的预测吗?比如我的X_train.shape = [8252, 1000, 1],我想对每个样本中的1000个时间步求出一个预测概率,相当于概率>0.5为1类,
-
### 🐛 Describe the bug
```
import torch
from triton.testing import do_bench
from torch.nn.attention.flex_attention import create_block_mask, flex_attention, noop_mask, BlockMask
import torch.nn.f…