berlino / gated_linear_attention

MIT License
95 stars 2 forks source link

Using useage is an error encountered #5

Closed JL-er closed 8 months ago

JL-er commented 8 months ago

from gla import GatedLinearAttention import torch d_model = 1024 num_head = 4 use_gk = True # alpha use_gv = False # beta device = "cuda:0"

gla_layer = GatedLinearAttention(d_model, num_head, use_gk, use_gv).to(device)

bsz, seq_len, d_model = 32, 2048, 1024 x = torch.randn(bsz, seq_len, d_model).to(device) y = gla_layer(x)

assert y.shape == x.shape

UVNM{~AJSYB(OQA CFLS1VB

sustcsonglin commented 8 months ago

Hello, which triton version are you using? Can you delete the triton cache file? which should be located in ~/.triton

JL-er commented 8 months ago

Hello, which triton version are you using? Can you delete the triton cache file? which should be located in ~/.triton

Thanks, the problem was solved after I updated the version