issues
search
lucidrains
/
FLASH-pytorch
Implementation of the Transformer variant proposed in "Transformer Quality in Linear Time"
MIT License
342
stars
24
forks
source link
issues
Newest
Newest
Most commented
Recently updated
Oldest
Least commented
Least recently updated
I would like to ask if your model can be applied to other text classification tasks?
#14
ZoeLct
opened
2 months ago
0
About the "/n"
#13
kj01239876
closed
6 months ago
0
AttributeError: module 'torch' has no attribute 'special'
#12
bibo-msft
closed
10 months ago
2
The speed.
#11
wangyuxin87
opened
1 year ago
0
Is it a typo in FLASH module?
#10
marsggbo
closed
1 year ago
1
rel_pos_bias in GAU
#9
SunderlandAJ-1130
opened
1 year ago
1
About negative values in my input sentence embeddings
#8
justinwoo97
opened
1 year ago
0
Laplace Activation Function Implementation
#7
boweny-cerebras
closed
1 year ago
1
Speed on TPU
#6
magicknight
closed
1 year ago
1
About the "shift_tokens"
#5
kangzhao2
opened
1 year ago
2
Cross-Attention?
#4
amorehead
opened
2 years ago
2
minor change to align with paper for better readliness
#3
chivee
opened
2 years ago
0
einsum operation in Linear Attention Part
#2
ShomyLiu
closed
2 years ago
5
mask error
#1
keyunluo
closed
2 years ago
6