issues
search
ita9naiwa
/
attention-impl
attention implemenation
4
stars
0
forks
source link
issues
Newest
Newest
Most commented
Recently updated
Oldest
Least commented
Least recently updated
Speculative Decoding
#8
ita9naiwa
opened
2 months ago
1
attention softmax 할 때 max를 구해서, exp^(val - max)로 계산하기
#7
ita9naiwa
closed
7 months ago
0
Write blog posting on implementing Attention
#6
ita9naiwa
closed
7 months ago
0
add pagedAttention
#5
ita9naiwa
closed
7 months ago
1
Faster Matmul
#4
ita9naiwa
closed
7 months ago
1
Add FP16 support
#3
ita9naiwa
closed
7 months ago
1
Add mask
#2
ita9naiwa
closed
7 months ago
0
make block reduction faster
#1
ita9naiwa
closed
2 months ago
0