issues
search
NetEase-FuXi
/
EETQ
Easy and Efficient Quantization for Transformers
Apache License 2.0
157
stars
12
forks
source link
issues
Newest
Newest
Most commented
Recently updated
Oldest
Least commented
Least recently updated
ImportError: cannot import name 'EetqConfig' from 'transformers', despite using using 4.38.2 which satisfies >=4.27.0
#23
moruga123
opened
1 day ago
1
Repetition with Llama3-70b and EETQ
#22
mjsteele12
opened
4 weeks ago
1
Does it support Vision Transformers?
#21
PaulaDelgado-Santos
opened
1 month ago
1
Create LICENSE
#20
dtlzhuangz
closed
1 month ago
0
Support CPU quantization
#19
xgal
opened
1 month ago
3
License
#18
AlpinDale
closed
1 month ago
1
Qlora with eetq is quite slow
#17
hjh0119
opened
1 month ago
3
FIX: Use `matmul` instead of `mm` in `backward`
#16
younesbelkada
closed
2 months ago
0
PEFT compatible GEMM
#15
dtlzhuangz
closed
2 months ago
0
how to dequant a EETQ model?
#14
mxjmtxrm
closed
2 months ago
4
Integration with Hugging Face transformers library
#13
younesbelkada
closed
1 month ago
2
Supports H100
#12
mwbyeon
opened
3 months ago
1
Modify code to support CUDA Graph
#11
khj94
closed
4 months ago
1
Quantization takes a very long time
#10
timohear
opened
5 months ago
3
[docs] Update readme
#9
SidaZh
closed
5 months ago
0
Add LoRAX to usage options in README.
#8
arnavgarg1
closed
5 months ago
4
rm dist
#7
dtlzhuangz
closed
5 months ago
0
gemv optimization
#6
dtlzhuangz
closed
5 months ago
0
Understanding EETQ and 8 bit quantization
#5
RonanKMcGovern
closed
6 months ago
3
How to handle bfloat16?
#4
vgoklani
closed
6 months ago
7
Why does EETQ take up all VRAM
#3
RonanKMcGovern
closed
6 months ago
2
安装出错ERROR: Could not build wheels for EETQ, which is required to install pyproject.toml-based projects
#2
linshuijin
closed
9 months ago
5
Question on outlier handling
#1
0xymoro
closed
6 months ago
1