A fast inference library for running LLMs locally on modern consumer-class GPUs
3.69k
stars
283
forks
source link
[BUG] exllamav2-0.2.2+cu118.torch2.4.0-cp310-cp310-win_amd64.whl Version seems missing under releases. #633
Closed
Nrgte closed 1 month ago
OS
Windows
GPU Library
CUDA 11.8
Python version
3.10
Pytorch version
2.4.0
Model
No response
Describe the bug
whl for windows 3.10 is missing.
Reproduction steps
-
Expected behavior
-
Logs
No response
Additional context
No response
Acknowledgements