turboderp / exllamav2

A fast inference library for running LLMs locally on modern consumer-class GPUs
MIT License
3.69k stars 283 forks source link

[BUG] exllamav2-0.2.2+cu118.torch2.4.0-cp310-cp310-win_amd64.whl Version seems missing under releases. #633

Closed Nrgte closed 1 month ago

Nrgte commented 2 months ago

OS

Windows

GPU Library

CUDA 11.8

Python version

3.10

Pytorch version

2.4.0

Model

No response

Describe the bug

whl for windows 3.10 is missing.

Reproduction steps

-

Expected behavior

-

Logs

No response

Additional context

No response

Acknowledgements

turboderp commented 1 month ago

This is added now and should be compiled for 0.2.3+