tonyctalope / gpu_poor

Calculate token/s & GPU memory requirement for any LLM. Supports llama.cpp/ggml/bnb/QLoRA quantization
https://rahulschand.github.io/gpu_poor/
0 stars 1 forks source link