Closed Luciennnnnnn closed 2 months ago
16GB*3
16GB*3
where *3 comes from?
16GB*3
where *3 comes from?
I deployed the FP16 model on my 3*3090 server, nvidia-smi like:
Would a single A100 with 40G vram be ok?
work,but only few conversation and it will overflow , more than 40G
Feature request / 功能建议
How many GPU memory is required to inference CogVLM2 in float16
Motivation / 动机
no
Your contribution / 您的贡献
no