Open HeChengHui opened 4 days ago
Thank you for the question. Video-SALMONN is trained with vicuna-13b so the input dimension for the 7b model would not match the Q-Former output of video-SALMONN. It would be helpful to try quantization (with a bit of performance loss).
@BriansIDP does that mean i can use something like TheBloke/vicuna-13B-v1.5-16K-AWQ by just setting it in the config?
@BriansIDP Thank you for your work.
what is the vram requirement to run inference? i am having OOM using
lmsys/vicuna-13b
, butlmsys/vicuna-7b
is giving me size mismatch error.Or am i using the wrong model?