Yxxxb / VoCo-LLaMA

VoCo-LLaMA: This repo is the official implementation of "VoCo-LLaMA: Towards Vision Compression with Large Language Models".
https://yxxxb.github.io/VoCo-LLaMA-page/
Apache License 2.0
84 stars 4 forks source link

Could you please provide the code for ablation study with LLaMA-VID #14

Closed YuchenLiu98 closed 1 week ago

YuchenLiu98 commented 3 months ago

Thanks the authors for amazing job and released code. I wonder could you please provide the code for ablation study with LLaMA-VID in Table 3

Yxxxb commented 1 week ago

Thanks for your interest!

The experimental setup in Table 3 is to replace the encoder with EVA-G and train with exactly the same setup as LLaMA-VID, as you can achieve by modifying the build encoder. We will provide the relevant code later.