Vision-CAIR / MiniGPT4-video

Official code for Goldfish model for long video understanding and MiniGPT4-video for short video understanding
https://vision-cair.github.io/Goldfish_website/
BSD 3-Clause "New" or "Revised" License
559 stars 60 forks source link

About configs in 224_v2_llama2_video_stage_3.yaml #19

Open tianguang2525 opened 6 months ago

tianguang2525 commented 6 months ago

1、Why in 224_v2_llama2_video_stage_3.yaml the config is llama_model: "meta-llama/Meta-Llama-3-8B-Instruct".While in stage_2, it is "meta-llama/Llama-2-7b-chat-hf". 2、Why the task is "image_text_pretrain" in stead of "fituning " in stage 3. Thanks.