DAMO-NLP-SG / Video-LLaMA

[EMNLP 2023 Demo] Video-LLaMA: An Instruction-tuned Audio-Visual Language Model for Video Understanding
BSD 3-Clause "New" or "Revised" License
2.77k stars 255 forks source link

modelling_llama.py #166

Open zeroQiaoba opened 3 months ago

zeroQiaoba commented 3 months ago

In Video-LLaMA, we notice that you load LlamaForCausalLM from ./models/modelling_llama.py. I wonder why not directly load it by "from transformers import LlamaForCausalLM". Do you make any change of the original code in Transforms packages?

plischwe commented 2 months ago

I was wondering this same thing, any clarification here?