X-PLUG / mPLUG-Owl

mPLUG-Owl: The Powerful Multi-modal Large Language Model Family
https://www.modelscope.cn/studios/damo/mPLUG-Owl
MIT License
2.25k stars 171 forks source link

The code and detailed implementation of Figure 4 and Figure 5 in the paper mPLUG-Owl2 #197

Open Zlatan-Ibrahi opened 9 months ago

Zlatan-Ibrahi commented 9 months ago

I would like to analyze the attention map of my own trained model, but I am not very clear about some details. For example, do we take the average of the attention maps across multiple heads? Could you provide the code for this?

GasolSun36 commented 6 months ago

same question, any solutions?