Sejong-VLI / V2T-Action-Graph-JKSUCIS-2023

The implementation of a paper entitled "Action Knowledge for Video Captioning with Graph Neural Networks" (JKSUCIS 2023).
MIT License
14 stars 3 forks source link

CLIP model #4

Open Mornnn opened 9 months ago

Mornnn commented 9 months ago

Hello, thank you very much for your team's excellent work. When I reproduced your code, I obtained metrics lower than those in your paper. One possible reason is that the CLIP features I used may not be the best. I would like to ask, at which iteration was the best CLIP model you used trained?