rese1f / MovieChat

[CVPR 2024] MovieChat: From Dense Token to Sparse Memory for Long Video Understanding
https://rese1f.github.io/MovieChat/
BSD 3-Clause "New" or "Revised" License
534 stars 41 forks source link

Inference fails #19

Closed fazliimam closed 1 year ago

fazliimam commented 1 year ago

When I inference.py, it reaches upto this point and then the process gets killed. Can I know the reason for it? Initializing Chat Loading VIT Loading VIT Done Loading Q-Former Using pad_token, but it is not set yet. Loading checkpoint shards: 0%| | 0/2 [00:00<?, ?it/s]

How much memory is required? I tried with 100gb RAM machine and it still failed

Espere-1119-Song commented 1 year ago

The information displayed indicates that you are loading the checkpoint required by MovieChat. Although it does take a long time to load the checkpoint, I have not encountered a situation where the process was killed. When I execute inference.py, the GPU required is about 16GB.