Closed terryyz closed 1 year ago
I would try the following things:
Hi @antoyang, Thanks for pointing out! Please let me know if you can get expected results via the non-distribution way. Not too sure if this is due to my environment issue or some bugs in the implementation... Cheers
I have just verified that the performance of a pretrained checkpoint is the same without distributing. Some notes:
Hi,
I tried to evaluate the fine-tuned checkpoints provided in the repo. My environment has been correctly configured and I followed all steps up to Zero-shot VideoQA section. As I only have one GPU, I didn't use distributed inference. Here is what I used to run the evaluation:
python videoqa.py --test --eval --combine_datasets <dataset> --combine_datasets_val <dataset> --save_dir=zs<dataset> --ds_factor_ff=8 --ds_factor_attn=8 --suffix="." --batch_size_val=32 --max_tokens=256 --load=checkpoints/frozenbilm_<dataset>.pth --<dataset>_vocab_path <data_folder>/vocab1000.json
I tried with ActivityNet-VQA and iVQA and couldn't get any expected results. For instance, here is what got by testing on ActivityNet-VQA:And results on iVQA: number of params: 29735424
Do you have any ideas on this issue?
Cheers