Open JACOBWHY opened 6 months ago
Hi, thanks for your interest in our work!
Hi, I download the checkpoints and put them in the output directory, then I use the commands to train the three models, but I can't get the same results as in the table. Is there anything else that needs to be changed?
I think the results are mostly in line with what I have. The reported results are the average of running the training process 5 times, and the standard deviation (+/-) calculated from the 5 runs. The standard deviation can sometimes be quite large unfortunately, if you run the training process a few more times and take the average, it should be closer to the numbers I saw.
since I only have a 24G 3090 GPU, when run the vidsegin_teachstd_kd from scratch I set --batch_size=8 --batch_size_unlb=4
For this task the batch size makes a big difference. You can verify this by training a standard regression model on the video dataset with different batch sizes.
Besides, How to calculate the std ? Is it in the log.csv?
The std is calculated by running the training process 5 separate times. Calculating it from larger sample sizes would have been better, but it was too computationally expensive to do so.
Hope this helps
Thank you very much for your reply! I'll try and I wonder if there is any relationship between the Settings of batch_size and batch_size_unlb? For example, the batch_size is twice as many as the batch_size_unlb?
Hi,
I did not investigate specifically the relationship between labeled and unlabeled batch size, if I recall I set the ratio that way mostly due to constraints on GPU memory :( .
My steps:
since I only have a 24G 3090 GPU, when run the vidsegin_teachstd_kd from scratch I set --batch_size=8 --batch_size_unlb=4
Besides, How to calculate the std ? Is it in the log.csv?