xiexh20 / VisTracker

Official implementation for the CVPR'23 paper: Visibility Aware Human-Object Interaction Tracking from Single RGB Camera
http://virtualhumans.mpi-inf.mpg.de/VisTracker/
67 stars 2 forks source link

Data preparation #7

Open faithbotbbot opened 2 weeks ago

faithbotbbot commented 2 weeks ago

Hello, thank you for your contribution. I would like to ask you three questions.

  1. When I use the official script to convert videos to RGB images with an FPS of 30, the speed is too slow. Is this normal? (It takes about two days to process all the videos.) And I would like to ask how much storage space the processed data occupies.
  2. Can you list the directory structure of the training data required by Vistracker in detail?
  3. The article does not mention the training time issue. How long does it take to train Vistracker using A100 80G?
xiexh20 commented 5 days ago

Hi, thank you for your interest.

  1. Yes, this is normal. There are in total ~1.2k videos to be processed, so it can take long time. I usually use multiple CPUs in parallel for this kind of jobs.
  2. I basically follows the same structure as in the original BEHAVE dataset. Specifically for training, these data are:
faithbotbbot commented 3 days ago

Thank you so much for your detailed and helpful response! Your guidance is invaluable for my project. Best regards!