Closed mioyeah closed 1 year ago
What difficulties did you encounter? I have finished my PhD and I am no longer working at the university of Surrey, so it will be difficult for me to get access to the data.
I encountered an error while running the script
src_video=/home/mio/work/project/PanopticProcessing-master/datasets/Panoptic/171026_pose1/Videos/00.mp4 && dst_video=/home/mio/work/project/PanopticProcessing-master/datasets/Panoptic/171026_pose1/Masks/SOLO/00.mp4 && policy=aggregate && labels= && SOLOv2_ROOT=$WRKSPCE/SOLOv2 && \ git clone https://github.com/GuillaumeRochette/SOLOv2.git $SOLOv2_ROOT && \ cd $SOLOv2_ROOT && \ exec python run_video.py \ --src_video $src_video \ --dst_video $dst_video \ --cfg $CFG_SOLOv2_X101_DCN_3x \ --ckpt $CKPT_SOLOv2_X101_DCN_3x \ --policy $policy \ --labels $labels
Cloning into '/workspace/SOLOv2'...
Matplotlib created a temporary config/cache directory at /tmp/matplotlib-ky3g5ll7 because the default path (/.config/matplotlib) is not a writable directory; it is highly recommended to set the MPLCONFIGDIR environment variable to a writable directory, in particular to speed up the import of Matplotlib and to better support multiprocessing.
0it [00:00, ?it/s]/home/mio/work/project/PanopticProcessing-master/datasets/Panoptic/171026_pose1/Videos/00.mp4
/home/mio/work/project/PanopticProcessing-master/datasets/Panoptic/171026_pose1/Masks/SOLO/00.mp4
/workspace/SOLO/configs/solov2/solov2_x101_dcn_fpn_8gpu_3x.py
/workspace/SOLO/checkpoints/SOLOv2_X101_DCN_3x.pth
/tmp/TEMP_1692355881.2927341
Traceback (most recent call last):
File "run_video.py", line 63, in
Perhaps because the cuda version of the built Dorcker container is 10.1 and my computer graphics card is RTX3090, it seems that cuda 10.1 cannot be used.
Yes indeed that's a problem ... I guess you could either run not pass the GPU in the Docker container and run these on CPU, but that would be slow or run an other, maybe more recent, semantic segmentation model on the data, such as SwinV2 or SAM. The idea there was just to segment the human from the background, by extracting binary masks. I don't think the exact same model is required.
I encountered many difficulties in preprocessing the dataset. Can you provide a processed dataset?