mks0601 / V2V-PoseNet_RELEASE

Official Torch7 implementation of "V2V-PoseNet: Voxel-to-Voxel Prediction Network for Accurate 3D Hand and Human Pose Estimation from a Single Depth Map", CVPR 2018
https://arxiv.org/abs/1711.07399
MIT License
377 stars 69 forks source link

video of real-time V2V-PoseNet hand tracking #64

Open pythonsql21 opened 2 years ago

pythonsql21 commented 2 years ago

Hi, do you have a demo (video) of real-time V2V-PoseNet hand tracking? Thx.

mks0601 commented 2 years ago

https://drive.google.com/file/d/1KorXzUlQG7rhZ3oJPk28dyALZ528nq-u/view?usp=sharing https://drive.google.com/file/d/1dMmFU4TpLgY88nLaN8uN-jEQsVzJIXKP/view?usp=sharing https://drive.google.com/file/d/1jaugcdhIgThpo-aYueXSstNKqpTNy52x/view?usp=sharing

pythonsql21 commented 2 years ago

Hi @mks0601 , thanks a lot for the demo video. One more question, in the demo video what type of camera did you use for real-time tracking? RGB camera or RGB-D camera.

mks0601 commented 2 years ago

Those are results on ICVL, MSRA, and NYU datasets, which provide depth maps. My method takes a depth map, so one of RGB-D camera or depth camera is used.

pythonsql21 commented 2 years ago

Hi @mks0601, for the input images for training the V2V-PoseNet (as shown in Fig. 3 of your paper), what sofware did you use to compute 3D voxilized depth map (3D image)? do you have the 3D voxilized depth map (3D images) in your dataset?

mks0601 commented 2 years ago

https://github.com/mks0601/V2V-PoseNet_RELEASE/blob/b91a69f3031869a192ccaf6b7339a3e4c5fb7eb7/src/util.lua#L92

This function generates a 3D voxelized depth map from a 2D depth map.

pythonsql21 commented 2 years ago

Hi @mks0601, what program did you use to visualize 3D voxel? Did you use matplotlib?thx