jiahaoLjh / PlaneSweepPose

Code for "Multi-View Multi-Person 3D Pose Estimation with Plane Sweep Stereo"
MIT License
94 stars 7 forks source link

Further explanation on the extrinsic parameters used in the campus dataset #11

Open AntonioEscamilla opened 2 years ago

AntonioEscamilla commented 2 years ago

Can you explain how did you obtain the translation parameter (T) in the calibration_campus.json? What special consideration should be made to obtain such values? How should they be interpreted?

From the original calib info in the campus dataset, I got T = [-1.787557e+00, 1.361094e+00, 5.226973e+00] for the cam 0 . But the T you provide in the json is way too different T=[1774.89, -5051.69, 1923.35]. It is the same for all three cameras, I am just mentioning cam0 as an example.

When I used the calib data you provide I was able to obtain 3d poses successfully, but I cannot plot the camera system coherently. When I use the original T parameters, I can plot the camera system but the model inference doesn't work.

I would appreciate if you can elaborate more on it

Thank you for your time!! And nice work, by the way

Shariy commented 2 years ago

Can you explain how did you obtain the translation parameter (T) in the calibration_campus.json? What special consideration should be made to obtain such values? How should they be interpreted?

From the original calib info in the campus dataset, I got T = [-1.787557e+00, 1.361094e+00, 5.226973e+00] for the cam 0 . But the T you provide in the json is way too different T=[1774.89, -5051.69, 1923.35]. It is the same for all three cameras, I am just mentioning cam0 as an example.

When I used the calib data you provide I was able to obtain 3d poses successfully, but I cannot plot the camera system coherently. When I use the original T parameters, I can plot the camera system but the model inference doesn't work.

I would appreciate if you can elaborate more on it

Thank you for your time!! And nice work, by the way

The T value used by the campus dataset is: Not the Translation vector value of the World -> Camera It was found that it was a position value of the camera calculated as -R.inv() * tvec. The R value was calculated based on the contents of the getRT.m file. In the case of the tvec value, the tvec obtained using opencv solvepnp was used.

When the data was entered by modifying the corresponding value, it was confirmed that the model worked.