hassony2 / obman_train

[cvpr19] Demo, training and evaluation code for generating dense hand+object reconstructions from single rgb images
https://hassony2.github.io/obman.html
GNU General Public License v3.0
186 stars 27 forks source link

Relation between 3D joint locations and camera space. #5

Closed erezposner closed 5 years ago

erezposner commented 5 years ago

Hi, I would like to understand the relation between the MANO 3d joint and vertices 3d location with the camera space.

Let's assume that I capture an RGB image using a calibrated camera and use "Learning joint reconstruction of hands and manipulated objects" to estimate MANO 3d joints. Are the 3d joints are in normalized camera space?

Would the MANO estimation is oriented towards camera? Thank you

hassony2 commented 5 years ago

Hi @erezposner

MANO 3d joint and vertices are predicted aligned with the camera view, but root centered. This means that:

I hope this answers your question !

Best,

Yana

erezposner commented 5 years ago

Are the scaling and translation estimated within the net? or using closed form solution? could you kindly direct me this part in the code? thank you

hassony2 commented 5 years ago

Sure !

Here are some pointers:

erezposner commented 5 years ago

Thanks! got it. I have another question, more in the context of MANO layer. How can one generate multiple perspectives of the same MANO generated hand? In the sense of beta,thetas.

If I understand correctly, for the same hand viewed from two different perspectives I would have two different theta vectors, Is that correct? If this is the case, How can I determine the theta vector of a hand viewed from another perspective? thank you

hassony2 commented 5 years ago

This is correct, the first 3 parameters of theta are the global axis-angle rotation vector, so this is the part that needs to be modified to generate the vector from a different perspective.

erezposner commented 5 years ago

Got it, Thank you