Open rfeinman opened 2 months ago
Hi,
sorry for the unclear description of the data. As we want to make our model more stable to the unperfect mv diffusion prediction, we need some argumentation for the conditioned 4 views images (not the exact 0, 90, 180, 270). It is more convenient for rotating the object rather than the 4 cameras for argumentation, we choose to randomly rotate the objects and save the matrices into 'rots.npy'
Best, Weiyu
@wyysf-98 could you explain how the rotation matrix is computed from camera parameters like elevation & azimuth? I see a function called eul2rot
(see link below) but I'm not sure what the input parameter theta
refers to. Is this the correct function to use? If so, what is the meaning of parameter theta
?
Perhaps theta[0]
is elevation, and theta[1]
is azimuth?
Hi @rfeinman , the code you mentioned is for debugging, which is not used during training.
We only load the rots.npy
of the object and transform the camera position like the code linked here.
I will add some comments about the data later and I think training our diffusion model using the single image is also a good idea. Thus, you do not need the camera parameters.
Hi,
Thanks for the great code. I am trying to reproduce your diffusion training experiment, however there is only a single object included in the data sample and I cannot find instructions about how to render/prepare more examples. Can you explain the camera format that you are using in
rots.npy
? I do not see this file included in the officialObjaverse-MIX
dataset release.Thank you!