Hi, I am confused about the transform in multiviewvideo.py and volumetric.py
Specifically,
self.campos[cam] = (-np.dot(krt[cam]['extrin'][:3, :3].T, krt[cam]['extrin'][:3, 3])).astype(np.float32) the line calculate the camera position in world space ?
self.camrot[cam] = (krt[cam]['extrin'][:3, :3]).astype(np.float32), this matrix is the camera rot in world space ?
result[k] = to4x4(np.dot( np.linalg.inv(to4x4(frametransf)), to4x4(self.basetransf))[:3, :4]) what is the meaning of this transform?
result[k] = to4x4(np.dot( np.linalg.inv(to4x4(self.basetransf)), to4x4(frametransf))[:3, :4]) what is the meaning of this transform?
result["campos"] = np.dot(self.basetransf[:3, :3].T, self.campos[cam] - self.basetransf[:3, 3]) meaning of this ?
result["camrot"] = np.dot(self.basetransf[:3, :3].T, self.camrot[cam].T).T meaning of this?
viewrot = torch.bmm(camrot, modelmatrixinv[:, :3, :3]) meaning of this?
viewpos = torch.bmm((campos[:, :] - modelmatrixinv[:, :3, 3])[:, None, :], modelmatrixinv[:, :3, :3])[:, 0, :] and meaning of this ?
I want to use the multiface dataset to train a NeRF, which need multi-view and correspondence camera parameters. To get these data, i think it's a good way to understand the coordinates transform of mvp. So i am pleased if you can answer above questions ?
Hi, I am confused about the transform in multiviewvideo.py and volumetric.py Specifically,
self.campos[cam] = (-np.dot(krt[cam]['extrin'][:3, :3].T, krt[cam]['extrin'][:3, 3])).astype(np.float32)
the line calculate the camera position in world space ?self.camrot[cam] = (krt[cam]['extrin'][:3, :3]).astype(np.float32)
, this matrix is the camera rot in world space ?result[k] = to4x4(np.dot( np.linalg.inv(to4x4(frametransf)), to4x4(self.basetransf))[:3, :4])
what is the meaning of this transform?result[k] = to4x4(np.dot( np.linalg.inv(to4x4(self.basetransf)), to4x4(frametransf))[:3, :4])
what is the meaning of this transform?result["campos"] = np.dot(self.basetransf[:3, :3].T, self.campos[cam] - self.basetransf[:3, 3])
meaning of this ?result["camrot"] = np.dot(self.basetransf[:3, :3].T, self.camrot[cam].T).T
meaning of this?viewrot = torch.bmm(camrot, modelmatrixinv[:, :3, :3])
meaning of this?viewpos = torch.bmm((campos[:, :] - modelmatrixinv[:, :3, 3])[:, None, :], modelmatrixinv[:, :3, :3])[:, 0, :]
and meaning of this ?I want to use the multiface dataset to train a NeRF, which need multi-view and correspondence camera parameters. To get these data, i think it's a good way to understand the coordinates transform of mvp. So i am pleased if you can answer above questions ?