I've followed ATISS pre-processing code to create boxes.npz. However, the boxes.npz generated by ATISS is inconsistent with the one required in dataset.py. I can see boxes["camera_coords"] and boxes["target_coords"] in KITTI preprocessing scripts (https://github.com/QhelDIV/kitti360_renderer), but have no idea how to define a camera pose distribution in 3D-FRONT, which is required in inference. The class_labels in boxes.npz is a 23-dimentional one-hot embedding instead of 16 in supplementary.
Hi,
Thank you for your great work.
I've followed ATISS pre-processing code to create boxes.npz. However, the boxes.npz generated by ATISS is inconsistent with the one required in dataset.py. I can see boxes["camera_coords"] and boxes["target_coords"] in KITTI preprocessing scripts (https://github.com/QhelDIV/kitti360_renderer), but have no idea how to define a camera pose distribution in 3D-FRONT, which is required in inference. The class_labels in boxes.npz is a 23-dimentional one-hot embedding instead of 16 in supplementary.
ATISS
dataset.py
I also followed (https://github.com/DLR-RM/BlenderProc/tree/main/examples/datasets/front_3d_with_improved_mat) to generate the images. However, the results are a bunch of hdf5 files and the orientation is not set toward a dominant object in the scene. Did you change anything like the camera pose in their scripts? Or Did you do any post-processing?
Is it possible for you to provide a small demo dataset.zip with one scene in the format of image below that can be used directly for inference?
Thank you for your help in advance!