Closed qgq99 closed 4 months ago
Thank you for your comments! I've just added a detailed description of data preparation on README.md. Frames without camera poses or instance masks are excluded during data preparation. Moreover, target frames without at least 16 source frames are excluded during frame sampling. As you mentioned, the number of frames that can be used for our auto-labeling is less than that in the KITTI-360 dataset for the above reasons. However, our code requires the whole KITTI-360 dataset. If you still have any questions, feel free to make comments on this thread!
Thank you for your comments! I've just added a detailed description of data preparation on README.md. Frames without camera poses or instance masks are excluded during data preparation. Moreover, target frames without at least 16 source frames are excluded during frame sampling. As you mentioned, the number of frames that can be used for our auto-labeling is less than that in the KITTI-360 dataset for the above reasons. However, our code requires the whole KITTI-360 dataset. If you still have any questions, feel free to make comments on this thread!
Thank you very muchοΌ I have finished the preparation of the dataset! Now, I got some new questions, which is:
π€
Firstly, very appreciate for your excellent work! I am kind of confused about the preparation of the dataset, as it 's said in the paper:
So the total number of images for experiments is 47599(less than 50000), but the total number of images (winch in the directory "data_2d_raw", about 150000) of the whole KITTI-360 dataset is much more than 50000.
Since the size of this dataset is too huge, I got some difficulty to train the model on our limited device resource, so I wonder whether it is right to download the whole KITTI-360 dataset, if not, could you please provide some more exact instructions of the dataset preparation?
I will appreciate it very very much! ππ