depth pre-training on DDAD15M is more important than COCO pre-training
depth pre-training is more important than detection pre-training on nuscenes dataset
The 3d detection accuracy scales against the amount of depth pre-training data up to 15M
PL requires depth fine-tuning(Eigen clean) on the target domain. The end-to-end approaches can be worked for a new domain.
PL achieves excellent performance on KITTI 3d val, yet it has not been reflected on the test set. It shows the difficulty of generalization on the PL-based method.