Closed SeaBird-Go closed 6 months ago
It's strange. Do you use slurm for distributed training and testing? The script is used for slurm-based testing. May also need to wait for other feedback from the community.
Thanks for your reply. I guess I have found the reasons. The reason is that the running time is so long that it seems to be stuck. When I changed the logging interval to 1, I noticed that the data_time is almost 40 seconds. So, why is it so slow to load the dataset?
Also, I have identified that the main time-consuming part is in the MultiViewPipeline
pipeline, it needs 30
s.
The batch size is 12, and the num_workers is 8.
@SeaBird-Go In MultiViewPipeline
, depth images will be converted to a 3D point cloud in global coordinates, which consumes a lot of time. To speed up this process, you can calculate the point cloud corresponding to each depth image in advance and store it in a file.
I got it. Thanks a lot. By the way, I also want to know how long did you run evaluation on the testing set, and submit a baseline result into the challenge benchmark?
On my side, the evaluation in the testing set would need 4 days, it's much longer than the time in your provided log files. And I also want to know whether can I use less number of images to speed up when testing, the default is 50.
It typically takes several hours, so your case is abnormal. I am not sure whether it is related to your machines (such as CPU and GPU settings or other programs occupying a lot of related resources). For the data_time, you can refer to issue #39, your case is obviously much slower.
For the number of images used for inference, you can refer to Fig. 7-(b) in the appendix of our paper. The detection performance would significantly increase until more than 40 images were used. I am unsure about the curve for multi-view grounding experiments, but you can try to ablate this issue similarly. I would recommend reducing the number of points to shorten the data time for the Matterport3D part.
Prerequisite
Task
I'm using the official example scripts/configs for the officially supported tasks/models/datasets.
Branch
main branch https://github.com/open-mmlab/mmdetection3d
Environment
Reproduces the problem - code sample
Reproduces the problem - command or script
It would cause error by using your provided command
python tools/test.py configs/detection/mv-det3d_8xb4_embodiedscan-3d-284class-9dof.py work_dirs/mv-3ddet/epoch_12.pth --launcher="pytorch"
.So I use the following command to run testing:
Reproduces the problem - error message
It's always stucked with the following output.
Additional information
I want to test the official provided baseline results.