Closed Steven-m2ai closed 2 years ago
Hi @Steven-m2ai ,
The only place, where this matrix is used is here. You can somehow check that the 3d points are mapped to the correspornding image pixels here.
I probably have never tried browse_dataset
. As our visualization script is correct, you can use it to visualize your ground truth boxes instead of our predicted.
Can you share your Dockerfile? I think something like this should be fine.
FROM pytorch/pytorch:1.7.0-cuda11.0-cudnn8-devel
...
RUN pip install mmcv-full==1.2.7 -f https://download.openmmlab.com/mmcv/dist/cu110/torch1.7.0/index.html
...
Hello @filaPro Thank you for your timely response. Appreciate that.
python tools/test.py {CONFIG} {CHECKPOINT} --eval 'mAP' --options 'show=True' 'out_dir={OUT_DIR}'
I run this but the results in my output folder do not have ground truth boxes in the images. Are you referring to a different script to run visualizations? I refer to here, which notes that GT visualizations can be seen through this command.Thanks again!
readme.md
?Hello @filaPro
Yes, I ran this command initially, python tools/test.py ./configs/imvoxelnet/imvoxelnet_sunrgbd_fast.py ./checkpoint/20211007_105255.pth --show --show-dir ./vis_results
but to no avail. The results inside the "vis_results" directory I created look like predictions only (as shown below)
Yes there are only predictions for now, but you can probably visualize ground truth boxes in the same coordinate system with the same function.
Are you training on 8 GPUs?
Yes, I looked into the code and modified the test script with a show_gt flag. Looks like the same function can produce ground truth well.
I am training on 2 x RTX3090s. Do you mean that the number of GPUs affect the mAP? I read in your paper you use 8 Nvidia Tesla P40s.
Hello,
Thank you for your work. I have a few questions I wish to have clarified. Context: I am creating a dataset in SUN-RGBD format, and so I would like to understand the format structure.
Looks like the "calib" file (once you run the matlab files in SUN-RGBD folder) contains two rows. The first, is the camera extrinsic. However, it is named "Rt" which in my mind should be a 3x4 matrix, but it is stored as a column-major 3x3 matrix. Which coordinates system does this extrinsic parameter transform? From what I understand it rotates from depth coordinate system to camera coordinate system. Then in the ground truth labeling the translation and yaw angle will take care of the bounding box position and orientation. Is this understanding correct?
In MMDetection3D there is a "browse_dataset" file that allows you to view your ground truths of your dataset to confirm it is correct before training. I was wondering if there is one for the SUN-RGBD in ImVoxelNet, as it would be helpful to see if my custom labels in SUN-RGBD format is correct.
I am trying to use the Dockerfile provided, however my machine runs CUDA version 11.1 (RTX 3090 so from my understanding i cannot downgrade to 10.1), which means pytorch>=1.8.0. I change the mmcv-full and mmdet to compatible, most recent versions, but i run into Runtime error "... is not complied with GPU support". Any suggestions here to make the Dockerfile compatible with cuda 11.1? (Running with provided dockerfile gives "CUDA error: no kernel image is available for execution on the device")
Again, thank you for your time!