Closed zhangxiao696 closed 8 months ago
I don’t think the mini split is sufficient to train the model, but if you test the pretrained model, it should be fine.
Besides, debug config is only used for testing training flow, which only updates the model with very few steps. The model is far from convergence.
okey. another problem, use my own dataset, but no road map, can this project run?
Yes. Our pre-trained model also supports generation when setting the map to all zeros, but the generated road structure is not correct. For example, you may see cars in bushes.
Training another model without the road map may be better.
Yes. Our pre-trained model also supports generation when setting the map to all zeros, but the generated road structure is not correct. For example, you may see cars in bushes.
Training another model without the road map may be better.
I would like to ask again, use my own dataset, no map, and no annotation results, how can I use bevfusion detect objects in your project, is this method feasible?
If you are asking about using pseudo-label to train MagicDrive, it is possible but we didn’t try it. If not, please specify your question again.
If you are asking about using pseudo-label to train MagicDrive, it is possible but we didn’t try it. If not, please specify your question again.
In fact, without road map and 3D bounding boxes in my dataset, I want test your pre-trained model, I don't know if it's possible. If possible, how can I generate 3D bounding boxes first?
Our method is used to generate camera views by providing boxes, maps, etc. If you want to test our pre-trained model, basically, all you need would be boxes and maps. For example in our paper, we use the boxes and maps from the dataset. We do not investigate how to "generate" boxes or maps.
Yes. Our pre-trained model also supports generation when setting the map to all zeros, but the generated road structure is not correct. For example, you may see cars in bushes.
Training another model without the road map may be better.
Could you please share an example to perform inference generation while setting all maps to zero? An input example for the StableDiffusionBEVControlNetPipeline
class would be really helpful, thanks!
You only need to change one line of code. I will not add the change to the current repo. If it is needed for any application, please consider opening a PR.
You only need to change one line of code. I will not add the change to the current repo. If it is needed for any application, please consider opening a PR.
sorry, I didn't update in a timely manner, I am currently researching. thanks!
Sure, no problem. I closed due to inactivity. If there is the same issue, please feel free to reopen.
Hi,
Thanks for your excellent work! Here, you mentioned that "We do not investigate how to "generate" boxes or maps". If so, how can you perform 3D Object Detection or BEV Segmentation without boxes or maps using generated new perspective images?
Perhaps I misunderstand or miss something. Could you give me some advice?
Thanks for your attention!
Our method is used to generate camera views by providing boxes, maps, etc. If you want to test our pre-trained model, basically, all you need would be boxes and maps. For example in our paper, we use the boxes and maps from the dataset. We do not investigate how to "generate" boxes or maps.
we use the boxes and maps from the dataset
FYI
follow your readme, I can run demo successfully. But when I train and test by nuScenes mini dataset, but the generated images is abnormal. Can you help us see where the problem lies?
use mini nuScenes dataset
python tools/create_data.py nuscenes --root-path ./data/nuscenes \ --out-dir ./data/nuscenes_mmdet3d_2 --extra-tag nuscenes --version v1.0-mini
train model in debug config with 1xV100
accelerate launch --mixed_precision fp16 --gpu_ids all --num_processes 1 tools/train.py +exp=224x400 runner=debug runner.validation_before_run=true --version
test
python tools/test.py resume_from_checkpoint=magicdrive-log/debug/SDv1.5mv-rawbox_2024-02-27_09-57_224x400