cuteyyt / or-nerf

OR-NeRF: Object Removing from 3D Scenes Guided by Multiview Segmentation with Neural Radiance Fields
https://ornerf.github.io
44 stars 3 forks source link

some small mistakes while getting the sparse folder #3

Closed Starryskyswt closed 1 year ago

Starryskyswt commented 1 year ago

1690122958354 Here I find a small mistake while warping a video of the images_downfactor. It should be corrected as out_dir while it was in_dir at first.Hopefully you can check it again,and reply to me if I Ignored other issues.

cuteyyt commented 1 year ago

Thanks for your interest in our project. Which file do these codes belong to? I'll check this.

cuteyyt commented 1 year ago

Thanks for your interest in our project. Which file do these codes belong to? I'll check this.

I have fixed this in datasets/pre_sparse.py (commit f72c358ca78c2ffe8aa8075e9db91f8e7bc247e6). Thank you!

Starryskyswt commented 1 year ago

1690197287287 1690197549192

Yes,that's the place where I found the mistake.Now I encounter the same issue as the other one has put forward.I use the nerf_liff_data(room dataset)to run pre_sparse.py and get two folders including image(Sorted images from the original dataset)and image4(Image after downsampling 1/4), and next step is to use LLFF to reconstruct cam params.Becuase I didn't install colmap and I found camera.bin from the nerf_liff_data room dataset. My question is that after sorted the image of the dataset,can I use camera. bin directly? Or in other words,Does the order of the images in the dataset correspond one-to-one to camera. bin?Now,I may directly use sparse folder of the dataset and skip the first step. I just wonder if I change the order of the pictures in the dataset(like what pre_sparese.py does),could I use the raw camera. bin from the dataset?Please do me a favour.

cuteyyt commented 1 year ago

Of course, you can use the camera parameters provided in the original dataset directly. But the file names of different views in the bin file and assigned to the image file may not match neatly. For example, in the room scene from nerf_llff_data, the image name in the images.bin file is _DJI_20200226_143850006.JPG corresponds to the images folder while images_4 uses a png suffix like _DJI_20200226_143850006.png.

To make the pipeline work properly, the best way is to ensure a unique name for each view in every usage. However, some reconstruction schemes used in this project, like NeRF, load camera parameters from file poses_bounds.npy, which omits the view name. In this case, the order determines the correspondence between image files and camera parameters. And as we warp all datasets into the same format, we may also use some sort functions lazily to match cameras and image files, which may be dangerous to use the original camera names.

Please carefully check the correspondence to use the camera parameters provided directly in the datasets. If you have any specific questions, I am willing to help.

Starryskyswt commented 1 year ago

Thanks for your asistance,I have figured out the corresdence between camera parameters and pictures.Becasue the nerf_liff dataset already has contained the necessary sparese folder,I have modified the code in order to directly use the existing sparse folder. Now I encounter another problem when running the run_sam_text.py,but I find the ckpt folder may need to be downloaded somewhere(and your code seemingly doesn't include it.). 1690297093259 Befor I realize this issue,I have run the code twice.Here are the results. bad9b9d09eeaf4dccfc9a087123ad32 First I use the ladder and find the progress stuck,then I close the laddder and try it again.It seems that the progress gets stuck again.My question is that is it necessary for me to downlaod the ckpts somewhere and whether the running results goes well (what I need to do is just wait?) I would appreciate it if you could help me figure out the issue. Additionally,I find the ckpt folder may need to be downloaded somewhere similarly in the run_sam_point.py,I have no idea where to download it.(I have downloaded all your code, including the code in the Prior folder link)

cuteyyt commented 1 year ago

Yes, you need to download some pre-trained checkpoints.

SAM: generate masks for different views with points or box prompts. Grounding-DINO: generate bounding boxes for images with text prompts. LaMa: inpainting 2D images with masks.

You can download pre-trained checkpoints and install the necessary packages with their guidance.

image

You can see from the above picture that scripts using sam ckpts already have specified the ckpts. You can use your path as well. Other paths are similar.

Starryskyswt commented 1 year ago

1690396262485

Thank you for your answer,I have downloaded some pre-trained checkpoints above.But I am not sure it's enough or not,cause I find other checkpoints on the link you gave as follow(eg) 1690396323688 I wonder if I need to download it,and where to place (because I notice here the default path relates to it) 1690396750630 I am confused about it and sincerely expect your reply.

cuteyyt commented 1 year ago

The following files are necessary, and I may miss something. If you have any problems in the future, we can continue this discussion.

ckpts
├── grounding_dino
│   └── groundingdino_swint_ogc.pth
├── lama
│   └── big-lama
│       ├── config.yaml
│       └── models
│           └── best.ckpt
└── sam
    └── sam_vit_h_4b8939.pth

I think you have downloaded all files. The default path means if you do not pass command line parameters to Python scripts, then that default parameter will be used. You can replace the parameter value with your own path.

cuteyyt commented 1 year ago

The file you underlined in the picture is not a model ckpt file but a config file for Grounding-DINO. I think this file is already contained in the project (GroundingDINO_SwinT_OGC.py).