This repository contains MegEngine implementation of our paper:
Practical Stereo Matching via Cascaded Recurrent Network with Adaptive Correlation\ Jiankun Li, Peisen Wang, Pengfei Xiong, Tao Cai, Ziwei Yan, Lei Yang, Jiangyu Liu, Haoqiang Fan, Shuaicheng Liu \ CVPR 2022 (Oral)
There are two ways to download the dataset(~400GB) proposed in our paper:
dataset_download.sh
sh dataset_download.sh
the dataset will be downloaded and extracted in ./stereo_trainset/crestereo
aa3g
) and extract the tar files manually.The disparity is saved as .png
uint16 format which can be loaded using opencv imread
function:
def get_disp(disp_path):
disp = cv2.imread(disp_path, cv2.IMREAD_UNCHANGED)
return disp.astype(np.float32) / 32
Other public datasets we use including
CUDA Version: 10.1, Python Version: 3.6.9
python3 -m pip install -r requirements.txt
We also provide docker to run the code quickly:
docker run --gpus all -it -v /tmp:/tmp ylmegvii/crestereo
shotwell /tmp/disparity.png
Download the pretrained MegEngine model from here and run:
python3 test.py --model_path path_to_mge_model --left img/test/left.png --right img/test/right.png --size 1024x1536 --output disparity.png
Modify the configurations in cfgs/train.yaml
and run the following command:
python3 train.py
You can launch a TensorBoard to monitor the training process:
tensorboard --logdir ./train_log
and navigate to the page at http://localhost:6006
in your browser.
Part of the code is adapted from previous works:
We thank all the authors for their awesome repos.
If you find the code or datasets helpful in your research, please cite:
@inproceedings{li2022practical,
title={Practical stereo matching via cascaded recurrent network with adaptive correlation},
author={Li, Jiankun and Wang, Peisen and Xiong, Pengfei and Cai, Tao and Yan, Ziwei and Yang, Lei and Liu, Jiangyu and Fan, Haoqiang and Liu, Shuaicheng},
booktitle={Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition},
pages={16263--16272},
year={2022}
}