CBMV: A Coalesced Bidirectional Matching Volume for Disparity Estimation code repository. If you use this code please cite our paper CBMV: A Coalesced Bidirectional Matching Volume for Disparity Estimation.
@inproceedings{batsos2018cbmv,
title={CBMV: A Coalesced Bidirectional Matching Volume for Disparity Estimation},
author={Batsos, Konstantinos and Cai, Changjiang and Mordohai, Philipos},
booktitle={IEEE Conference on Computer Vision and Pattern Recognition (CVPR)},
year={2018}
}
The repository contains the code, models and procudures for training and testing. The code includes the cost-optimization and post-processing of MC-CNN Stereo Matching by Training a Convolutional Neural Network to Compare Image Patches as implemented by Jure Zbontar, and modified to interface with python. We aslo incorporate, as an alternative cost-optimization method, the code from Continuous 3D Label Stereo Matching using Local Expansion Moves.
CBMV model: the trained random forest model, with version 0.17.1 sklearn.
To run CBMV you need the following python packages:
- numpy
- scipy
- sklearn (version 0.17.1)
- cPickle
Much of CBMV code is written as C++ libraries and interface with python via boost python. To compile the C++ libraries you will need the following:
- OpenCV 3.0
- Boost ( boost python )
- CUDA
Assuming this repository is rooted at "~/cbmv-cvpr18/", the OpenCV library is installed at /usr/local/opencv-3.2.0.
-
- To compile
matchers,featextractandrectification:
We provide bothmakefileandCMakeLists.txt(preferred) for compilation. If you have installed OpenCV at different directory, please update the lineg++ -I/usr/local/opencv-3.2.0/include/ ...in the file*/Debug/subdir.mkfor including header files; and modify the lineg++ ... -L/usr/local/opencv-3.2.0/lib ...in the file*/Debug/makefilefor library linking. Or to modify the lineset(OpenCV_DIR "/usr/local/opencv-3.2.0/share/OpenCV/")in the fileCMakeList.txt.
- To compile
# 1) Method 1 : using CMake, will generate the libraries in the directory `~/cbmv-cvpr18/cpp/lib`.
cd ~/cbmv-cvpr18/cpp/ && mkdir build && cd build && cmake .. && build
# 2) Method 2 : using makefile
#To compile `matchers` for four basic matchers:
cd ~/cbmv-cvpr18/cpp/matchers/Debug && make clean && make
# To compile `featextract` for feature extraction:
cd ~/cbmv-cvpr18/cpp/featextract/Debug && make clean && make
# To compile `rectification` for rectifying stereo image pair:
cd ~/cbmv-cvpr18/cpp/rectification/Debug && make clean && make-
- To compile
postfor postprocessing:
This part contains the GPU code. We assume the CUDA was installed at/usr/local/cuda-8.0/. If not, please modify the lineCUDA_LIBS = /usr/local/cuda-8.0/lib64in the file*/post/makefile.
- To compile
cd ~/cbmv-cvpr18/cpp/post/ && make clean && make-
- To compile
localexpfor Local Expansion Moves:
This part requires the Maxflow code by Boykov and Kolmogorov [Code v3.01] [Code v3.04]. For your convenience, we already include and confiure it. But please note they are freely available for research purposes only. You could also check the Local Expansion Move repository for the license. Forlocalexpcompilation, we provide theCMakeLists.txtfile. Run the following
- To compile
# will generate the libraries in the directory `*/localexp/lib`.
cd ~/cbmv-cvpr18/cpp/localexp && mkdir build && cd build && cmake .. && makeYou can change the OMP threads number defined by #define THREADS_NUM_USED 16 at file ~/cbmv-cvpr18/cpp/paramSetting.hpp, for efficient parallel computation.
After compiling the C++ libraries, downloading the required python packages and precomputed model (Note: pay attention to the version of sklearn), CBMV can be executed using the following command
python main.py --l ./datasets/ArtL/im0.png --r ./datasets/ArtL/im1.png --w 694 --h 554 --d 128 --model ./modelall_4m.rf --prob_save_path ./
For a complete list of arguments please see tha main.py file. We provide a script run_test_cbmv.sh to run CBMV for testing and see more details in section Disparity Estimation.
The above link includes the data we used to trained CBMV. To train CBMV you can use the following command:
python main.py --train --data_path "./datasets/" --train_add ./mb/additional.txt --train_set ./mb/trainall.txt --model ./models/model.rf
The *.txt files specify the training sets used during the training process. For a complete list of arguments please see tha main.py file. We also provide a script run_train_cbmv.sh to train CBMV.
Giveing the cost volume, there are two methods for cost-optimization in order to generate the disparity map as output.
- Post-processing used by MC-CNN: see the code
~/cbmv-cvpr18/cpp/post/post.cufor details. - Local Expansion Moves: see the codes
~/cbmv-cvpr18/cpp/localexp/local_exp_stereo.cppfor more details. You can also modify the hyperparameters defined in function__postprocessing_localExp(...)in the file~/cbmv-cvpr18/test.py.