MVSNet is a deep learning architecture for depth map inference from unstructured multi-view images. If you find this project useful for your research, please cite:
@article{yao2018mvsnet,
title={MVSNet: Depth Inference for Unstructured Multi-view Stereo},
author={Yao, Yao and Luo, Zixin and Li, Shiwei and Fang, Tian and Quan, Long},
journal={arXiv preprint arXiv:1804.02505},
year={2018}
}
- check out the source code
git clone https://github.com/YoYo000/MVSNet
- Install cuda 9.0 and cudnn 7.0
- Install Tensorflow and other dependencies by
sudo pip install -r requirements.txt
- Download the preprocessed DTU training data (see the paper), and upzip it as the
MVS_TRANING
folder. - Enter the
MVSNet/mvsnet
folder, intrain.py
, setdtu_data_root
to yourMVS_TRANING
path. - Create a log folder and a model folder in wherever you like to save the training outputs. Set the
log_dir
andsave_dir
intrain.py
correspondingly. - Train the network
python train.py
- Download the test data for scan9 and unzip it as the
TEST_DATA_FOLDER
folder, which should contain onecams
folder, oneimages
folder and onepair.txt
file. - Download the pre-trained MVSNet model and upzip it as
MODEL_FOLDER
. - Enter the
MVSNet/mvsnet
folder, intest.py
, setpretrained_model_ckpt_path
toMODEL_FOLDER/model.ckpt
- Depth map inference for this test data by
python test.py --dense_folder TEST_DATA_FOLDER
. - Inspect the .pfm format outputs in
TEST_DATA_FOLDER/depths_mvsnet
usingpython visualize.py .pfm
- File formats
- Post-processing.