https://github.com/dyhes/pointmvsnet
Code for "Point-based Multi-view Stereo Network" (ICCV 2019 Oral) & "Visibility-aware Point-based Multi-view Stereo Network" (TPAMI)
Science Score: 10.0%
This score indicates how likely this project is to be science-related based on various indicators:
-
○CITATION.cff file
-
○codemeta.json file
-
○.zenodo.json file
-
○DOI references
-
✓Academic publication links
Links to: ieee.org -
○Academic email domains
-
○Institutional organization owner
-
○JOSS paper metadata
-
○Scientific vocabulary similarity
Low similarity (11.3%) to scientific vocabulary
Last synced: 5 months ago
·
JSON representation
Repository
Code for "Point-based Multi-view Stereo Network" (ICCV 2019 Oral) & "Visibility-aware Point-based Multi-view Stereo Network" (TPAMI)
Statistics
- Stars: 0
- Watchers: 0
- Forks: 0
- Open Issues: 0
- Releases: 0
Fork of callmeray/PointMVSNet
Created almost 4 years ago
· Last pushed over 5 years ago
https://github.com/dyhes/PointMVSNet/blob/master/
# Point-based Multi-view Stereo Network & Visibility-aware Point-based Multi-view Stereo Network
## Introduction
[PointMVSNet](http://hansf.me/projects/PMVSNet/) is a deep point-based deep framework for multi-view stereo (MVS). PointMVSNet directly processes the target scene as point clouds and predicts the depth in a coarse-to-fine manner. Our network leverages 3D geometry priors and 2D texture information jointly and effectively by fusing them into a feature-augmented point cloud, and processes the point cloud to estimate the 3D flow for each point.
[VAPointMVSNet](https://ieeexplore.ieee.org/abstract/document/9076298) extends PointMVSNet with visibility-aware multi-view feature aggregations, which allows the network to aggregate multi-view appearance cues while taking into account occlusions.
If you find this project useful for your research, please cite:
```
@ARTICLE{ChenVAPMVSNet2020TPAMI,
author={Chen, Rui and Han, Songfang and Xu, Jing and Su, Hao},
journal={IEEE Transactions on Pattern Analysis and Machine Intelligence},
title={Visibility-Aware Point-Based Multi-View Stereo Network},
year={2020},
volume={},
number={},
pages={1-1},}
```
```
@InProceedings{ChenPMVSNet2019ICCV,
author = {Chen, Rui and Han, Songfang and Xu, Jing and Su, Hao},
title = {Point-based Multi-view Stereo Network},
booktitle = {The IEEE International Conference on Computer Vision (ICCV)},
year = {2019}
}
```
## How to use
### Environment
The environment requirements are listed as follows:
- Pytorch 1.0.1
- CUDA 9.0
- CUDNN 7.4.2
- GCC5
### Installation
* Check out the source code
```git clone https://github.com/callmeray/PointMVSNet && cd PointMVSNet```
* Install dependencies
```bash install_dependencies.sh```
* Compile CUDA extensions
```bash compile.sh```
### Training
* Download the preprocessed [DTU training data](https://drive.google.com/file/d/1eDjh-_bxKKnEuz5h-HXS7EDJn59clx6V/view) from [MVSNet](https://github.com/YoYo000/MVSNet) and unzip it to ```data/dtu```.
* Train the network
```python pointmvsnet/train.py --cfg configs/dtu_wde3.yaml```
You could change the batch size in the configuration file according to your own pc.
### Testing
* Download the [rectified images](http://roboimagedata2.compute.dtu.dk/data/MVS/Rectified.zip) from [DTU benchmark](http://roboimagedata.compute.dtu.dk/?page_id=36) and unzip it to ```data/dtu/Eval```.
* Test with your own model
```python pointmvsnet/test.py --cfg configs/dtu_wde3.yaml```
* Test with the pretrained model
```python pointmvsnet/test.py --cfg configs/dtu_wde3.yaml TEST.WEIGHT outputs/dtu_wde3/model_pretrained.pth```
### Depth Fusion
PointMVSNet generates per-view depth map. We need to apply depth fusion ```tools/depthfusion.py``` to get the complete point cloud. Please refer to [MVSNet](https://github.com/YoYo000/MVSNet) for more details.
Owner
- Name: heslin
- Login: dyhes
- Kind: user
- Location: Beijing
- Company: Beijing Institute of Technology
- Website: https://www.cnblogs.com/heslin
- Twitter: heslinn_
- Repositories: 2
- Profile: https://github.com/dyhes
Don't just make a living, but try to make a difference.