This repository contains the code to reproduce the results from the paper. Reconstructing Surfaces for Sparse Point Clouds with On-Surface Priors.
You can find detailed usage instructions for training your own models and using pretrained models below.
If you find our code or paper useful, please consider citing
@inproceedings{On-SurfacePriors,
title = {Reconstructing Surfaces for Sparse Point Clouds with On-Surface Priors},
author = {Baorui, Ma and Yu-Shen, Liu and Zhizhong, Han},
booktitle = {Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)},
year = {2022}
}
This work was originally implemented by tensorflow, pytroch version of the code will be released soon that is easier to use.
Related work
Pytorch
https://github.com/mabaorui/NeuralPull-Pytorch
Tensorflow
https://github.com/mabaorui/NeuralPull
https://github.com/mabaorui/OnSurfacePrior
https://github.com/mabaorui/PredictableContextPrior
First you have to make sure that you have all dependencies in place. The simplest way to do so, is to use anaconda.
You can create an anaconda environment called tf
using
conda env create -f tf.yaml
conda activate tf
You can train a new network from pre-train On-Surface Prior Networks, run
python onSurPrior.py --data_dir ./data/downsample/ --out_dir ./train_net/ --CUDA 0 --INPUT_NUM 500 --epoch 30000 --input_ply_file input.ply --train
You should put the point cloud file(--input_ply_file, only ply format) into the '--out_dir' folder, '--INPUT_NUM' is the number of points in the '--input_ply_file'.
You can extract the mesh model from the trained network, run
python onSurPrior.py --data_dir ./data/downsample --out_dir ./train_net/ --CUDA 0 --INPUT_NUM 500 --epoch 30000 --input_ply_file input.ply --test
In different datasets or your own data, because of the variation in point cloud density, this '0.25' parameter has a very strong influence on the final result, which controls the distance between the query points and the point cloud. So if you want to get better results, you should adjust this parameter. We give '0.25' here as a reference value, and this value can be used for most object-level reconstructions. For the scene dataset, we will later publish the reference values for the hyperparameter settings for the scene dataset.