This repo contains code for training and evaluating OmniSyn.
OmniSyn predicts depth from two panoramas and then renders meshes using PyTorch3D from intermediate positions.
-
Setup a conda environment
conda create -n omnisyn python=3.7 conda activate omnisyn conda install pytorch torchvision torchaudio cudatoolkit=11.3 -c pytorch pip install -r requirements.txt conda install habitat-sim=0.1.6 headless -c conda-forge -c aihabitat
-
Download our processed CARLA dataset.
-
Download our CARLA model weights to
example/
and run our selected example to check everything is working.python train_inpainting.py -c example/config_example.txt
Currently, the main scripts are:
train_depth.py
which trains the depth estimator only.train_inpainting.py
to train the inpainting component.
Our pretrained carla model uses an older version of PyTorch3D.
pip install git+https://github.com/facebookresearch/pytorch3d.git@07d7e12644ee48ea0b544c07becacdafe93c260a
Some code in this repo is borrowed from facebookresearch/synsin and nianticlabs/monodepth2.
If you use this in your research, please reference it as:
@INPROCEEDINGS{Li2022Omnisyn,
author={Li, David and Zhang, Yinda and H\"{a}ne, Christian and Tang, Danhang and Varshney, Amitabh and Du, Ruofei},
booktitle={2022 IEEE Conference on Virtual Reality and 3D User Interfaces Abstracts and Workshops (VRW)},
title={{OmniSyn}: Synthesizing 360 Videos with Wide-baseline Panoramas},
year={2022},
volume={},
number={},
pages={670-671},
doi={10.1109/VRW55335.2022.00186}}