Skip to content

Latest commit

 

History

History
46 lines (38 loc) · 1.88 KB

README.md

File metadata and controls

46 lines (38 loc) · 1.88 KB

OmniSyn

This repo contains code for training and evaluating OmniSyn.
OmniSyn predicts depth from two panoramas and then renders meshes using PyTorch3D from intermediate positions.

Getting Started

  1. Setup a conda environment

    conda create -n omnisyn python=3.7
    conda activate omnisyn
    conda install pytorch torchvision torchaudio cudatoolkit=11.3 -c pytorch
    pip install -r requirements.txt
    conda install habitat-sim=0.1.6 headless -c conda-forge -c aihabitat
  2. Download our processed CARLA dataset.

  3. Download our CARLA model weights to example/ and run our selected example to check everything is working.

    python train_inpainting.py -c example/config_example.txt

Currently, the main scripts are:

  • train_depth.py which trains the depth estimator only.
  • train_inpainting.py to train the inpainting component.

Notes

Our pretrained carla model uses an older version of PyTorch3D.
pip install git+https://github.com/facebookresearch/pytorch3d.git@07d7e12644ee48ea0b544c07becacdafe93c260a

Acknowledgements

Some code in this repo is borrowed from facebookresearch/synsin and nianticlabs/monodepth2.

Reference

If you use this in your research, please reference it as:

@INPROCEEDINGS{Li2022Omnisyn,
  author={Li, David and Zhang, Yinda and H\"{a}ne, Christian and Tang, Danhang and Varshney, Amitabh and Du, Ruofei},
  booktitle={2022 IEEE Conference on Virtual Reality and 3D User Interfaces Abstracts and Workshops (VRW)}, 
  title={{OmniSyn}: Synthesizing 360 Videos with Wide-baseline Panoramas}, 
  year={2022},
  volume={},
  number={},
  pages={670-671},
  doi={10.1109/VRW55335.2022.00186}}