Home

Awesome

Unifying Correspondence, Pose and NeRF for Generalized Pose-Free Novel View Synthesis from Stereo Pairs CVPR-2024 (Highlight)

This is our official implementation of CoPoNeRF CVPR-2024 (Highlight)!

[arXiv] [Project] <br>

by Sunghwan Hong, Jaewoo Jung, Heeseong Shin, Jiaolong Yang, Seungryong Kim, Chong Luo,

Introduction

We delve into the task of generalized pose-free novel view synthesis from stereo pairs, a challenging and pioneering task in 3D vision.

For further details and visualization results, please check out our paper and our project page.

❗️Update: This repository includes refactored codes:

Installation

Requirements

An example of installation is shown below:

git clone https://github.com/KU-CVLAB/CoPoNeRF.git
cd CoPoNeRF
conda create -n CoPoNeRF python=3.8
conda activate CoPoNeRF
conda install pytorch==1.13.1 torchvision==0.14.1 torchaudio==0.13.1 pytorch-cuda=11.7 -c pytorch -c nvidia
pip install -r assets/requirements.txt

Data Preparation

Please follow dataset preperation.

Training

python train.py --experiment_name [EXPERIMENT_NAME] 

If you want to specify batch size or number of gpus, include them as arguments. Also, you can freely add the losses, e.g., cycle loss. To train on ACID, you can do it by simply changing the directories and the dataset in train.py.

Evaluation

python test.py --checkpoint_path [CHECKPOINT_PATH]

Pretrained Models

We provide pretrained weights here. These models were trained with 4 A6000.

Acknowledgement

We would like to acknowledge the contributions of public projects, such as Du et al. and UFC, whose code has been utilized in this repository.

Citing CoPoNeRF:

@article{hong2023unifying,
  title={Unifying Correspondence, Pose and NeRF for Pose-Free Novel View Synthesis from Stereo Pairs},
  author={Hong, Sunghwan and Jung, Jaewoo and Shin, Heeseong and Yang, Jiaolong and Kim, Seungryong and Luo, Chong},
  journal={arXiv preprint arXiv:2312.07246},
  year={2023}
}