Home

Awesome

Unbounded Neural Radiance Fields in Pytorch

<!-- ALL-CONTRIBUTORS-BADGE:START - Do not remove or modify this section -->

All Contributors

<!-- ALL-CONTRIBUTORS-BADGE:END -->

1. Introduction

This is still a research project in progress.

This project aims for benchmarking several state-of-the-art large-scale radiance fields algorithms. We exchangely use terms "unbounded NeRF" and "large-scale NeRF" because we find the techniques behind them are closely related.

Instead of pursuing a big and complicated code system, we pursue a simple code repo with SOTA performance for unbounded NeRFs.

You are expected to get the following results in this repository:

BenchmarkMethodsPSNR
Unbounded Tanks & TemplesNeRF++20.49
Unbounded Tanks & TemplesPlenoxels20.40
Unbounded Tanks & TemplesDVGO20.10
Unbounded Tanks & TemplesOurs20.85
Mip-NeRF-360 BenchmarkNeRF24.85
Mip-NeRF-360 BenchmarkNeRF++26.21
Mip-NeRF-360 BenchmarkMip-NeRF-36028.94
Mip-NeRF-360 BenchmarkDVGO25.42
Mip-NeRF-360 BenchmarkOurs28.98
<details> <summary> Expand / collapse qualitative results. </summary>

Tanks and Temples:

https://user-images.githubusercontent.com/31123348/220946729-d88db335-0618-4b75-9fc2-8de577e1ddb5.mp4

https://user-images.githubusercontent.com/31123348/220946857-0f4b7239-8be6-4fca-9bba-2f2425e857a5.mp4

https://user-images.githubusercontent.com/31123348/220947063-068b94f6-3afb-421d-8746-43bcf9643a37.mp4

https://user-images.githubusercontent.com/31123348/220947239-6528d542-b2b8-45e3-8e69-6e0eff869720.mp4

Mip-NeRF-360 Benchmark:

https://user-images.githubusercontent.com/31123348/220947385-ab31c646-c671-4522-8e4f-a1982d98c753.mp4

https://user-images.githubusercontent.com/31123348/220947472-47dc4716-095b-45ec-890b-d6afd97de9e9.mp4

https://user-images.githubusercontent.com/31123348/220947597-68f7ec32-c761-4253-955a-a2acc6a2eb25.mp4

https://user-images.githubusercontent.com/31123348/220947686-d8957a2e-ef52-46cf-b437-28de91f55871.mp4

https://user-images.githubusercontent.com/31123348/220947771-bbd249c0-3d0b-4d25-9b79-d4de9af17c4a.mp4

https://user-images.githubusercontent.com/31123348/220947818-e5c6b07f-c930-48b2-8aa7-363182dea6be.mp4

https://user-images.githubusercontent.com/31123348/220948025-25ce5cc1-3c9a-450c-920d-98a8f153a0fa.mp4

San Francisco Mission Bay (dataset released by Block-NeRF):

</details>

Hope our efforts could help your research or projects!

2. News

<details> <summary> Expand / collapse older news. </summary> </details>

3. Installation

<details> <summary>Expand / collapse installation steps.</summary>
  1. Clone this repository. Use depth == 1 to avoid download a large history.

    git clone --depth=1 git@github.com:sjtuytc/LargeScaleNeRFPytorch.git
    mkdir data
    mkdir logs
    
  2. Create conda environment.

    conda create -n large-scale-nerf python=3.9
    conda activate large-scale-nerf
    
  3. Install pytorch and other libs. Make sure your Pytorch version is compatible with your CUDA.

    pip install --upgrade pip
    conda install pytorch==1.13.1 torchvision==0.14.1 torchaudio==0.13.1 pytorch-cuda=11.6 -c pytorch -c nvidia
    pip install -r requirements.txt
    
  4. Install grid-based operators to avoid running them every time, cuda lib required. (Check via "nvcc -V" to ensure that you have a latest cuda.)

    apt-get install g++ build-essential  # ensure you have g++ and other build essentials, sudo access required.
    cd FourierGrid/cuda
    python setup.py install
    cd ../../
    
  5. Install other libs used for reconstructing custom scenes. This is only needed when you need to build your scenes.

    sudo apt-get install colmap
    sudo apt-get install imagemagick  # required sudo accesss
    conda install pytorch-scatter -c pyg  # or install via https://github.com/rusty1s/pytorch_scatter
    

    You can use laptop version of COLMAP as well if you do not have access to sudo access on your server. However, we found if you do not set up COLMAP parameters properly, you would not get the SOTA performance.

</details>

4. Unbounded NeRF on the public datasets

Click the following sub-section titles to expand / collapse steps.

<details> <summary> 4.1 Download processed data.</summary>

(1) Unbounded Tanks & Temples. Download data from here. Then unzip the data.

cd data
gdown --id 11KRfN91W1AxAW6lOFs4EeYDbeoQZCi87
unzip tanks_and_temples.zip
cd ../

(2) The Mip-NeRF-360 dataset.

cd data
wget http://storage.googleapis.com/gresearch/refraw360/360_v2.zip
mkdir 360_v2
unzip 360_v2.zip -d 360_v2
cd ../

(3) San Fran Cisco Mission Bay. What you should know before downloading the data:

The downloaded data would look like this:

data
   |
   |——————360_v2                                    // the root folder for the Mip-NeRF-360 benchmark
   |        └——————bicycle                          // one scene under the Mip-NeRF-360 benchmark
   |        |         └——————images                 // rgb images
   |        |         └——————images_2               // rgb images downscaled by 2
   |        |         └——————sparse                 // camera poses
   |        ...
   |——————tanks_and_temples                         // the root folder for Tanks&Temples
   |        └——————tat_intermediate_M60             // one scene under Tanks&Temples
   |        |         └——————camera_path            // render split camera poses, intrinsics and extrinsics
   |        |         └——————test                   // test split
   |        |         └——————train                  // train split
   |        |         └——————validation             // validation split
   |        ...
   |——————pytorch_waymo_dataset                     // the root folder for San Fran Cisco Mission Bay
   |        └——————cam_info.json                    // extracted cam2img information in dict.
   |        └——————coordinates.pt                   // global camera information used in Mega-NeRF, deprecated
   |        └——————train                            // train data
   |        |         └——————metadata               // meta data per image (camera information, etc)
   |        |         └——————rgbs                   // rgb images
   |        |         └——————split_block_train.json // split block informations
   |        |         └——————train_all_meta.json    // all meta informations in train folder
   |        └——————val                              // val data with the same structure as train
</details> <details> <summary> 4.2 Train models and see the results!</summary>

You only need to run "python run_FourierGrid.py" to finish the train-test-render cycle. Explanations of some arguments:

--program: the program to run, normally --program train will be all you need.
--config: the config pointing to the scene file, e.g., --config FourierGrid/configs/tankstemple_unbounded/truck_single.py.
--num_per_block: number of blocks used in large-scale NeRFs, normally this is set to -1, unless specially needed.
--render_train: render the trained model on the train split.
--render_train: render the trained model on the test split.
--render_train: render the trained model on the render split.
--exp_id: add some experimental ids to identify different experiments. E.g., --exp_id 5.
--eval_ssim / eval_lpips_vgg: report SSIM / LPIPS(VGG) scores.

While we list major of the commands in scripts/train_FourierGrid.sh, we list some of commands below for better reproducibility.

# Unbounded tanks and temples
python run_FourierGrid.py --program train --config FourierGrid/configs/tankstemple_unbounded/playground_single.py --num_per_block -1 --render_train --render_test --render_video --exp_id 57
python run_FourierGrid.py --program train --config FourierGrid/configs/tankstemple_unbounded/train_single.py --num_per_block -1 --render_train --render_test --render_video --exp_id 12
python run_FourierGrid.py --program train --config FourierGrid/configs/tankstemple_unbounded/truck_single.py --num_per_block -1 --render_train --render_test --render_video --exp_id 4
python run_FourierGrid.py --program train --config FourierGrid/configs/tankstemple_unbounded/m60_single.py --num_per_block -1 --render_train --render_test --render_video --exp_id 6

# 360 degree dataset
python run_FourierGrid.py --program train --config FourierGrid/configs/nerf_unbounded/room_single.py --num_per_block -1 --eval_ssim --eval_lpips_vgg --render_train --render_test --render_video --exp_id 9
python run_FourierGrid.py --program train --config FourierGrid/configs/nerf_unbounded/stump_single.py --num_per_block -1 --eval_ssim --eval_lpips_vgg --render_train --render_test --render_video --exp_id 10
python run_FourierGrid.py --program train --config FourierGrid/configs/nerf_unbounded/bicycle_single.py --num_per_block -1 --eval_ssim --eval_lpips_vgg --render_train --render_test --render_video --exp_id 11
python run_FourierGrid.py --program train --config FourierGrid/configs/nerf_unbounded/bonsai_single.py --num_per_block -1 --eval_ssim --eval_lpips_vgg --render_train --render_test --render_video --exp_id 3
python run_FourierGrid.py --program train --config FourierGrid/configs/nerf_unbounded/garden_single.py --num_per_block -1 --eval_ssim --eval_lpips_vgg --render_train --render_test --render_video --exp_id 2
python run_FourierGrid.py --program train --config FourierGrid/configs/nerf_unbounded/kitchen_single.py --num_per_block -1 --eval_ssim --eval_lpips_vgg --render_train --render_test --render_video --exp_id 2
python run_FourierGrid.py --program train --config FourierGrid/configs/nerf_unbounded/counter_single.py --num_per_block -1 --eval_ssim --eval_lpips_vgg --render_train --render_test --render_video --exp_id 2

# San Francisco Mission Bay dataset
python run_FourierGrid.py --program train --config FourierGrid/configs/waymo/waymo_no_block.py --num_per_block 100 --render_video --exp_id 30

The old version of Block-NeRF is still provided to serve as a baseline, but it will be deprecated soon. We will mainly work on grid-based models later because they are simple and fast. Run the following command to reproduce the old Block-NeRF experiments:

bash scripts/block_nerf_train.sh
bash scripts/block_nerf_eval.sh
</details>

5. Build your custom unbounded NeRF (deprecated)

<details> <summary>Expand / collapse steps for building custom NeRF world.</summary>
  1. Put your images under data folder. The structure should be like:

    data
       |——————Madoka          // Your folder name here.
       |        └——————source // Source images should be put here.
       |                 └——————---|1.png
       |                 └——————---|2.png
       |                 └——————---|...
    

    The sample data is provided in our Google drive folder. The Madoka and Otobai can be found at this link.

  2. Run COLMAP to reconstruct scenes. This would probably cost a long time.

    python FourierGrid/tools/imgs2poses.py data/Madoka
    

    You can replace data/Madoka by your data folder. If your COLMAP version is larger than 3.6 (which should not happen if you use apt-get), you need to change export_path to output_path in the colmap_wrapper.py.

  3. Training NeRF scenes.

    python FourierGrid/run_FourierGrid.py --config configs/custom/Madoka.py
    

    You can replace configs/custom/Madoka.py by other configs.

  4. Validating the training results to generate a fly-through video.

    python FourierGrid/run_FourierGrid.py --config configs/custom/Madoka.py --render_only --render_video --render_video_factor 8
    
</details>

6. Citations & acknowledgements

Our latest theoretical work about grid-based models (Oral & Best Paper Award Candidate & Full Review Score (5/5/5) at CVPR24):

@misc{zhao2024grounding,
      title={Grounding and Enhancing Grid-based Models for Neural Fields}, 
      author={Zelin Zhao and Fenglei Fan and Wenlong Liao and Junchi Yan},
      year={2024},
      eprint={2403.20002},
      archivePrefix={arXiv},
      primaryClass={cs.CV}
}

Consider citing the following great works:

@inproceedings{dvgo,
  title={Direct voxel grid optimization: Super-fast convergence for radiance fields reconstruction},
  author={Sun, Cheng and Sun, Min and Chen, Hwann-Tzong},
  booktitle={Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition},
  pages={5459--5469},
  year={2022}
}

 @InProceedings{Tancik_2022_CVPR,
    author    = {Tancik, Matthew and Casser, Vincent and Yan, Xinchen and Pradhan, Sabeek and Mildenhall, Ben and Srinivasan, Pratul P. and Barron, Jonathan T. and Kretzschmar, Henrik},
    title     = {Block-NeRF: Scalable Large Scene Neural View Synthesis},
    booktitle = {Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)},
    month     = {June},
    year      = {2022},
    pages     = {8248-8258}
}

We refer to the code and data from DVGO, nerf-pl and SVOX2, thanks for their great work!

Weekly classified NeRF

We track weekly NeRF papers and classify them. All previous published NeRF papers have been added to the list. We provide an English version and a Chinese version. We welcome contributions and corrections via PR.

We also provide an excel version (the meta data) of all NeRF papers, you can add your own comments or make your own paper analysis tools based on the structured meta data.

Contributors ✨

Thanks goes to these wonderful people (emoji key):

<!-- ALL-CONTRIBUTORS-LIST:START - Do not remove or modify this section --> <!-- prettier-ignore-start --> <!-- markdownlint-disable --> <table> <tbody> <tr> <td align="center" valign="top" width="14.28%"><a href="https://sjtuytc.github.io/"><img src="https://avatars.githubusercontent.com/u/31123348?v=4?s=100" width="100px;" alt="Zelin Zhao"/><br /><sub><b>Zelin Zhao</b></sub></a><br /><a href="https://github.com/sjtuytc/LargeScaleNeRFPytorch/commits?author=sjtuytc" title="Code">💻</a> <a href="#maintenance-sjtuytc" title="Maintenance">🚧</a></td> <td align="center" valign="top" width="14.28%"><a href="https://github.com/SEUleaderYang"><img src="https://avatars.githubusercontent.com/u/55042050?v=4?s=100" width="100px;" alt="EZ-Yang"/><br /><sub><b>EZ-Yang</b></sub></a><br /><a href="https://github.com/sjtuytc/LargeScaleNeRFPytorch/commits?author=SEUleaderYang" title="Code">💻</a></td> <td align="center" valign="top" width="14.28%"><a href="https://github.com/Alex-Alison-Zhang"><img src="https://avatars.githubusercontent.com/u/71915735?v=4?s=100" width="100px;" alt="Alex-Zhang"/><br /><sub><b>Alex-Zhang</b></sub></a><br /><a href="https://github.com/sjtuytc/LargeScaleNeRFPytorch/issues?q=author%3AAlex-Alison-Zhang" title="Bug reports">🐛</a></td> <td align="center" valign="top" width="14.28%"><a href="https://fanlu97.github.io/"><img src="https://avatars.githubusercontent.com/u/45007531?v=4?s=100" width="100px;" alt="Fan Lu"/><br /><sub><b>Fan Lu</b></sub></a><br /><a href="https://github.com/sjtuytc/LargeScaleNeRFPytorch/issues?q=author%3AFanLu97" title="Bug reports">🐛</a></td> <td align="center" valign="top" width="14.28%"><a href="https://maybeshewill-cv.github.io"><img src="https://avatars.githubusercontent.com/u/15725187?v=4?s=100" width="100px;" alt="MaybeShewill-CV"/><br /><sub><b>MaybeShewill-CV</b></sub></a><br /><a href="https://github.com/sjtuytc/LargeScaleNeRFPytorch/issues?q=author%3AMaybeShewill-CV" title="Bug reports">🐛</a></td> <td align="center" valign="top" width="14.28%"><a href="https://github.com/buer1121"><img src="https://avatars.githubusercontent.com/u/48516434?v=4?s=100" width="100px;" alt="buer1121"/><br /><sub><b>buer1121</b></sub></a><br /><a href="https://github.com/sjtuytc/LargeScaleNeRFPytorch/issues?q=author%3Abuer1121" title="Bug reports">🐛</a></td> </tr> </tbody> </table> <!-- markdownlint-restore --> <!-- prettier-ignore-end --> <!-- ALL-CONTRIBUTORS-LIST:END -->

This project follows the all-contributors specification. Contributions of any kind welcome!