README.md 5.97 KB
Newer Older
Matthew Tancik's avatar
Matthew Tancik committed
1
2
<picture>
    <source media="(prefers-color-scheme: dark)" srcset="https://user-images.githubusercontent.com/3310961/199083722-881a2372-62c1-4255-8521-31a95a721851.png" />
Ruilong Li(李瑞龙)'s avatar
Ruilong Li(李瑞龙) committed
3
    <img alt="nerfacc logo" src="https://user-images.githubusercontent.com/3310961/199084143-0d63eb40-3f35-48d2-a9d5-78d1d60b7d66.png" width="350px" />
Matthew Tancik's avatar
Matthew Tancik committed
4
5
</picture>

Matthew Tancik's avatar
Matthew Tancik committed
6
[![Core Tests.](https://github.com/KAIR-BAIR/nerfacc/actions/workflows/code_checks.yml/badge.svg)](https://github.com/KAIR-BAIR/nerfacc/actions/workflows/code_checks.yml)
7
[![Documentation Status](https://readthedocs.com/projects/plenoptix-nerfacc/badge/?version=latest)](https://www.nerfacc.com/en/latest/?badge=latest)
Ruilong Li(李瑞龙)'s avatar
Ruilong Li(李瑞龙) committed
8
[![Downloads](https://pepy.tech/badge/nerfacc)](https://pepy.tech/project/nerfacc)
Ruilong Li's avatar
readme  
Ruilong Li committed
9

10
https://www.nerfacc.com/
Ruilong Li(李瑞龙)'s avatar
Ruilong Li(李瑞龙) committed
11

12
13
14
15
16
NerfAcc is a PyTorch Nerf acceleration toolbox for both training and inference. It focus on
efficient volumetric rendering of radiance fields, which is universal and plug-and-play for most of the NeRFs.

Using NerfAcc, 

Ruilong Li(李瑞龙)'s avatar
Ruilong Li(李瑞龙) committed
17
- The `vanilla NeRF` model with 8-layer MLPs can be trained to *better quality* (+~0.5 PNSR)
Ruilong Li(李瑞龙)'s avatar
Ruilong Li(李瑞龙) committed
18
  in *1 hour* rather than *days* as in the paper.
Ruilong Li(李瑞龙)'s avatar
Ruilong Li(李瑞龙) committed
19
20
- The `Instant-NGP NeRF` model can be trained to *equal quality* in *4.5 minutes*,
  comparing to the official pure-CUDA implementation.
Ruilong Li(李瑞龙)'s avatar
Ruilong Li(李瑞龙) committed
21
- The `D-NeRF` model for *dynamic* objects can also be trained in *1 hour*
22
  rather than *2 days* as in the paper, and with *better quality* (+~2.5 PSNR).
23
24
25
- Both *bounded* and *unbounded* scenes are supported.

**And it is pure Python interface with flexible APIs!**
Ruilong Li's avatar
readme  
Ruilong Li committed
26

27
28
29
30
31
32
## Installation

```
pip install nerfacc
```

33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
## Usage

The idea of NerfAcc is to perform efficient ray marching and volumetric rendering. So NerfAcc can work with any user-defined radiance field. To plug the NerfAcc rendering pipeline into your code and enjoy the acceleration, you only need to define two functions with your radience field.
- `sigma_fn`: Compute density at each sample. It will be used by `nerfacc.ray_marching()` to skip the empty and occluded space during ray marching, which is where the major speedup comes from. 
- `rgb_sigma_fn`: Compute color and density at each sample. It will be used by `nerfacc.rendering()` to conduct differentiable volumetric rendering. This function will receive gradients to update your network.

An simple example is like this:

``` python
import torch
from torch import Tensor
import nerfacc 

radiance_field = ...  # network: a NeRF model
rays_o: Tensor = ...  # ray origins. (n_rays, 3)
rays_d: Tensor = ...  # ray normalized directions. (n_rays, 3)
49
optimizer = ...  # optimizer
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83

def sigma_fn(
    t_starts: Tensor, t_ends:Tensor, ray_indices: Tensor
) -> Tensor:
    """ Query density values from a user-defined radiance field.
    :params t_starts: Start of the sample interval along the ray. (n_samples, 1).
    :params t_ends: End of the sample interval along the ray. (n_samples, 1).
    :params ray_indices: Ray indices that each sample belongs to. (n_samples,).
    :returns The post-activation density values. (n_samples, 1).
    """
    t_origins = rays_o[ray_indices]  # (n_samples, 3)
    t_dirs = rays_d[ray_indices]  # (n_samples, 3)
    positions = t_origins + t_dirs * (t_starts + t_ends) / 2.0
    sigmas = radiance_field.query_density(positions) 
    return sigmas  # (n_samples, 1)

def rgb_sigma_fn(
    t_starts: Tensor, t_ends: Tensor, ray_indices: Tensor
) -> Tuple[Tensor, Tensor]:
    """ Query rgb and density values from a user-defined radiance field.
    :params t_starts: Start of the sample interval along the ray. (n_samples, 1).
    :params t_ends: End of the sample interval along the ray. (n_samples, 1).
    :params ray_indices: Ray indices that each sample belongs to. (n_samples,).
    :returns The post-activation rgb and density values. 
        (n_samples, 3), (n_samples, 1).
    """
    t_origins = rays_o[ray_indices]  # (n_samples, 3)
    t_dirs = rays_d[ray_indices]  # (n_samples, 3)
    positions = t_origins + t_dirs * (t_starts + t_ends) / 2.0
    rgbs, sigmas = radiance_field(positions, condition=t_dirs)  
    return rgbs, sigmas  # (n_samples, 3), (n_samples, 1)

# Efficient Raymarching: Skip empty and occluded space, pack samples from all rays.
# packed_info: (n_rays, 2). t_starts: (n_samples, 1). t_ends: (n_samples, 1).
84
85
86
87
88
with torch.no_grad():
    packed_info, t_starts, t_ends = nerfacc.ray_marching(
        rays_o, rays_d, sigma_fn=sigma_fn, near_plane=0.2, far_plane=1.0, 
        early_stop_eps=1e-4, alpha_thre=1e-2, 
    )
89
90
91
92
93

# Differentiable Volumetric Rendering.
# colors: (n_rays, 3). opaicity: (n_rays, 1). depth: (n_rays, 1).
color, opacity, depth = nerfacc.rendering(rgb_sigma_fn, packed_info, t_starts, t_ends)

94
# Optimize: Both the network and rays will receive gradients
95
96
97
98
99
100
optimizer.zero_grad()
loss = F.mse_loss(color, color_gt)
loss.backward()
optimizer.step()
```

Ruilong Li(李瑞龙)'s avatar
Ruilong Li(李瑞龙) committed
101
## Examples: 
Ruilong Li's avatar
readme  
Ruilong Li committed
102

Ruilong Li(李瑞龙)'s avatar
Ruilong Li(李瑞龙) committed
103
Before running those example scripts, please check the script about which dataset it is needed, and download
Ruilong Li(李瑞龙)'s avatar
Ruilong Li(李瑞龙) committed
104
105
the dataset first.

Ruilong Li's avatar
Ruilong Li committed
106
``` bash
Ruilong Li(李瑞龙)'s avatar
Ruilong Li(李瑞龙) committed
107
# Instant-NGP NeRF in 4.5 minutes with reproduced performance!
108
# See results at here: https://www.nerfacc.com/en/latest/examples/ngp.html
Ruilong Li(李瑞龙)'s avatar
Ruilong Li(李瑞龙) committed
109
python examples/train_ngp_nerf.py --train_split train --scene lego
Ruilong Li's avatar
readme  
Ruilong Li committed
110
111
```

Ruilong Li's avatar
Ruilong Li committed
112
``` bash
113
# Vanilla MLP NeRF in 1 hour with better performance!
114
# See results at here: https://www.nerfacc.com/en/latest/examples/vanilla.html
Ruilong Li(李瑞龙)'s avatar
Ruilong Li(李瑞龙) committed
115
python examples/train_mlp_nerf.py --train_split train --scene lego
Ruilong Li's avatar
Ruilong Li committed
116
117
```

Ruilong Li(李瑞龙)'s avatar
Ruilong Li(李瑞龙) committed
118
```bash
119
# D-NeRF for Dynamic objects in 1 hour with better performance!
120
# See results at here: https://www.nerfacc.com/en/latest/examples/dnerf.html
Ruilong Li(李瑞龙)'s avatar
Ruilong Li(李瑞龙) committed
121
python examples/train_mlp_dnerf.py --train_split train --scene lego
122
123
```

Ruilong Li(李瑞龙)'s avatar
Ruilong Li(李瑞龙) committed
124
```bash
125
# Instant-NGP on unbounded scenes in 20 minutes!
126
# See results at here: https://www.nerfacc.com/en/latest/examples/unbounded.html
Ruilong Li(李瑞龙)'s avatar
Ruilong Li(李瑞龙) committed
127
python examples/train_ngp_nerf.py --train_split train --scene garden --auto_aabb --unbounded --cone_angle=0.004
Ruilong Li(李瑞龙)'s avatar
Ruilong Li(李瑞龙) committed
128
```
129
130
131
132
133
134
135
136
137
138
139


## Citation

```bibtex
@article{li2022nerfacc,
  title={NerfAcc: A General NeRF Accleration Toolbox.},
  author={Li, Ruilong and Tancik, Matthew and Kanazawa, Angjoo},
  journal={arXiv preprint arXiv:2210.04847},
  year={2022}
}
Ruilong Li(李瑞龙)'s avatar
Ruilong Li(李瑞龙) committed
140
```