Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
OpenDAS
nerfacc
Commits
62e70d39
Unverified
Commit
62e70d39
authored
Oct 12, 2022
by
Ruilong Li(李瑞龙)
Committed by
GitHub
Oct 12, 2022
Browse files
update ngp perf. (#68)
parent
76c0f981
Changes
6
Show whitespace changes
Inline
Side-by-side
Showing
6 changed files
with
30 additions
and
25 deletions
+30
-25
README.md
README.md
+4
-4
docs/source/examples/dnerf.rst
docs/source/examples/dnerf.rst
+1
-1
docs/source/examples/ngp.rst
docs/source/examples/ngp.rst
+21
-16
docs/source/examples/unbounded.rst
docs/source/examples/unbounded.rst
+1
-1
docs/source/examples/vanilla.rst
docs/source/examples/vanilla.rst
+1
-1
docs/source/index.rst
docs/source/index.rst
+2
-2
No files found.
README.md
View file @
62e70d39
...
@@ -12,8 +12,8 @@ Using NerfAcc,
...
@@ -12,8 +12,8 @@ Using NerfAcc,
-
The
`vanilla NeRF`
model with 8-layer MLPs can be trained to
*better quality*
(+~0.5 PNSR)
-
The
`vanilla NeRF`
model with 8-layer MLPs can be trained to
*better quality*
(+~0.5 PNSR)
in
*1 hour*
rather than
*days*
as in the paper.
in
*1 hour*
rather than
*days*
as in the paper.
-
The
`Instant-NGP NeRF`
model can be trained to
*
better
quality*
(+~0.7 PSNR) with
*9/10th*
of
-
The
`Instant-NGP NeRF`
model can be trained to
*
equal
quality*
in
*4.5 minutes*
,
the training time (4.5 minutes)
comparing to the official pure-CUDA implementation.
comparing to the official pure-CUDA implementation.
-
The
`D-NeRF`
model for
*dynamic*
objects can also be trained in
*1 hour*
-
The
`D-NeRF`
model for
*dynamic*
objects can also be trained in
*1 hour*
rather than
*2 days*
as in the paper, and with
*better quality*
(+~2.5 PSNR).
rather than
*2 days*
as in the paper, and with
*better quality*
(+~2.5 PSNR).
-
Both
*bounded*
and
*unbounded*
scenes are supported.
-
Both
*bounded*
and
*unbounded*
scenes are supported.
...
@@ -100,9 +100,9 @@ Before running those example scripts, please check the script about which datase
...
@@ -100,9 +100,9 @@ Before running those example scripts, please check the script about which datase
the dataset first.
the dataset first.
```
bash
```
bash
# Instant-NGP NeRF in 4.5 minutes with
better
performance!
# Instant-NGP NeRF in 4.5 minutes with
reproduced
performance!
# See results at here: https://www.nerfacc.com/en/latest/examples/ngp.html
# See results at here: https://www.nerfacc.com/en/latest/examples/ngp.html
python examples/train_ngp_nerf.py
--train_split
train
val
--scene
lego
python examples/train_ngp_nerf.py
--train_split
train
--scene
lego
```
```
```
bash
```
bash
...
...
docs/source/examples/dnerf.rst
View file @
62e70d39
...
@@ -32,5 +32,5 @@ single NVIDIA TITAN RTX GPU. The training memory footprint is about 11GB.
...
@@ -32,5 +32,5 @@ single NVIDIA TITAN RTX GPU. The training memory footprint is about 11GB.
.. _`D-Nerf`: https://arxiv.org/abs/2011.13961
.. _`D-Nerf`: https://arxiv.org/abs/2011.13961
.. _`D-Nerf dataset`: https://www.dropbox.com/s/0bf6fl0ye2vz3vr/data.zip?dl=0
.. _`D-Nerf dataset`: https://www.dropbox.com/s/0bf6fl0ye2vz3vr/data.zip?dl=0
.. _`github repository`: https://github.com/KAIR-BAIR/nerfacc/tree/
5637cc9a1565b2685c02250eb1ee1c53d3b07af1
.. _`github repository`: https://github.com/KAIR-BAIR/nerfacc/tree/
76c0f9817da4c9c8b5ccf827eb069ee2ce854b75
docs/source/examples/ngp.rst
View file @
62e70d39
...
@@ -7,29 +7,34 @@ See code `examples/train_ngp_nerf.py` at our `github repository`_ for details.
...
@@ -7,29 +7,34 @@ See code `examples/train_ngp_nerf.py` at our `github repository`_ for details.
Benchmarks
Benchmarks
------------
------------
*updated on 2022-10-
08
*
*updated on 2022-10-
12
*
Here we trained a `Instant-NGP Nerf`_ model on the `Nerf-Synthetic dataset`_. We follow the same
Here we trained a `Instant-NGP Nerf`_ model on the `Nerf-Synthetic dataset`_. We follow the same
settings with the Instant-NGP paper, which uses train
val
split for training and test split for
settings with the Instant-NGP paper, which uses train split for training and test split for
evaluation.
Our
experiments are conducted on a single NVIDIA TITAN RTX GPU. The training
evaluation.
All
experiments are conducted on a single NVIDIA TITAN RTX GPU. The training
memory footprint is about 3GB.
memory footprint is about 3GB.
.. note::
.. note::
The Instant-NGP paper makes use of the alpha channel in the images to apply random background
The Instant-NGP paper makes use of the alpha channel in the images to apply random background
augmentation during training. Yet we only uses RGB values with a constant white background.
augmentation during training. For fair comparision, we rerun their code with a constant white
background during both training and testing. Also it is worth to mention that we didn't strictly
follow the training receipe in the Instant-NGP paper, such as the learning rate schedule etc, as
the purpose of this benchmark is to showcase instead of reproducing the paper.
+----------------------+-------+-------+---------+-------+-------+-------+-------+-------+-------+
+----------------------
-
+-------+-------+---------+-------+-------+-------+-------+-------+-------+
| PSNR | Lego | Mic |Materials| Chair |Hotdog | Ficus | Drums | Ship | MEAN |
| PSNR | Lego | Mic |Materials| Chair |Hotdog | Ficus | Drums | Ship | MEAN |
| | | | | | | | | | |
| | | | | | | | | | |
+======================+=======+=======+=========+=======+=======+=======+=======+=======+=======+
+=======================+=======+=======+=========+=======+=======+=======+=======+=======+=======+
| Instant-NGP (5min) | 36.39 | 36.22 | 29.78 | 35.00 | 37.40 | 33.51 | 26.02 | 31.10 | 33.18 |
|Instant-NGP 35k steps | 35.87 | 36.22 | 29.08 | 35.10 | 37.48 | 30.61 | 23.85 | 30.62 | 32.35 |
+----------------------+-------+-------+---------+-------+-------+-------+-------+-------+-------+
+-----------------------+-------+-------+---------+-------+-------+-------+-------+-------+-------+
| Ours (~4.5min) | 36.82 | 37.61 | 30.18 | 36.13 | 38.11 | 34.48 | 26.62 | 31.37 | 33.92 |
|(training time) | 309s | 258s | 256s | 316s | 292s | 207s | 218s | 250s | 263s |
+----------------------+-------+-------+---------+-------+-------+-------+-------+-------+-------+
+-----------------------+-------+-------+---------+-------+-------+-------+-------+-------+-------+
| Ours (Training time)| 288s | 259s | 256s | 324s | 288s | 245s | 262s | 257s | 272s |
|Ours 20k steps | 35.50 | 36.16 | 29.14 | 35.23 | 37.15 | 31.71 | 24.88 | 29.91 | 32.46 |
+----------------------+-------+-------+---------+-------+-------+-------+-------+-------+-------+
+-----------------------+-------+-------+---------+-------+-------+-------+-------+-------+-------+
|(training time) | 287s | 274s | 269s | 317s | 269s | 244s | 249s | 257s | 271s |
.. _`Instant-NGP Nerf`: https://arxiv.org/abs/2201.05989
+-----------------------+-------+-------+---------+-------+-------+-------+-------+-------+-------+
.. _`github repository`: https://github.com/KAIR-BAIR/nerfacc/tree/5637cc9a1565b2685c02250eb1ee1c53d3b07af1
.. _`Instant-NGP Nerf`: https://github.com/NVlabs/instant-ngp/tree/51e4107edf48338e9ab0316d56a222e0adf87143
.. _`github repository`: https://github.com/KAIR-BAIR/nerfacc/tree/76c0f9817da4c9c8b5ccf827eb069ee2ce854b75
.. _`Nerf-Synthetic dataset`: https://drive.google.com/drive/folders/1JDdLGDruGNXWnM1eqY1FNL9PlStjaKWi
.. _`Nerf-Synthetic dataset`: https://drive.google.com/drive/folders/1JDdLGDruGNXWnM1eqY1FNL9PlStjaKWi
docs/source/examples/unbounded.rst
View file @
62e70d39
...
@@ -40,4 +40,4 @@ that takes from `MipNerf360`_.
...
@@ -40,4 +40,4 @@ that takes from `MipNerf360`_.
.. _`Instant-NGP Nerf`: https://arxiv.org/abs/2201.05989
.. _`Instant-NGP Nerf`: https://arxiv.org/abs/2201.05989
.. _`MipNerf360`: https://arxiv.org/abs/2111.12077
.. _`MipNerf360`: https://arxiv.org/abs/2111.12077
.. _`Nerf++`: https://arxiv.org/abs/2010.07492
.. _`Nerf++`: https://arxiv.org/abs/2010.07492
.. _`github repository`: https://github.com/KAIR-BAIR/nerfacc/tree/
5637cc9a1565b2685c02250eb1ee1c53d3b07af1
.. _`github repository`: https://github.com/KAIR-BAIR/nerfacc/tree/
76c0f9817da4c9c8b5ccf827eb069ee2ce854b75
docs/source/examples/vanilla.rst
View file @
62e70d39
...
@@ -29,5 +29,5 @@ conducted on a single NVIDIA TITAN RTX GPU. The training memory footprint is abo
...
@@ -29,5 +29,5 @@ conducted on a single NVIDIA TITAN RTX GPU. The training memory footprint is abo
| Ours (Training time)| 58min | 53min | 46min | 62min | 56min | 42min | 52min | 49min | 52min |
| Ours (Training time)| 58min | 53min | 46min | 62min | 56min | 42min | 52min | 49min | 52min |
+----------------------+-------+-------+---------+-------+-------+-------+-------+-------+-------+
+----------------------+-------+-------+---------+-------+-------+-------+-------+-------+-------+
.. _`github repository`: : https://github.com/KAIR-BAIR/nerfacc/tree/
5637cc9a1565b2685c02250eb1ee1c53d3b07af1
.. _`github repository`: : https://github.com/KAIR-BAIR/nerfacc/tree/
76c0f9817da4c9c8b5ccf827eb069ee2ce854b75
.. _`vanilla Nerf`: https://arxiv.org/abs/2003.08934
.. _`vanilla Nerf`: https://arxiv.org/abs/2003.08934
docs/source/index.rst
View file @
62e70d39
...
@@ -8,8 +8,8 @@ Using NerfAcc,
...
@@ -8,8 +8,8 @@ Using NerfAcc,
- The `vanilla Nerf`_ model with 8-layer MLPs can be trained to *better quality* (+~0.5 PNSR) \
- The `vanilla Nerf`_ model with 8-layer MLPs can be trained to *better quality* (+~0.5 PNSR) \
in *1 hour* rather than *1~2 days* as in the paper.
in *1 hour* rather than *1~2 days* as in the paper.
- The `Instant-NGP Nerf`_ model can be trained to *
better
quality*
(+~0.7 PSNR) with *9/10th* of
\
- The `Instant-NGP Nerf`_ model can be trained to *
equal
quality*
in *4.5 minutes*,
\
the training time (4.5 minutes)
comparing to the official pure-CUDA implementation.
comparing to the official pure-CUDA implementation.
- The `D-Nerf`_ model for *dynamic* objects can also be trained in *1 hour* \
- The `D-Nerf`_ model for *dynamic* objects can also be trained in *1 hour* \
rather than *2 days* as in the paper, and with *better quality* (+~2.5 PSNR).
rather than *2 days* as in the paper, and with *better quality* (+~2.5 PSNR).
- Both *bounded* and *unbounded* scenes are supported.
- Both *bounded* and *unbounded* scenes are supported.
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment