Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
OpenDAS
nerfacc
Commits
62e70d39
Unverified
Commit
62e70d39
authored
Oct 12, 2022
by
Ruilong Li(李瑞龙)
Committed by
GitHub
Oct 12, 2022
Browse files
update ngp perf. (#68)
parent
76c0f981
Changes
6
Show whitespace changes
Inline
Side-by-side
Showing
6 changed files
with
30 additions
and
25 deletions
+30
-25
README.md
README.md
+4
-4
docs/source/examples/dnerf.rst
docs/source/examples/dnerf.rst
+1
-1
docs/source/examples/ngp.rst
docs/source/examples/ngp.rst
+21
-16
docs/source/examples/unbounded.rst
docs/source/examples/unbounded.rst
+1
-1
docs/source/examples/vanilla.rst
docs/source/examples/vanilla.rst
+1
-1
docs/source/index.rst
docs/source/index.rst
+2
-2
No files found.
README.md
View file @
62e70d39
...
...
@@ -12,8 +12,8 @@ Using NerfAcc,
-
The
`vanilla NeRF`
model with 8-layer MLPs can be trained to
*better quality*
(+~0.5 PNSR)
in
*1 hour*
rather than
*days*
as in the paper.
-
The
`Instant-NGP NeRF`
model can be trained to
*
better
quality*
(+~0.7 PSNR) with
*9/10th*
of
the training time (4.5 minutes)
comparing to the official pure-CUDA implementation.
-
The
`Instant-NGP NeRF`
model can be trained to
*
equal
quality*
in
*4.5 minutes*
,
comparing to the official pure-CUDA implementation.
-
The
`D-NeRF`
model for
*dynamic*
objects can also be trained in
*1 hour*
rather than
*2 days*
as in the paper, and with
*better quality*
(+~2.5 PSNR).
-
Both
*bounded*
and
*unbounded*
scenes are supported.
...
...
@@ -100,9 +100,9 @@ Before running those example scripts, please check the script about which datase
the dataset first.
```
bash
# Instant-NGP NeRF in 4.5 minutes with
better
performance!
# Instant-NGP NeRF in 4.5 minutes with
reproduced
performance!
# See results at here: https://www.nerfacc.com/en/latest/examples/ngp.html
python examples/train_ngp_nerf.py
--train_split
train
val
--scene
lego
python examples/train_ngp_nerf.py
--train_split
train
--scene
lego
```
```
bash
...
...
docs/source/examples/dnerf.rst
View file @
62e70d39
...
...
@@ -32,5 +32,5 @@ single NVIDIA TITAN RTX GPU. The training memory footprint is about 11GB.
.. _`D-Nerf`: https://arxiv.org/abs/2011.13961
.. _`D-Nerf dataset`: https://www.dropbox.com/s/0bf6fl0ye2vz3vr/data.zip?dl=0
.. _`github repository`: https://github.com/KAIR-BAIR/nerfacc/tree/
5637cc9a1565b2685c02250eb1ee1c53d3b07af1
.. _`github repository`: https://github.com/KAIR-BAIR/nerfacc/tree/
76c0f9817da4c9c8b5ccf827eb069ee2ce854b75
docs/source/examples/ngp.rst
View file @
62e70d39
...
...
@@ -7,29 +7,34 @@ See code `examples/train_ngp_nerf.py` at our `github repository`_ for details.
Benchmarks
------------
*updated on 2022-10-
08
*
*updated on 2022-10-
12
*
Here we trained a `Instant-NGP Nerf`_ model on the `Nerf-Synthetic dataset`_. We follow the same
settings with the Instant-NGP paper, which uses train
val
split for training and test split for
evaluation.
Our
experiments are conducted on a single NVIDIA TITAN RTX GPU. The training
settings with the Instant-NGP paper, which uses train split for training and test split for
evaluation.
All
experiments are conducted on a single NVIDIA TITAN RTX GPU. The training
memory footprint is about 3GB.
.. note::
The Instant-NGP paper makes use of the alpha channel in the images to apply random background
augmentation during training. Yet we only uses RGB values with a constant white background.
augmentation during training. For fair comparision, we rerun their code with a constant white
background during both training and testing. Also it is worth to mention that we didn't strictly
follow the training receipe in the Instant-NGP paper, such as the learning rate schedule etc, as
the purpose of this benchmark is to showcase instead of reproducing the paper.
+----------------------+-------+-------+---------+-------+-------+-------+-------+-------+-------+
+----------------------
-
+-------+-------+---------+-------+-------+-------+-------+-------+-------+
| PSNR | Lego | Mic |Materials| Chair |Hotdog | Ficus | Drums | Ship | MEAN |
| | | | | | | | | | |
+======================+=======+=======+=========+=======+=======+=======+=======+=======+=======+
| Instant-NGP (5min) | 36.39 | 36.22 | 29.78 | 35.00 | 37.40 | 33.51 | 26.02 | 31.10 | 33.18 |
+----------------------+-------+-------+---------+-------+-------+-------+-------+-------+-------+
| Ours (~4.5min) | 36.82 | 37.61 | 30.18 | 36.13 | 38.11 | 34.48 | 26.62 | 31.37 | 33.92 |
+----------------------+-------+-------+---------+-------+-------+-------+-------+-------+-------+
| Ours (Training time)| 288s | 259s | 256s | 324s | 288s | 245s | 262s | 257s | 272s |
+----------------------+-------+-------+---------+-------+-------+-------+-------+-------+-------+
.. _`Instant-NGP Nerf`: https://arxiv.org/abs/2201.05989
.. _`github repository`: https://github.com/KAIR-BAIR/nerfacc/tree/5637cc9a1565b2685c02250eb1ee1c53d3b07af1
+=======================+=======+=======+=========+=======+=======+=======+=======+=======+=======+
|Instant-NGP 35k steps | 35.87 | 36.22 | 29.08 | 35.10 | 37.48 | 30.61 | 23.85 | 30.62 | 32.35 |
+-----------------------+-------+-------+---------+-------+-------+-------+-------+-------+-------+
|(training time) | 309s | 258s | 256s | 316s | 292s | 207s | 218s | 250s | 263s |
+-----------------------+-------+-------+---------+-------+-------+-------+-------+-------+-------+
|Ours 20k steps | 35.50 | 36.16 | 29.14 | 35.23 | 37.15 | 31.71 | 24.88 | 29.91 | 32.46 |
+-----------------------+-------+-------+---------+-------+-------+-------+-------+-------+-------+
|(training time) | 287s | 274s | 269s | 317s | 269s | 244s | 249s | 257s | 271s |
+-----------------------+-------+-------+---------+-------+-------+-------+-------+-------+-------+
.. _`Instant-NGP Nerf`: https://github.com/NVlabs/instant-ngp/tree/51e4107edf48338e9ab0316d56a222e0adf87143
.. _`github repository`: https://github.com/KAIR-BAIR/nerfacc/tree/76c0f9817da4c9c8b5ccf827eb069ee2ce854b75
.. _`Nerf-Synthetic dataset`: https://drive.google.com/drive/folders/1JDdLGDruGNXWnM1eqY1FNL9PlStjaKWi
docs/source/examples/unbounded.rst
View file @
62e70d39
...
...
@@ -40,4 +40,4 @@ that takes from `MipNerf360`_.
.. _`Instant-NGP Nerf`: https://arxiv.org/abs/2201.05989
.. _`MipNerf360`: https://arxiv.org/abs/2111.12077
.. _`Nerf++`: https://arxiv.org/abs/2010.07492
.. _`github repository`: https://github.com/KAIR-BAIR/nerfacc/tree/
5637cc9a1565b2685c02250eb1ee1c53d3b07af1
.. _`github repository`: https://github.com/KAIR-BAIR/nerfacc/tree/
76c0f9817da4c9c8b5ccf827eb069ee2ce854b75
docs/source/examples/vanilla.rst
View file @
62e70d39
...
...
@@ -29,5 +29,5 @@ conducted on a single NVIDIA TITAN RTX GPU. The training memory footprint is abo
| Ours (Training time)| 58min | 53min | 46min | 62min | 56min | 42min | 52min | 49min | 52min |
+----------------------+-------+-------+---------+-------+-------+-------+-------+-------+-------+
.. _`github repository`: : https://github.com/KAIR-BAIR/nerfacc/tree/
5637cc9a1565b2685c02250eb1ee1c53d3b07af1
.. _`github repository`: : https://github.com/KAIR-BAIR/nerfacc/tree/
76c0f9817da4c9c8b5ccf827eb069ee2ce854b75
.. _`vanilla Nerf`: https://arxiv.org/abs/2003.08934
docs/source/index.rst
View file @
62e70d39
...
...
@@ -8,8 +8,8 @@ Using NerfAcc,
- The `vanilla Nerf`_ model with 8-layer MLPs can be trained to *better quality* (+~0.5 PNSR) \
in *1 hour* rather than *1~2 days* as in the paper.
- The `Instant-NGP Nerf`_ model can be trained to *
better
quality*
(+~0.7 PSNR) with *9/10th* of
\
the training time (4.5 minutes)
comparing to the official pure-CUDA implementation.
- The `Instant-NGP Nerf`_ model can be trained to *
equal
quality*
in *4.5 minutes*,
\
comparing to the official pure-CUDA implementation.
- The `D-Nerf`_ model for *dynamic* objects can also be trained in *1 hour* \
rather than *2 days* as in the paper, and with *better quality* (+~2.5 PSNR).
- Both *bounded* and *unbounded* scenes are supported.
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment