Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
ModelZoo
ResNet50_tensorflow
Commits
4e7a22e0
Commit
4e7a22e0
authored
Mar 13, 2016
by
Jack Zhang
Committed by
Jack Zhang
Mar 13, 2016
Browse files
Fix some typos
parent
cee4aff1
Changes
1
Hide whitespace changes
Inline
Side-by-side
Showing
1 changed file
with
7 additions
and
7 deletions
+7
-7
inception/README.md
inception/README.md
+7
-7
No files found.
inception/README.md
View file @
4e7a22e0
...
@@ -35,7 +35,7 @@ validation data set.
...
@@ -35,7 +35,7 @@ validation data set.
*
Retraining an Inception v3 network on a novel task and back-propagating the
*
Retraining an Inception v3 network on a novel task and back-propagating the
errors to fine tune the network weights.
errors to fine tune the network weights.
The training procedure employs synchronous stochastic gradient desent across
The training procedure employs synchronous stochastic gradient des
c
ent across
multiple GPUs. The user may specify the number of GPUs they wish harness.
multiple GPUs. The user may specify the number of GPUs they wish harness.
The synchronous training performs
*batch-splitting*
by dividing a given batch
The synchronous training performs
*batch-splitting*
by dividing a given batch
across multiple GPUs.
across multiple GPUs.
...
@@ -80,7 +80,7 @@ To begin, you will need to sign up for an account with
...
@@ -80,7 +80,7 @@ To begin, you will need to sign up for an account with
sign up page, create an account and request an access key to download the data.
sign up page, create an account and request an access key to download the data.
After you have
`USERNAME`
and
`PASSWORD`
, you are ready to run our script.
After you have
`USERNAME`
and
`PASSWORD`
, you are ready to run our script.
Make sure that your hard disk has at least 500 GB of free space for do
n
wloading
Make sure that your hard disk has at least 500 GB of free space for dow
n
loading
and storing the data. Here we select
`DATA_DIR=$HOME/imagenet-data`
as such a
and storing the data. Here we select
`DATA_DIR=$HOME/imagenet-data`
as such a
location but feel free to edit accordingly.
location but feel free to edit accordingly.
...
@@ -131,7 +131,7 @@ across the GPUs.
...
@@ -131,7 +131,7 @@ across the GPUs.
processing a batch of data.
processing a batch of data.
The training procedure is encapsulated by this diagram of how operations and
The training procedure is encapsulated by this diagram of how operations and
variables are placed on CPU and GPUs respec
i
tvely.
variables are placed on CPU and GPUs respect
i
vely.
<div
style=
"width:40%; margin:auto; margin-bottom:10px; margin-top:20px;"
>
<div
style=
"width:40%; margin:auto; margin-bottom:10px; margin-top:20px;"
>
<img
style=
"width:100%"
src=
"https://www.tensorflow.org/images/Parallelism.png"
>
<img
style=
"width:100%"
src=
"https://www.tensorflow.org/images/Parallelism.png"
>
...
@@ -590,7 +590,7 @@ NUM_EPOCHS_PER_DECAY = 30.0 # Epochs after which learning rate decays.
...
@@ -590,7 +590,7 @@ NUM_EPOCHS_PER_DECAY = 30.0 # Epochs after which learning rate decays.
LEARNING_RATE_DECAY_FACTOR
=
0.16
# Learning rate decay factor.
LEARNING_RATE_DECAY_FACTOR
=
0.16
# Learning rate decay factor.
```
```
There are many papers tha discuss the various tricks and trade-offs associated
There are many papers tha
t
discuss the various tricks and trade-offs associated
with training a model with stochastic gradient descent. For those new to the
with training a model with stochastic gradient descent. For those new to the
field, some great references are:
field, some great references are:
...
@@ -600,7 +600,7 @@ field, some great references are:
...
@@ -600,7 +600,7 @@ field, some great references are:
What follows is a summary of some general advice for identifying appropriate
What follows is a summary of some general advice for identifying appropriate
model hyper-parameters in the context of this particular
model hyper-parameters in the context of this particular
model training setup. Namely,
model training setup. Namely,
this library provides
*sy
c
nhronous*
updates to model parameters based on
this library provides
*syn
c
hronous*
updates to model parameters based on
batch-splitting the model across multiple GPUs.
batch-splitting the model across multiple GPUs.
*
Higher learning rates leads to faster training. Too high of learning rate
*
Higher learning rates leads to faster training. Too high of learning rate
...
@@ -641,7 +641,7 @@ CPU memory would be ideal.
...
@@ -641,7 +641,7 @@ CPU memory would be ideal.
If that is not possible, you can tune down the memory demands of the model
If that is not possible, you can tune down the memory demands of the model
via lowering
`--input_queue_memory_factor`
. Images are preprocessed
via lowering
`--input_queue_memory_factor`
. Images are preprocessed
asyncronously with respect to the main training across
async
h
ronously with respect to the main training across
`--num_preprocess_threads`
threads. The preprocessed images are stored in
`--num_preprocess_threads`
threads. The preprocessed images are stored in
shuffling queue in which each GPU performs a dequeue operation in order
shuffling queue in which each GPU performs a dequeue operation in order
to receive a
`batch_size`
worth of images.
to receive a
`batch_size`
worth of images.
...
@@ -651,7 +651,7 @@ shuffling queue of 1024 x `input_queue_memory_factor` images. For the current
...
@@ -651,7 +651,7 @@ shuffling queue of 1024 x `input_queue_memory_factor` images. For the current
model architecture, this corresponds to 16GB of CPU memory. You may lower
model architecture, this corresponds to 16GB of CPU memory. You may lower
`input_queue_memory_factor`
in order to decrease the memory footprint. Keep
`input_queue_memory_factor`
in order to decrease the memory footprint. Keep
in mind though that lowering this value drastically may result in a model
in mind though that lowering this value drastically may result in a model
with sligh
l
ty lower predictive accuracy when training from scratch. Please
with slight
l
y lower predictive accuracy when training from scratch. Please
see comments in
[
`image_processing.py`
](
./image_processing.py
)
for more details.
see comments in
[
`image_processing.py`
](
./image_processing.py
)
for more details.
## Troubleshooting
## Troubleshooting
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment