Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
ModelZoo
ResNet50_tensorflow
Commits
c646928d
Commit
c646928d
authored
Mar 14, 2016
by
Martin Wicke
Browse files
Merge pull request #14 from moonboots/master
Fix typos and broken link
parents
2e4f31a1
4e7a22e0
Changes
1
Hide whitespace changes
Inline
Side-by-side
Showing
1 changed file
with
8 additions
and
8 deletions
+8
-8
inception/README.md
inception/README.md
+8
-8
No files found.
inception/README.md
View file @
c646928d
...
...
@@ -35,7 +35,7 @@ validation data set.
*
Retraining an Inception v3 network on a novel task and back-propagating the
errors to fine tune the network weights.
The training procedure employs synchronous stochastic gradient desent across
The training procedure employs synchronous stochastic gradient des
c
ent across
multiple GPUs. The user may specify the number of GPUs they wish harness.
The synchronous training performs
*batch-splitting*
by dividing a given batch
across multiple GPUs.
...
...
@@ -70,7 +70,7 @@ encoded) as well as metadata such as label and bounding box information. See
[
`parse_example_proto`
](
image_processing.py
)
for details.
We provide a single
[
script
](
data/download_and_preprocess_imagenet.sh
)
[
script
](
inception/
data/download_and_preprocess_imagenet.sh
)
for downloading and converting ImageNet data to TFRecord format. Downloading
and preprocessing the data may take several hours (up to half a day) depending
on your network and computer speed. Please be patient.
...
...
@@ -80,7 +80,7 @@ To begin, you will need to sign up for an account with
sign up page, create an account and request an access key to download the data.
After you have
`USERNAME`
and
`PASSWORD`
, you are ready to run our script.
Make sure that your hard disk has at least 500 GB of free space for do
n
wloading
Make sure that your hard disk has at least 500 GB of free space for dow
n
loading
and storing the data. Here we select
`DATA_DIR=$HOME/imagenet-data`
as such a
location but feel free to edit accordingly.
...
...
@@ -131,7 +131,7 @@ across the GPUs.
processing a batch of data.
The training procedure is encapsulated by this diagram of how operations and
variables are placed on CPU and GPUs respec
i
tvely.
variables are placed on CPU and GPUs respect
i
vely.
<div
style=
"width:40%; margin:auto; margin-bottom:10px; margin-top:20px;"
>
<img
style=
"width:100%"
src=
"https://www.tensorflow.org/images/Parallelism.png"
>
...
...
@@ -590,7 +590,7 @@ NUM_EPOCHS_PER_DECAY = 30.0 # Epochs after which learning rate decays.
LEARNING_RATE_DECAY_FACTOR
=
0.16
# Learning rate decay factor.
```
There are many papers tha discuss the various tricks and trade-offs associated
There are many papers tha
t
discuss the various tricks and trade-offs associated
with training a model with stochastic gradient descent. For those new to the
field, some great references are:
...
...
@@ -600,7 +600,7 @@ field, some great references are:
What follows is a summary of some general advice for identifying appropriate
model hyper-parameters in the context of this particular
model training setup. Namely,
this library provides
*sy
c
nhronous*
updates to model parameters based on
this library provides
*syn
c
hronous*
updates to model parameters based on
batch-splitting the model across multiple GPUs.
*
Higher learning rates leads to faster training. Too high of learning rate
...
...
@@ -641,7 +641,7 @@ CPU memory would be ideal.
If that is not possible, you can tune down the memory demands of the model
via lowering
`--input_queue_memory_factor`
. Images are preprocessed
asyncronously with respect to the main training across
async
h
ronously with respect to the main training across
`--num_preprocess_threads`
threads. The preprocessed images are stored in
shuffling queue in which each GPU performs a dequeue operation in order
to receive a
`batch_size`
worth of images.
...
...
@@ -651,7 +651,7 @@ shuffling queue of 1024 x `input_queue_memory_factor` images. For the current
model architecture, this corresponds to 16GB of CPU memory. You may lower
`input_queue_memory_factor`
in order to decrease the memory footprint. Keep
in mind though that lowering this value drastically may result in a model
with sligh
l
ty lower predictive accuracy when training from scratch. Please
with slight
l
y lower predictive accuracy when training from scratch. Please
see comments in
[
`image_processing.py`
](
./image_processing.py
)
for more details.
## Troubleshooting
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment