Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
ModelZoo
ResNet50_tensorflow
Commits
0102dfb6
Commit
0102dfb6
authored
May 12, 2016
by
Vincent Ohprecio
Committed by
Martin Wicke
May 12, 2016
Browse files
fixed script relative link (#58)
parent
bf60abf8
Changes
1
Hide whitespace changes
Inline
Side-by-side
Showing
1 changed file
with
10 additions
and
10 deletions
+10
-10
inception/README.md
inception/README.md
+10
-10
No files found.
inception/README.md
View file @
0102dfb6
...
@@ -50,7 +50,7 @@ primary differences with that setup are:
...
@@ -50,7 +50,7 @@ primary differences with that setup are:
language called TensorFlow-Slim.
language called TensorFlow-Slim.
For more details about TensorFlow-Slim, please see the [Slim README]
For more details about TensorFlow-Slim, please see the [Slim README]
(slim/README.md). Please note that this higher-level language is still
(
inception/
slim/README.md). Please note that this higher-level language is still
*experimental*
and the API may change over time depending on usage and
*experimental*
and the API may change over time depending on usage and
subsequent research.
subsequent research.
...
@@ -66,9 +66,9 @@ and convert the ImageNet data to native TFRecord format. The TFRecord format
...
@@ -66,9 +66,9 @@ and convert the ImageNet data to native TFRecord format. The TFRecord format
consists of a set of sharded files where each entry is a serialized
`tf.Example`
consists of a set of sharded files where each entry is a serialized
`tf.Example`
proto. Each
`tf.Example`
proto contains the ImageNet image (JPEG encoded) as
proto. Each
`tf.Example`
proto contains the ImageNet image (JPEG encoded) as
well as metadata such as label and bounding box information. See
well as metadata such as label and bounding box information. See
[
`parse_example_proto`
](
image_processing.py
)
for details.
[
`parse_example_proto`
](
inception/
image_processing.py
)
for details.
We provide a single
[
script
](
data/download_and_preprocess_imagenet.sh
)
for
We provide a single
[
script
](
inception/
data/download_and_preprocess_imagenet.sh
)
for
downloading and converting ImageNet data to TFRecord format. Downloading and
downloading and converting ImageNet data to TFRecord format. Downloading and
preprocessing the data may take several hours (up to half a day) depending on
preprocessing the data may take several hours (up to half a day) depending on
your network and computer speed. Please be patient.
your network and computer speed. Please be patient.
...
@@ -444,7 +444,7 @@ There is a single automated script that downloads the data set and converts it
...
@@ -444,7 +444,7 @@ There is a single automated script that downloads the data set and converts it
to the TFRecord format. Much like the ImageNet data set, each record in the
to the TFRecord format. Much like the ImageNet data set, each record in the
TFRecord format is a serialized
`tf.Example`
proto whose entries include a
TFRecord format is a serialized
`tf.Example`
proto whose entries include a
JPEG-encoded string and an integer label. Please see [
`parse_example_proto`
]
JPEG-encoded string and an integer label. Please see [
`parse_example_proto`
]
(image_processing.py) for details.
(
inception/
image_processing.py) for details.
The script just takes a few minutes to run depending your network connection
The script just takes a few minutes to run depending your network connection
speed for downloading and processing the images. Your hard disk requires 200MB
speed for downloading and processing the images. Your hard disk requires 200MB
...
@@ -474,10 +474,10 @@ files in the `DATA_DIR`. The files will match the patterns `train-????-of-00001`
...
@@ -474,10 +474,10 @@ files in the `DATA_DIR`. The files will match the patterns `train-????-of-00001`
and
`validation-?????-of-00001`
, respectively.
and
`validation-?????-of-00001`
, respectively.
**NOTE**
If you wish to prepare a custom image data set for transfer learning,
**NOTE**
If you wish to prepare a custom image data set for transfer learning,
you will need to invoke
[
`build_image_data.py`
](
data/build_image_data.py
)
on
you will need to invoke
[
`build_image_data.py`
](
inception/
data/build_image_data.py
)
on
your custom data set. Please see the associated options and assumptions behind
your custom data set. Please see the associated options and assumptions behind
this script by reading the comments section of [
`build_image_data.py`
]
this script by reading the comments section of [
`build_image_data.py`
]
(data/build_image_data.py).
(
inception/
data/build_image_data.py).
The second piece you will need is a trained Inception v3 image model. You have
The second piece you will need is a trained Inception v3 image model. You have
the option of either training one yourself (See [How to Train from Scratch]
the option of either training one yourself (See [How to Train from Scratch]
...
@@ -607,7 +607,7 @@ Succesfully loaded model from /tmp/flowers/model.ckpt-1999 at step=1999.
...
@@ -607,7 +607,7 @@ Succesfully loaded model from /tmp/flowers/model.ckpt-1999 at step=1999.
One can use the existing scripts supplied with this model to build a new dataset
One can use the existing scripts supplied with this model to build a new dataset
for training or fine-tuning. The main script to employ is
for training or fine-tuning. The main script to employ is
[
`build_image_data.py`
](
.
/build_image_data.py
)
. Briefly, this script takes a
[
`build_image_data.py`
](
inception/data
/build_image_data.py
)
. Briefly, this script takes a
structured directory of images and converts it to a sharded
`TFRecord`
that can
structured directory of images and converts it to a sharded
`TFRecord`
that can
be read by the Inception model.
be read by the Inception model.
...
@@ -714,7 +714,7 @@ considerations for novices.
...
@@ -714,7 +714,7 @@ considerations for novices.
Roughly 5-10 hyper-parameters govern the speed at which a network is trained. In
Roughly 5-10 hyper-parameters govern the speed at which a network is trained. In
addition to
`--batch_size`
and
`--num_gpus`
, there are several constants defined
addition to
`--batch_size`
and
`--num_gpus`
, there are several constants defined
in
[
inception_train.py
](
.
/inception_train.py
)
which dictate the learning
in
[
inception_train.py
](
inception
/inception_train.py
)
which dictate the learning
schedule.
schedule.
```
shell
```
shell
...
@@ -788,7 +788,7 @@ model architecture, this corresponds to about 4GB of CPU memory. You may lower
...
@@ -788,7 +788,7 @@ model architecture, this corresponds to about 4GB of CPU memory. You may lower
`input_queue_memory_factor`
in order to decrease the memory footprint. Keep in
`input_queue_memory_factor`
in order to decrease the memory footprint. Keep in
mind though that lowering this value drastically may result in a model with
mind though that lowering this value drastically may result in a model with
slightly lower predictive accuracy when training from scratch. Please see
slightly lower predictive accuracy when training from scratch. Please see
comments in
[
`image_processing.py`
](
.
/image_processing.py
)
for more details.
comments in
[
`image_processing.py`
](
inception
/image_processing.py
)
for more details.
## Troubleshooting
## Troubleshooting
...
@@ -824,7 +824,7 @@ input image size, then you may need to redesign the entire model architecture.
...
@@ -824,7 +824,7 @@ input image size, then you may need to redesign the entire model architecture.
We targeted a desktop with 128GB of CPU ram connected to 8 NVIDIA Tesla K40 GPU
We targeted a desktop with 128GB of CPU ram connected to 8 NVIDIA Tesla K40 GPU
cards but we have run this on desktops with 32GB of CPU ram and 1 NVIDIA Tesla
cards but we have run this on desktops with 32GB of CPU ram and 1 NVIDIA Tesla
K40. You can get a sense of the various training configurations we tested by
K40. You can get a sense of the various training configurations we tested by
reading the comments in
[
`inception_train.py`
](
.
/inception_train.py
)
.
reading the comments in
[
`inception_train.py`
](
inception
/inception_train.py
)
.
#### How do I continue training from a checkpoint in distributed setting?
#### How do I continue training from a checkpoint in distributed setting?
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment