Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
ModelZoo
ResNet50_tensorflow
Commits
1bbd359d
Commit
1bbd359d
authored
Apr 12, 2021
by
Reed Wanderman-Milne
Committed by
A. Unique TensorFlower
Apr 12, 2021
Browse files
Internal change
PiperOrigin-RevId: 368067415
parent
ecbc3fd8
Changes
10
Show whitespace changes
Inline
Side-by-side
Showing
10 changed files
with
12 additions
and
30 deletions
+12
-30
official/vision/beta/projects/deepmac_maskrcnn/train.py
official/vision/beta/projects/deepmac_maskrcnn/train.py
+1
-3
official/vision/beta/projects/simclr/tasks/simclr.py
official/vision/beta/projects/simclr/tasks/simclr.py
+2
-4
official/vision/beta/projects/simclr/train.py
official/vision/beta/projects/simclr/train.py
+1
-3
official/vision/beta/projects/yolo/train.py
official/vision/beta/projects/yolo/train.py
+1
-3
official/vision/beta/projects/yt8m/train.py
official/vision/beta/projects/yt8m/train.py
+1
-3
official/vision/beta/train.py
official/vision/beta/train.py
+1
-3
official/vision/beta/train_spatial_partitioning.py
official/vision/beta/train_spatial_partitioning.py
+1
-3
official/vision/image_classification/classifier_trainer.py
official/vision/image_classification/classifier_trainer.py
+2
-4
official/vision/image_classification/resnet/resnet_ctl_imagenet_main.py
...n/image_classification/resnet/resnet_ctl_imagenet_main.py
+1
-2
official/vision/image_classification/resnet/resnet_runnable.py
...ial/vision/image_classification/resnet/resnet_runnable.py
+1
-2
No files found.
official/vision/beta/projects/deepmac_maskrcnn/train.py
View file @
1bbd359d
...
@@ -48,9 +48,7 @@ def main(_):
...
@@ -48,9 +48,7 @@ def main(_):
# GPUs, and bfloat16 in the case of TPUs. loss_scale takes effect only when
# GPUs, and bfloat16 in the case of TPUs. loss_scale takes effect only when
# dtype is float16
# dtype is float16
if
params
.
runtime
.
mixed_precision_dtype
:
if
params
.
runtime
.
mixed_precision_dtype
:
performance
.
set_mixed_precision_policy
(
params
.
runtime
.
mixed_precision_dtype
,
performance
.
set_mixed_precision_policy
(
params
.
runtime
.
mixed_precision_dtype
)
params
.
runtime
.
loss_scale
,
use_experimental_api
=
True
)
distribution_strategy
=
distribute_utils
.
get_distribution_strategy
(
distribution_strategy
=
distribute_utils
.
get_distribution_strategy
(
distribution_strategy
=
params
.
runtime
.
distribution_strategy
,
distribution_strategy
=
params
.
runtime
.
distribution_strategy
,
all_reduce_alg
=
params
.
runtime
.
all_reduce_alg
,
all_reduce_alg
=
params
.
runtime
.
all_reduce_alg
,
...
...
official/vision/beta/projects/simclr/tasks/simclr.py
View file @
1bbd359d
...
@@ -91,8 +91,7 @@ class SimCLRPretrainTask(base_task.Task):
...
@@ -91,8 +91,7 @@ class SimCLRPretrainTask(base_task.Task):
optimizer
=
performance
.
configure_optimizer
(
optimizer
=
performance
.
configure_optimizer
(
optimizer
,
optimizer
,
use_float16
=
runtime_config
.
mixed_precision_dtype
==
'float16'
,
use_float16
=
runtime_config
.
mixed_precision_dtype
==
'float16'
,
loss_scale
=
runtime_config
.
loss_scale
,
loss_scale
=
runtime_config
.
loss_scale
)
use_experimental_api
=
True
)
return
optimizer
return
optimizer
...
@@ -397,8 +396,7 @@ class SimCLRFinetuneTask(base_task.Task):
...
@@ -397,8 +396,7 @@ class SimCLRFinetuneTask(base_task.Task):
optimizer
=
performance
.
configure_optimizer
(
optimizer
=
performance
.
configure_optimizer
(
optimizer
,
optimizer
,
use_float16
=
runtime_config
.
mixed_precision_dtype
==
'float16'
,
use_float16
=
runtime_config
.
mixed_precision_dtype
==
'float16'
,
loss_scale
=
runtime_config
.
loss_scale
,
loss_scale
=
runtime_config
.
loss_scale
)
use_experimental_api
=
True
)
return
optimizer
return
optimizer
...
...
official/vision/beta/projects/simclr/train.py
View file @
1bbd359d
...
@@ -59,9 +59,7 @@ def main(_):
...
@@ -59,9 +59,7 @@ def main(_):
# GPUs, and bfloat16 in the case of TPUs. loss_scale takes effect only when
# GPUs, and bfloat16 in the case of TPUs. loss_scale takes effect only when
# dtype is float16
# dtype is float16
if
params
.
runtime
.
mixed_precision_dtype
:
if
params
.
runtime
.
mixed_precision_dtype
:
performance
.
set_mixed_precision_policy
(
params
.
runtime
.
mixed_precision_dtype
,
performance
.
set_mixed_precision_policy
(
params
.
runtime
.
mixed_precision_dtype
)
params
.
runtime
.
loss_scale
,
use_experimental_api
=
True
)
distribution_strategy
=
distribute_utils
.
get_distribution_strategy
(
distribution_strategy
=
distribute_utils
.
get_distribution_strategy
(
distribution_strategy
=
params
.
runtime
.
distribution_strategy
,
distribution_strategy
=
params
.
runtime
.
distribution_strategy
,
all_reduce_alg
=
params
.
runtime
.
all_reduce_alg
,
all_reduce_alg
=
params
.
runtime
.
all_reduce_alg
,
...
...
official/vision/beta/projects/yolo/train.py
View file @
1bbd359d
...
@@ -50,9 +50,7 @@ def main(_):
...
@@ -50,9 +50,7 @@ def main(_):
# GPUs, and bfloat16 in the case of TPUs. loss_scale takes effect only when
# GPUs, and bfloat16 in the case of TPUs. loss_scale takes effect only when
# dtype is float16
# dtype is float16
if
params
.
runtime
.
mixed_precision_dtype
:
if
params
.
runtime
.
mixed_precision_dtype
:
performance
.
set_mixed_precision_policy
(
params
.
runtime
.
mixed_precision_dtype
,
performance
.
set_mixed_precision_policy
(
params
.
runtime
.
mixed_precision_dtype
)
params
.
runtime
.
loss_scale
,
use_experimental_api
=
True
)
distribution_strategy
=
distribute_utils
.
get_distribution_strategy
(
distribution_strategy
=
distribute_utils
.
get_distribution_strategy
(
distribution_strategy
=
params
.
runtime
.
distribution_strategy
,
distribution_strategy
=
params
.
runtime
.
distribution_strategy
,
all_reduce_alg
=
params
.
runtime
.
all_reduce_alg
,
all_reduce_alg
=
params
.
runtime
.
all_reduce_alg
,
...
...
official/vision/beta/projects/yt8m/train.py
View file @
1bbd359d
...
@@ -46,9 +46,7 @@ def main(_):
...
@@ -46,9 +46,7 @@ def main(_):
# GPUs, and bfloat16 in the case of TPUs. loss_scale takes effect only when
# GPUs, and bfloat16 in the case of TPUs. loss_scale takes effect only when
# dtype is float16
# dtype is float16
if
params
.
runtime
.
mixed_precision_dtype
:
if
params
.
runtime
.
mixed_precision_dtype
:
performance
.
set_mixed_precision_policy
(
params
.
runtime
.
mixed_precision_dtype
,
performance
.
set_mixed_precision_policy
(
params
.
runtime
.
mixed_precision_dtype
)
params
.
runtime
.
loss_scale
,
use_experimental_api
=
True
)
distribution_strategy
=
distribute_utils
.
get_distribution_strategy
(
distribution_strategy
=
distribute_utils
.
get_distribution_strategy
(
distribution_strategy
=
params
.
runtime
.
distribution_strategy
,
distribution_strategy
=
params
.
runtime
.
distribution_strategy
,
all_reduce_alg
=
params
.
runtime
.
all_reduce_alg
,
all_reduce_alg
=
params
.
runtime
.
all_reduce_alg
,
...
...
official/vision/beta/train.py
View file @
1bbd359d
...
@@ -46,9 +46,7 @@ def main(_):
...
@@ -46,9 +46,7 @@ def main(_):
# GPUs, and bfloat16 in the case of TPUs. loss_scale takes effect only when
# GPUs, and bfloat16 in the case of TPUs. loss_scale takes effect only when
# dtype is float16
# dtype is float16
if
params
.
runtime
.
mixed_precision_dtype
:
if
params
.
runtime
.
mixed_precision_dtype
:
performance
.
set_mixed_precision_policy
(
params
.
runtime
.
mixed_precision_dtype
,
performance
.
set_mixed_precision_policy
(
params
.
runtime
.
mixed_precision_dtype
)
params
.
runtime
.
loss_scale
,
use_experimental_api
=
True
)
distribution_strategy
=
distribute_utils
.
get_distribution_strategy
(
distribution_strategy
=
distribute_utils
.
get_distribution_strategy
(
distribution_strategy
=
params
.
runtime
.
distribution_strategy
,
distribution_strategy
=
params
.
runtime
.
distribution_strategy
,
all_reduce_alg
=
params
.
runtime
.
all_reduce_alg
,
all_reduce_alg
=
params
.
runtime
.
all_reduce_alg
,
...
...
official/vision/beta/train_spatial_partitioning.py
View file @
1bbd359d
...
@@ -96,9 +96,7 @@ def main(_):
...
@@ -96,9 +96,7 @@ def main(_):
# GPUs, and bfloat16 in the case of TPUs. loss_scale takes effect only when
# GPUs, and bfloat16 in the case of TPUs. loss_scale takes effect only when
# dtype is float16
# dtype is float16
if
params
.
runtime
.
mixed_precision_dtype
:
if
params
.
runtime
.
mixed_precision_dtype
:
performance
.
set_mixed_precision_policy
(
params
.
runtime
.
mixed_precision_dtype
,
performance
.
set_mixed_precision_policy
(
params
.
runtime
.
mixed_precision_dtype
)
params
.
runtime
.
loss_scale
,
use_experimental_api
=
True
)
input_partition_dims
=
None
input_partition_dims
=
None
if
FLAGS
.
mode
==
'train_and_eval'
:
if
FLAGS
.
mode
==
'train_and_eval'
:
...
...
official/vision/image_classification/classifier_trainer.py
View file @
1bbd359d
...
@@ -227,8 +227,7 @@ def initialize(params: base_configs.ExperimentConfig,
...
@@ -227,8 +227,7 @@ def initialize(params: base_configs.ExperimentConfig,
dataset_builder
:
dataset_factory
.
DatasetBuilder
):
dataset_builder
:
dataset_factory
.
DatasetBuilder
):
"""Initializes backend related initializations."""
"""Initializes backend related initializations."""
keras_utils
.
set_session_config
(
enable_xla
=
params
.
runtime
.
enable_xla
)
keras_utils
.
set_session_config
(
enable_xla
=
params
.
runtime
.
enable_xla
)
performance
.
set_mixed_precision_policy
(
dataset_builder
.
dtype
,
performance
.
set_mixed_precision_policy
(
dataset_builder
.
dtype
)
use_experimental_api
=
False
)
if
tf
.
config
.
list_physical_devices
(
'GPU'
):
if
tf
.
config
.
list_physical_devices
(
'GPU'
):
data_format
=
'channels_first'
data_format
=
'channels_first'
else
:
else
:
...
@@ -341,8 +340,7 @@ def train_and_eval(
...
@@ -341,8 +340,7 @@ def train_and_eval(
optimizer
=
performance
.
configure_optimizer
(
optimizer
=
performance
.
configure_optimizer
(
optimizer
,
optimizer
,
use_float16
=
train_builder
.
dtype
==
'float16'
,
use_float16
=
train_builder
.
dtype
==
'float16'
,
loss_scale
=
get_loss_scale
(
params
),
loss_scale
=
get_loss_scale
(
params
))
use_experimental_api
=
True
)
metrics_map
=
_get_metrics
(
one_hot
)
metrics_map
=
_get_metrics
(
one_hot
)
metrics
=
[
metrics_map
[
metric
]
for
metric
in
params
.
train
.
metrics
]
metrics
=
[
metrics_map
[
metric
]
for
metric
in
params
.
train
.
metrics
]
...
...
official/vision/image_classification/resnet/resnet_ctl_imagenet_main.py
View file @
1bbd359d
...
@@ -99,8 +99,7 @@ def run(flags_obj):
...
@@ -99,8 +99,7 @@ def run(flags_obj):
"""
"""
keras_utils
.
set_session_config
(
keras_utils
.
set_session_config
(
enable_xla
=
flags_obj
.
enable_xla
)
enable_xla
=
flags_obj
.
enable_xla
)
performance
.
set_mixed_precision_policy
(
flags_core
.
get_tf_dtype
(
flags_obj
),
performance
.
set_mixed_precision_policy
(
flags_core
.
get_tf_dtype
(
flags_obj
))
use_experimental_api
=
False
)
if
tf
.
config
.
list_physical_devices
(
'GPU'
):
if
tf
.
config
.
list_physical_devices
(
'GPU'
):
if
flags_obj
.
tf_gpu_thread_mode
:
if
flags_obj
.
tf_gpu_thread_mode
:
...
...
official/vision/image_classification/resnet/resnet_runnable.py
View file @
1bbd359d
...
@@ -81,8 +81,7 @@ class ResnetRunnable(orbit.StandardTrainer, orbit.StandardEvaluator):
...
@@ -81,8 +81,7 @@ class ResnetRunnable(orbit.StandardTrainer, orbit.StandardEvaluator):
self
.
optimizer
,
self
.
optimizer
,
use_float16
=
self
.
dtype
==
tf
.
float16
,
use_float16
=
self
.
dtype
==
tf
.
float16
,
use_graph_rewrite
=
use_graph_rewrite
,
use_graph_rewrite
=
use_graph_rewrite
,
loss_scale
=
flags_core
.
get_loss_scale
(
flags_obj
,
default_for_fp16
=
128
),
loss_scale
=
flags_core
.
get_loss_scale
(
flags_obj
,
default_for_fp16
=
128
))
use_experimental_api
=
False
)
self
.
train_loss
=
tf
.
keras
.
metrics
.
Mean
(
'train_loss'
,
dtype
=
tf
.
float32
)
self
.
train_loss
=
tf
.
keras
.
metrics
.
Mean
(
'train_loss'
,
dtype
=
tf
.
float32
)
self
.
train_accuracy
=
tf
.
keras
.
metrics
.
SparseCategoricalAccuracy
(
self
.
train_accuracy
=
tf
.
keras
.
metrics
.
SparseCategoricalAccuracy
(
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment