- 15 May, 2019 3 commits
-
-
Rachel Lim authored
-
Igor authored
* Set the --clone_model_in_keras_dist_strat to None. Remove the separate no_cloning benchmarks and add a couple of cloning ones. Fixes the learning rate schedule to cache its ops per graph.
-
Rachel Lim authored
* Added 'tfdata_exp' version of all benchmarks which set FLAGS.tf_data_experimental_slack = True. Renamed `data_prefetch_with_slack` to `data_delay_prefetch` (haoyu's change) to make the names more distinct. * Add flag to resnet input pipeline and surface through keras_imagenet_main.py
-
- 11 May, 2019 1 commit
-
-
Toby Boyd authored
* Add FP16 and benchmarks. * add missing run and report. * Add loss_scale as option not included with dtype. * move loss_scale validation under dtype conditional. * add loss_scale to flags tested.
-
- 10 May, 2019 5 commits
-
-
Haoyu Zhang authored
* Fix trivial model to work properly with fp16 * Add comment on manual casting
-
Haoyu Zhang authored
Previously we had one dense layer in trivial model. The weight was [224*224*3, num_classes]. Using two dense layers, the weights are [224*224*3, 1] and [1, num_classes].
-
Haoyu Zhang authored
* Do not report metrics in performance benchmarks * Rename flag
-
Haoyu Zhang authored
-
Haoyu Zhang authored
* Modified tweaked tests to use tensor learning rate
-
- 09 May, 2019 1 commit
-
-
Haoyu Zhang authored
* Add learning rate tensor. This makes training slower * Improve LearningRateSchedule with better efficiency * Fix lint error * Replace constant definition with existing one
-
- 07 May, 2019 1 commit
-
-
Haoyu Zhang authored
-
- 06 May, 2019 1 commit
-
-
Haoyu Zhang authored
-
- 04 May, 2019 1 commit
-
-
Haoyu Zhang authored
* Enable CuDNN BatchNorm spatial persistent by default; Remove 2nd zero padding layer * Apply scale=False and fused=True consistently to BatchNorm layers * Undo remove padding layer * Replace zero padding with padding attribute in max pooling for better performance * Resolve comments * Revert "Replace zero padding with padding attribute in max pooling for better performance" This reverts commit ad49db057c800ecac008eec1057005bd2c08ac73.
-
- 03 May, 2019 1 commit
-
-
Reed authored
-
- 02 May, 2019 1 commit
-
-
Reed authored
-
- 01 May, 2019 1 commit
-
-
Reed authored
This options allows the new tf.train.experimental.enable_mixed_precision_graph_rewrite() function to be used for fp16, instead of manual casts.
-
- 30 Apr, 2019 1 commit
-
-
Toby Boyd authored
-
- 29 Apr, 2019 3 commits
-
-
Igor authored
Replace per_device with per_replica and PerDevice with PerReplica, because the PerDevice concept was renamed and doesn't exist anymore. (#6693) * Replace per_device with per_replica and PerDevice with PerReplica, because the PerReplica concept was renamed and doesn't exist anymore.
-
Shining Sun authored
* bug fix * bug fix
-
Igor authored
* Add benchmarks with the --cloning flag to Resnet and NFC. * Renamed cloning to clone_model_in_keras_dist_strat. Dropped a few tests that aren't essential. * Fixed up the formatting after re-naming the flag to a much longer name. Thanks, lint. * Fixed the lint error in nfc_common.py
-
- 26 Apr, 2019 2 commits
-
-
Toby Boyd authored
* Combined imagenet and cifar-10 benchmarks * Comments and epochs_between_evals. * Added tuned tests and cleaned up benchmark flags * Fix names. * Return results and add images/sec hook. * updated doc strings for return values. * 128 to 256 batch for FP16 test * added more doc strings to fix lint.
-
Ayush Dubey authored
* Add num_packs flag for MirroredStrategy's cross device ops. * fix parens * Fix lint errors and make all_reduce_alg more robust. * Set default num_packs to 1
-
- 25 Apr, 2019 1 commit
-
-
Haoyu Zhang authored
Reason: test failures because contrib is not available in V2 This reverts commit 325dd761.
-
- 24 Apr, 2019 5 commits
-
-
Toby Boyd authored
-
Shining Sun authored
* Added none check for output_dir * Change double quote to single
-
Haoyu Zhang authored
-
Haoyu Zhang authored
-
Haoyu Zhang authored
* Introduce a short sleep before ds.prefetch in tf.data. * Further limit dataset threads to reduce CPU contention * Tuned dataset sleep time * Rename dataset sleep flag; enable it only for Keras Graph mode
-
- 23 Apr, 2019 2 commits
-
-
Toby Boyd authored
* Small word tweak * Few more tweaks
-
Usama Muneeb authored
Added additional information on using the `SavedModel` for prediction purposes.
-
- 22 Apr, 2019 1 commit
-
-
Toby Boyd authored
-
- 18 Apr, 2019 1 commit
-
-
Haoyu Zhang authored
-
- 17 Apr, 2019 4 commits
-
-
Shining Sun authored
* before moving test cases to the base class * Added tests for keras cifar and keras imagenet * fix cifar10_test * add blank lines * fix lint errors * fix lint * Resolve comments * Modified two resnet keras tests * Tests passed * Remove keras_test_base * Remove gpu from the no-dist tests
-
Yuefeng Zhou authored
* Update resnet_run_loop.py * Update resnet_run_loop.py * Update resnet_run_loop.py * Update resnet_run_loop.py * Update resnet_run_loop.py
-
rxsang authored
-
- 12 Apr, 2019 2 commits
-
-
Dong Lin authored
-
Taylor Robie authored
-
- 11 Apr, 2019 1 commit
-
-
rxsang authored
* Revert "Revert " Ensure static shapes when enabling XLA in Resnet Keras model (#6508)" (#6517)" This reverts commit cc9eef76. * Set `batch_size` to keras.Input in non-eager mode. Eager mode currently has OOM problem. * Add comments for enable_eager flag. * Always set drop_remainder=True. * Only set drop_remainder=True for XLA.
-
- 10 Apr, 2019 1 commit
-
-
Vighnesh Birodkar authored
* Refactored ResNet code and added additional architectures. * Added numerical layer names instead of alphabetical. * Change dash to underscore. * Corrected return statement. * Use conv_strides argument. * Set classes=10 * Use partial to reduce code duplication.
-