- 19 Jun, 2019 2 commits
-
-
Toby Boyd authored
* set default steps to 300K. * Log flags to perfzero. * Add XLA support to transformer - Moved config logic to keras_utils - Added enable_xla flag to _performance flags - Did not refactor enable_xla flag from keras resnet due to reliance on calling FLAGs in estimator keras and that is a needed refactor for another time. * fix g3 lint complaint. * Refactor set config into keras_utils. * Move flags out of main. * pipe through enable_xla * Update official/transformer/v2/misc.py Co-Authored-By:Reed <reedwm@google.com>
-
Toby Boyd authored
-
- 14 Jun, 2019 2 commits
- 13 Jun, 2019 1 commit
-
-
Toby Boyd authored
-
- 10 Jun, 2019 1 commit
-
-
rxsang authored
-
- 06 Jun, 2019 3 commits
-
-
Reed authored
Before, there was a global default loss scale for all models. Currently, only resnet uses loss scaling, but this will be useful once more models support it.
-
Haoyu Zhang authored
-
Haoyu Zhang authored
* Modify tweaked tests for better performance in no cloning mode * Tweak trivial models
-
- 05 Jun, 2019 1 commit
-
-
rxsang authored
-
- 03 Jun, 2019 2 commits
-
-
Haoyu Zhang authored
Because we run warmup tests in all real data benchmarks, XLA bugs will cause non-XLA tests to fail as well.
-
Toby Boyd authored
* Add mlperf like test. * Final comments. * docstring wording tweak. * non-tweaked version
-
- 31 May, 2019 3 commits
-
-
Haoyu Zhang authored
-
Goldie Gadde authored
-
Haoyu Zhang authored
* Support pure eager execution in ResNet50 * Use smaller batch size
-
- 29 May, 2019 1 commit
-
-
Haoyu Zhang authored
-
- 28 May, 2019 2 commits
-
-
Haoyu Zhang authored
-
Haoyu Zhang authored
* Run different numbers of steps on different platforms * Add new tests for delayed performance measurement
-
- 24 May, 2019 3 commits
-
-
rxsang authored
* Add a graph optional_next Reset benchmark. * Fix lint error.
-
Toby Boyd authored
-
Tian Lin authored
* Merged commit includes the following changes: 249776315 by tianlin<tianlin@google.com>: Internal change 249763206 by tianlin<tianlin@google.com>: For TF 2.0 (related to Beam Search), expand cond dims in tf.where(cond, x, y) to make all parameters broadcastable. -- 249392724 by hongkuny<hongkuny@google.com>: Internal change PiperOrigin-RevId: 249776315 * Merged commit includes the following changes: 249823043 by tianlin<tianlin@google.com>: Bring back v2 test for predict and eval. -- PiperOrigin-RevId: 249823043
-
- 23 May, 2019 3 commits
-
-
rxsang authored
* Add a test enabling get_next_as_optional behavior. * Remove repeated flag. * Remove trailing space. * Make the name shorter. * Fix lint error. * Refine the benchmark name.
-
rxsang authored
-
rxsang authored
* Add enable_get_next_as_optional flag. * Set enable_get_next_as_optional to strategy. * Add comments to explain the flag. * Remove trailing whitespace. * Remove trailing space.
-
- 22 May, 2019 1 commit
-
-
Haoyu Zhang authored
-
- 21 May, 2019 1 commit
-
-
Haoyu Zhang authored
-
- 15 May, 2019 2 commits
-
-
Igor authored
* Set the --clone_model_in_keras_dist_strat to None. Remove the separate no_cloning benchmarks and add a couple of cloning ones. Fixes the learning rate schedule to cache its ops per graph.
-
Rachel Lim authored
* Added 'tfdata_exp' version of all benchmarks which set FLAGS.tf_data_experimental_slack = True. Renamed `data_prefetch_with_slack` to `data_delay_prefetch` (haoyu's change) to make the names more distinct. * Add flag to resnet input pipeline and surface through keras_imagenet_main.py
-
- 10 May, 2019 5 commits
-
-
Haoyu Zhang authored
* Fix trivial model to work properly with fp16 * Add comment on manual casting
-
Haoyu Zhang authored
Previously we had one dense layer in trivial model. The weight was [224*224*3, num_classes]. Using two dense layers, the weights are [224*224*3, 1] and [1, num_classes].
-
Haoyu Zhang authored
* Do not report metrics in performance benchmarks * Rename flag
-
Haoyu Zhang authored
-
Haoyu Zhang authored
* Modified tweaked tests to use tensor learning rate
-
- 09 May, 2019 1 commit
-
-
Haoyu Zhang authored
* Add learning rate tensor. This makes training slower * Improve LearningRateSchedule with better efficiency * Fix lint error * Replace constant definition with existing one
-
- 07 May, 2019 1 commit
-
-
Haoyu Zhang authored
-
- 06 May, 2019 1 commit
-
-
Haoyu Zhang authored
-
- 04 May, 2019 1 commit
-
-
Haoyu Zhang authored
* Enable CuDNN BatchNorm spatial persistent by default; Remove 2nd zero padding layer * Apply scale=False and fused=True consistently to BatchNorm layers * Undo remove padding layer * Replace zero padding with padding attribute in max pooling for better performance * Resolve comments * Revert "Replace zero padding with padding attribute in max pooling for better performance" This reverts commit ad49db057c800ecac008eec1057005bd2c08ac73.
-
- 30 Apr, 2019 1 commit
-
-
Toby Boyd authored
-
- 29 Apr, 2019 2 commits
-
-
Shining Sun authored
* bug fix * bug fix
-
Igor authored
* Add benchmarks with the --cloning flag to Resnet and NFC. * Renamed cloning to clone_model_in_keras_dist_strat. Dropped a few tests that aren't essential. * Fixed up the formatting after re-naming the flag to a much longer name. Thanks, lint. * Fixed the lint error in nfc_common.py
-