- 23 Jul, 2019 1 commit
-
-
Toby Boyd authored
* Add force_run_distributed tests. * Added enable_eager * r/force_run_distributed/force_v2_in_keras_compile * Adding force_v2 tests and FLAGs. * Rename method to avoid conflict. * Add cpu force_v2 tests. * fix lint, wrap line. * change to force_v2_in_keras_compile * Update method name. * Lower mlperf target to 0.736.
-
- 20 Jul, 2019 1 commit
-
-
Zongwei Zhou authored
-
- 11 Jul, 2019 3 commits
- 08 Jul, 2019 1 commit
-
-
Toby Boyd authored
* reduce iterations from 20 to 12. * add fp16 dynamic batch accuracy check. * fix existing lint issue.
-
- 03 Jul, 2019 1 commit
-
-
Toby Boyd authored
* Fix unit tests failures. * 96% of TF 2.0 tests on GPU are passing. * Currently all passing GPU and CPU TF 2.0 * Address code comments. * use tf 2.0 cast. * Comment about working on TF 2.0 CPU * Uses contrib turn off for TF 2.0. * Fix wide_deep and add keras_common_tests. * use context to get num_gpus. * Switch to tf.keras.metrics
-
- 28 Jun, 2019 1 commit
-
-
Toby Boyd authored
-
- 22 Jun, 2019 1 commit
-
-
Toby Boyd authored
-
- 21 Jun, 2019 2 commits
-
-
guptapriya authored
* trying fake merge call * make metrics optional * Remove extra print
-
Reed authored
-
- 20 Jun, 2019 2 commits
- 19 Jun, 2019 2 commits
-
-
Reed authored
-
Toby Boyd authored
* set default steps to 300K. * Log flags to perfzero. * Add XLA support to transformer - Moved config logic to keras_utils - Added enable_xla flag to _performance flags - Did not refactor enable_xla flag from keras resnet due to reliance on calling FLAGs in estimator keras and that is a needed refactor for another time. * fix g3 lint complaint. * Refactor set config into keras_utils. * Move flags out of main. * pipe through enable_xla * Update official/transformer/v2/misc.py Co-Authored-By:Reed <reedwm@google.com>
-
- 18 Jun, 2019 1 commit
-
-
Toby Boyd authored
-
- 11 Jun, 2019 1 commit
-
-
saberkun authored
252534787 by hongkuny<hongkuny@google.com>: Transformer vocab fix to strip correctly in py2 -- PiperOrigin-RevId: 252534787
-
- 06 Jun, 2019 3 commits
-
-
Reed authored
-
guptapriya authored
-
saberkun authored
251762562 by hongkuny<hongkuny@google.com>: Fix blue score inconsistency -- PiperOrigin-RevId: 251762562
-
- 05 Jun, 2019 7 commits
-
-
guptapriya authored
-
guptapriya authored
-
guptapriya authored
-
guptapriya authored
-
guptapriya authored
-
guptapriya authored
-
guptapriya authored
-
- 31 May, 2019 2 commits
-
-
Haoyu Zhang authored
-
Haoyu Zhang authored
* Fix various lint errors * Fix logging format
-
- 29 May, 2019 4 commits
-
-
Zhang Xunkai authored
* Make max_length and static_batch configurable. * Fix line length. * Fix incorrect parameters in building eval input. * Improve comments for readability.
-
guptapriya authored
-
guptapriya authored
-
guptapriya authored
-
- 28 May, 2019 6 commits
-
-
guptapriya authored
* Add static batch benchmarks to estimator So we can distinguish how much static vs dynamic batch matter. * change max_length for static_batch tests * Add flag for max length
-
Igor authored
-
guptapriya authored
-
guptapriya authored
this is not going to help with current tf.data semantics. so removing it.
-
Igor authored
* Fixes that make transformer run. * Remove debug print statements. * Changed the permissions to 644. * Fix the rest of the permissions. * enable static batch in all benchmarks * Restrict dist strat hack to training mode For now we will do predict/eval without dist strat, so remove that hack in non training cases. * Use `inputs` instead of `x` as arg name for call Keras has different behavior based on whether the inputs are called `inputs` or not. Using `inputs` gives expected behaviors. * Avoid extra map fn on input in dist strat case * Update how we handle custom metrics This new approach works with and without dist strat. The previous one didn't work with dist strat. We need to fix that but this is reasonable in meantime (b/133724664). * Update benchmarks * typo in metrics code * Revert metrics change Didn't actually work in distributed case..
-
guptapriya authored
This shuffling should help in getting shuffling each epoch.
-
- 24 May, 2019 1 commit
-
-
Toby Boyd authored
* Moved common keras code to utils. * Initial 1 gpu benchmark - Aligned flags with resnet example - removed code/features that are not super useful - eval as part of train if bleu source/ref provided - add exp_per_second hook * Rename benchmark classes, pass batch-size and log_steps. * fix docstring * Predict done with checkpoints inline - perfzero baseclass * steps not epochs with smoother training loop. * do not initialize history outside loop. * 5000 between eval not 500 * estimator to keras. * remove epochs var. * use range not xrange. * 200K steps for 1 gpu * fix global step
-