1. 11 Oct, 2019 2 commits
  2. 09 Sep, 2019 1 commit
  3. 04 Sep, 2019 1 commit
  4. 26 Aug, 2019 1 commit
  5. 23 Aug, 2019 1 commit
  6. 20 Aug, 2019 1 commit
  7. 19 Aug, 2019 1 commit
    • Reed Wanderman-Milne's avatar
      Do not expose --max_train_steps in models that do not use it. · 824ff2d6
      Reed Wanderman-Milne authored
      Only the V1 resnet model uses --max_train_steps. This unexposes the flag in the keras_application_models, mnist, keras resnet, CTL resnet Models. Before this change, such models allowed the flag to be specified, but ignored it.
      
      I also removed the "max_train" argument from the run_synthetic function, since this only had any meaning for the V1 resnet model. Instead, the V1 resnet model now directly passes --max_train_steps=1 to run_synthetic.
      
      PiperOrigin-RevId: 264269836
      824ff2d6
  8. 16 Aug, 2019 1 commit
    • Ayush Dubey's avatar
      Add multi-worker benchmarks to Keras ResNet model. · ff6c3b1e
      Ayush Dubey authored
      Also add `worker_hosts` and `task_index` flags.  These flags enable running the
      model over multiple hosts by passing the cluster information via command line.
      
      Setting `TF_CONFIG` will continue to work.
      
      PiperOrigin-RevId: 263825245
      ff6c3b1e
  9. 01 Aug, 2019 1 commit
  10. 21 Jun, 2019 1 commit
    • Toby Boyd's avatar
      NCF XLA and Eager tests with a refactor of resnet flags to make this cleaner. (#7067) · a68f65f8
      Toby Boyd authored
      * XLA FP32 and first test
      
      * More XLA benchmarks FP32.
      
      * Add eager to NCF and refactor resnet.
      
      * fix v2_0 calls and more flag refactor.
      
      * Remove extra flag args.
      
      * 90 epoch default
      
      * add return
      
      * remove xla not used by estimator.
      
      * Remove duplicate run_eagerly.
      
      * fix flag defaults.
      
      * Remove fp16_implementation flag option.
      
      * Remove stop early on mlperf test.
      
      * remove unneeded args.
      
      * load flags from keras mains.
      a68f65f8
  11. 19 Jun, 2019 1 commit
    • Toby Boyd's avatar
      Add XLA to transformer (#7048) · 269581dc
      Toby Boyd authored
      
      
      * set default steps to 300K.
      
      * Log flags to perfzero.
      
      * Add XLA support to transformer
      
      - Moved config logic to keras_utils
      - Added enable_xla flag to _performance flags
      - Did not refactor enable_xla flag from keras resnet due to
        reliance on calling FLAGs in estimator keras and that is
        a needed refactor for another time.
      
      * fix g3 lint complaint.
      
      * Refactor set config into keras_utils.
      
      * Move flags out of main.
      
      * pipe through enable_xla
      
      * Update official/transformer/v2/misc.py
      Co-Authored-By: default avatarReed <reedwm@google.com>
      269581dc
  12. 14 Jun, 2019 1 commit
  13. 06 Jun, 2019 1 commit
  14. 18 May, 2019 1 commit
  15. 15 May, 2019 2 commits
  16. 11 May, 2019 1 commit
  17. 01 May, 2019 1 commit
    • Reed's avatar
      Add --fp16_implementation option. (#6703) · b691578c
      Reed authored
      This options allows the new tf.train.experimental.enable_mixed_precision_graph_rewrite() function to be used for fp16, instead of manual casts.
      b691578c
  18. 29 Apr, 2019 1 commit
  19. 26 Apr, 2019 2 commits
  20. 18 Apr, 2019 1 commit
  21. 17 Apr, 2019 1 commit
  22. 11 Apr, 2019 1 commit
  23. 03 Apr, 2019 2 commits
  24. 02 Apr, 2019 1 commit
  25. 30 Mar, 2019 1 commit
  26. 28 Mar, 2019 2 commits
  27. 19 Mar, 2019 1 commit
  28. 12 Mar, 2019 1 commit
  29. 11 Mar, 2019 1 commit
  30. 07 Mar, 2019 2 commits
  31. 22 Feb, 2019 2 commits
  32. 21 Feb, 2019 1 commit
    • Ayush Dubey's avatar
      Multi-worker support for Resnet. (#6206) · f2e90945
      Ayush Dubey authored
      * Update official resnet for multi worker training with distribution strategies.
      
      * Fixes for multi worker training.
      
      * Fix call to `get_distribution_strategy`.
      
      * Undo test change.
      
      * Fix spacing.
      
      * Move cluster configuration to distribution_utils.
      
      * Move train_and_evaluate out of loop.  Also, update docstrings for multi-worker flags and add use_train_and_evaluate flag.
      
      * Update distribution_strategy flag to match exported name for collective strategy.
      f2e90945
  33. 19 Feb, 2019 1 commit