"src/include/vscode:/vscode.git/clone" did not exist on "12beb62250916daa07ca99912c26daedc8376652"
  1. 21 Jun, 2019 1 commit
  2. 20 Jun, 2019 2 commits
  3. 19 Jun, 2019 2 commits
    • Reed's avatar
      Add mixed precision support to Transformer (#7011) · f8ec01ae
      Reed authored
      f8ec01ae
    • Toby Boyd's avatar
      Add XLA to transformer (#7048) · 269581dc
      Toby Boyd authored
      
      
      * set default steps to 300K.
      
      * Log flags to perfzero.
      
      * Add XLA support to transformer
      
      - Moved config logic to keras_utils
      - Added enable_xla flag to _performance flags
      - Did not refactor enable_xla flag from keras resnet due to
        reliance on calling FLAGs in estimator keras and that is
        a needed refactor for another time.
      
      * fix g3 lint complaint.
      
      * Refactor set config into keras_utils.
      
      * Move flags out of main.
      
      * pipe through enable_xla
      
      * Update official/transformer/v2/misc.py
      Co-Authored-By: default avatarReed <reedwm@google.com>
      269581dc
  4. 18 Jun, 2019 1 commit
  5. 06 Jun, 2019 1 commit
  6. 05 Jun, 2019 7 commits
  7. 31 May, 2019 2 commits
  8. 29 May, 2019 3 commits
  9. 28 May, 2019 5 commits
    • Igor's avatar
      Make 'off' a string literal. · 3928d481
      Igor authored
      3928d481
    • guptapriya's avatar
      Turn dist strat off for 1 GPU benchmarks · 2be9ba5b
      guptapriya authored
      2be9ba5b
    • guptapriya's avatar
      undo shuffle change · df523d91
      guptapriya authored
      this is not going to help with current tf.data semantics. so removing it.
      df523d91
    • Igor's avatar
      Add distribute strategies to transformer. (#6883) · b9c1d1ca
      Igor authored
      * Fixes that make transformer run.
      
      * Remove debug print statements.
      
      * Changed the permissions to 644.
      
      * Fix the rest of the permissions.
      
      * enable static batch in all benchmarks
      
      * Restrict dist strat hack to training mode
      
      For now we will do predict/eval without dist strat, so remove that hack in non training cases.
      
      * Use `inputs` instead of `x` as arg name for call
      
      Keras has different behavior based on whether the inputs are called `inputs` or not. Using `inputs` gives expected behaviors.
      
      * Avoid extra map fn on input in dist strat case
      
      * Update how we handle custom metrics
      
      This new approach works with and without dist strat. The previous one didn't work with dist strat. We need to fix that but this is reasonable in meantime (b/133724664).
      
      * Update benchmarks
      
      * typo in metrics code
      
      * Revert metrics change
      
      Didn't actually work in distributed case..
      b9c1d1ca
    • guptapriya's avatar
      Add shuffle to dataset records · 733a752d
      guptapriya authored
      This shuffling should help in getting shuffling each epoch.
      733a752d
  10. 24 May, 2019 2 commits
    • Toby Boyd's avatar
      Transformer v2 benchmark (#6860) · f2ea2f53
      Toby Boyd authored
      * Moved common keras code to utils.
      
      * Initial 1 gpu benchmark
      
      - Aligned flags with resnet example
      - removed code/features that are not super useful
      - eval as part of train if bleu source/ref provided
      - add exp_per_second hook
      
      * Rename benchmark classes, pass batch-size and log_steps.
      
      * fix docstring
      
      * Predict done with checkpoints inline
      
      - perfzero baseclass
      
      * steps not epochs with smoother training loop.
      
      * do not initialize history outside loop.
      
      * 5000 between eval not 500
      
      * estimator to keras.
      
      * remove epochs var.
      
      * use range not xrange.
      
      * 200K steps for 1 gpu
      
      * fix global step
      f2ea2f53
    • Tian Lin's avatar
      Merged commit that fixes transformer's predict and eval. (#6874) · b9cab01b
      Tian Lin authored
      * Merged commit includes the following changes:
      249776315  by tianlin<tianlin@google.com>:
      
          Internal change
      
      249763206  by tianlin<tianlin@google.com>:
      
          For TF 2.0 (related to Beam Search), expand cond dims in tf.where(cond, x, y) to make all parameters broadcastable.
      
      --
      249392724  by hongkuny<hongkuny@google.com>:
      
          Internal change
      
      PiperOrigin-RevId: 249776315
      
      * Merged commit includes the following changes:
      249823043  by tianlin<tianlin@google.com>:
      
          Bring back v2 test for predict and eval.
      
      --
      
      PiperOrigin-RevId: 249823043
      b9cab01b
  11. 22 May, 2019 2 commits
    • Toby Boyd's avatar
      fix lint issues. (#6855) · 3a97b68c
      Toby Boyd authored
      3a97b68c
    • Tian Lin's avatar
      Merge Transformer V2 to Github (#6846) · c4f34e58
      Tian Lin authored
      * Merged commit includes the following changes:
      249218656  by tianlin<tianlin@google.com>:
      
          Deal with imports, fix a typo and make unit tests fast.
      
      --
      249198645  by tianlin<tianlin@google.com>:
      
          Trivial: Remove one empty line before "import tensorflow"
      
      --
      249195490  by tianlin<tianlin@google.com>:
      
          Initialize Transformer TF V2 Model with Keras subclassing implementation. (Compatible with TF V1)
      
      --
      249195008  by tianlin<tianlin@google.com>:
      
          Internal change
      
      249173564  by hongkuny<hongkuny@google.com>:
      
          Internal change
      
      249079258  by hongkuny<hongkuny@google.com>:
      
          Internal change
      
      247691534  by haoyuzhang<haoyuzhang@google.com>:
      
          Internal change
      
      247533725  by haoyuzhang<haoyuzhang@google.com>:
      
          Internal change
      
      247509295  by haoyuzhang<haoyuzhang@google.com>:
      
          Internal change
      
      247311355  by wangtz<wangtz@google.com>:
      
          Internal change
      
      247303127  by wangtz<wangtz@google.com>:
      
        ...
      c4f34e58