1. 30 Aug, 2019 1 commit
  2. 22 Aug, 2019 1 commit
  3. 20 Aug, 2019 1 commit
  4. 16 Aug, 2019 1 commit
  5. 15 Aug, 2019 1 commit
  6. 09 Aug, 2019 1 commit
  7. 23 Jul, 2019 1 commit
  8. 21 Jun, 2019 1 commit
  9. 19 Jun, 2019 2 commits
    • Reed's avatar
      Add mixed precision support to Transformer (#7011) · f8ec01ae
      Reed authored
      f8ec01ae
    • Toby Boyd's avatar
      Add XLA to transformer (#7048) · 269581dc
      Toby Boyd authored
      
      
      * set default steps to 300K.
      
      * Log flags to perfzero.
      
      * Add XLA support to transformer
      
      - Moved config logic to keras_utils
      - Added enable_xla flag to _performance flags
      - Did not refactor enable_xla flag from keras resnet due to
        reliance on calling FLAGs in estimator keras and that is
        a needed refactor for another time.
      
      * fix g3 lint complaint.
      
      * Refactor set config into keras_utils.
      
      * Move flags out of main.
      
      * pipe through enable_xla
      
      * Update official/transformer/v2/misc.py
      Co-Authored-By: default avatarReed <reedwm@google.com>
      269581dc
  10. 18 Jun, 2019 1 commit
  11. 29 May, 2019 1 commit
  12. 24 May, 2019 2 commits
    • Toby Boyd's avatar
      Transformer v2 benchmark (#6860) · f2ea2f53
      Toby Boyd authored
      * Moved common keras code to utils.
      
      * Initial 1 gpu benchmark
      
      - Aligned flags with resnet example
      - removed code/features that are not super useful
      - eval as part of train if bleu source/ref provided
      - add exp_per_second hook
      
      * Rename benchmark classes, pass batch-size and log_steps.
      
      * fix docstring
      
      * Predict done with checkpoints inline
      
      - perfzero baseclass
      
      * steps not epochs with smoother training loop.
      
      * do not initialize history outside loop.
      
      * 5000 between eval not 500
      
      * estimator to keras.
      
      * remove epochs var.
      
      * use range not xrange.
      
      * 200K steps for 1 gpu
      
      * fix global step
      f2ea2f53
    • Tian Lin's avatar
      Merged commit that fixes transformer's predict and eval. (#6874) · b9cab01b
      Tian Lin authored
      * Merged commit includes the following changes:
      249776315  by tianlin<tianlin@google.com>:
      
          Internal change
      
      249763206  by tianlin<tianlin@google.com>:
      
          For TF 2.0 (related to Beam Search), expand cond dims in tf.where(cond, x, y) to make all parameters broadcastable.
      
      --
      249392724  by hongkuny<hongkuny@google.com>:
      
          Internal change
      
      PiperOrigin-RevId: 249776315
      
      * Merged commit includes the following changes:
      249823043  by tianlin<tianlin@google.com>:
      
          Bring back v2 test for predict and eval.
      
      --
      
      PiperOrigin-RevId: 249823043
      b9cab01b
  13. 22 May, 2019 1 commit
    • Tian Lin's avatar
      Merge Transformer V2 to Github (#6846) · c4f34e58
      Tian Lin authored
      * Merged commit includes the following changes:
      249218656  by tianlin<tianlin@google.com>:
      
          Deal with imports, fix a typo and make unit tests fast.
      
      --
      249198645  by tianlin<tianlin@google.com>:
      
          Trivial: Remove one empty line before "import tensorflow"
      
      --
      249195490  by tianlin<tianlin@google.com>:
      
          Initialize Transformer TF V2 Model with Keras subclassing implementation. (Compatible with TF V1)
      
      --
      249195008  by tianlin<tianlin@google.com>:
      
          Internal change
      
      249173564  by hongkuny<hongkuny@google.com>:
      
          Internal change
      
      249079258  by hongkuny<hongkuny@google.com>:
      
          Internal change
      
      247691534  by haoyuzhang<haoyuzhang@google.com>:
      
          Internal change
      
      247533725  by haoyuzhang<haoyuzhang@google.com>:
      
          Internal change
      
      247509295  by haoyuzhang<haoyuzhang@google.com>:
      
          Internal change
      
      247311355  by wangtz<wangtz@google.com>:
      
          Internal change
      
      247303127  by wangtz<wangtz@google.com>:
      
        ...
      c4f34e58