1. 11 Oct, 2019 2 commits
  2. 16 Aug, 2018 1 commit
    • Jules Gagnon-Marchand's avatar
      Deterministic dataset order fix (#5098) · 468d8bb6
      Jules Gagnon-Marchand authored
      * Deterministic dataset order fix
      
      In order for the order of the files to be deterministic, in `tf.data.Dataset.list_files(..., shuffle)`, shuffle needs to be True, otherwise different iterator inits will yield different file orders
      
      * removed unnecessary shuffle of filenames
      
      * Removed the `_FILE_SHUFFLE_BUFFER` definition
      468d8bb6
  3. 12 Jun, 2018 1 commit
    • Katherine Wu's avatar
      Transformer multi gpu, remove multi_gpu flag, distribution helper functions (#4457) · 29c9f985
      Katherine Wu authored
      * Add DistributionStrategy to transformer model
      
      * add num_gpu flag
      
      * Calculate per device batch size for transformer
      
      * remove reference to flags_core
      
      * Add synthetic data option to transformer
      
      * fix typo
      
      * add import back in
      
      * Use hierarchical copy
      
      * address PR comments
      
      * lint
      
      * fix spaces
      
      * group train op together to fix single GPU error
      
      * Fix translate bug (sorted_keys is a dict, not a list)
      
      * Change params to a default dict (translate.py was throwing errors because params didn't have the TPU parameters.)
      
      * Address PR comments. Removed multi gpu flag + more
      
      * fix lint
      
      * fix more lints
      
      * add todo for Synthetic dataset
      
      * Update docs
      29c9f985
  4. 04 Jun, 2018 1 commit
    • Taylor Robie's avatar
      First pass at a TPU loop for Transformer (#4296) · 2eeb85fe
      Taylor Robie authored
      * port changes from previous branch now that transformer util changes are in master
      
      fix incorrect count
      
      correct (hopefully) treatment of batch_size
      
      set eval_metrics to a dummy function for now
      
      add some comments
      
      start bringing metrics to transformer TPU
      
      resolve logits shape
      
      metrics are now working except for tf.py_func metrics
      
      increase batch_size for tpu, and create summary host call
      
      fix host call
      
      reduce tpu default batch size
      
      further tune batch sizes
      
      add minibatch loss to summary
      
      handle case of single_iteration_train_steps > number points in an epoch
      
      begin to incorporate hooks
      
      add sleep workarounds
      
      disable hooks altogether
      
      generalize host call function and move to newly created tpu utils module
      
      remove all traces of params as an object
      
      switch from  to
      
      address some PR comments, and change the number of data points.
      
      minor tweaks
      
      add tpu dry run for testing, and use matmul for TPU embedding
      
      infeed/outfeed queue issue is fixed. Sleeps are no longer necessary
      
      add some documentation.
      
      cleanup and address PR comments
      
      delint
      
      add accelerator __init__
      
      fix embedding
      
      missed PR comment
      
      address PR comments
      
      fix validator bug
      
      rewrite cloud storage validator, and add oauth dependency to requirements.txt
      
      * delint
      2eeb85fe
  5. 11 May, 2018 1 commit
  6. 02 May, 2018 1 commit