1. 18 Aug, 2018 1 commit
    • Reed's avatar
      Speed up cache construction. (#5131) · 5aee67b4
      Reed authored
      This is done by using a higher Pickle protocol version, which the Python docs describe as being "slightly more efficient". This reduces the file write time at the beginning from 2 1/2 minutes to 5 seconds.
      5aee67b4
  2. 02 Aug, 2018 2 commits
  3. 01 Aug, 2018 1 commit
  4. 31 Jul, 2018 8 commits
  5. 30 Jul, 2018 1 commit
    • Taylor Robie's avatar
      NCF pipeline refactor (take 2) and initial TPU port. (#4935) · 6518c1c7
      Taylor Robie authored
      * intermediate commit
      
      * ncf now working
      
      * reorder pipeline
      
      * allow batched decode for file backed dataset
      
      * fix bug
      
      * more tweaks
      
      * parallize false negative generation
      
      * shared pool hack
      
      * workers ignore sigint
      
      * intermediate commit
      
      * simplify buffer backed dataset creation to fixed length record approach only. (more cleanup needed)
      
      * more tweaks
      
      * simplify pipeline
      
      * fix misplaced cleanup() calls. (validation works\!)
      
      * more tweaks
      
      * sixify memoryview usage
      
      * more sixification
      
      * fix bug
      
      * add future imports
      
      * break up training input pipeline
      
      * more pipeline tuning
      
      * first pass at moving negative generation to async
      
      * refactor async pipeline to use files instead of ipc
      
      * refactor async pipeline
      
      * move expansion and concatenation from reduce worker to generation workers
      
      * abandon complete async due to interactions with the tensorflow threadpool
      
      * cleanup
      
      * remove performance_comparison.py
      
      * experiment with rough generator + interleave pipeline
      
      * yet more pipeline tuning
      
      * update on-the-fly pipeline
      
      * refactor preprocessing, and move train generation behind a GRPC server
      
      * fix leftover call
      
      * intermediate commit
      
      * intermediate commit
      
      * fix index error in data pipeline, and add logging to train data server
      
      * make sharding more robust to imbalance
      
      * correctly sample with replacement
      
      * file buffers are no longer needed for this branch
      
      * tweak sampling methods
      
      * add README for data pipeline
      
      * fix eval sampling, and vectorize eval metrics
      
      * add spillover and static training batch sizes
      
      * clean up cruft from earlier iterations
      
      * rough delint
      
      * delint 2 / n
      
      * add type annotations
      
      * update run script
      
      * make run.sh a bit nicer
      
      * change embedding initializer to match reference
      
      * rough pass at pure estimator model_fn
      
      * impose static shape hack (revisit later)
      
      * refinements
      
      * fix dir error in run.sh
      
      * add documentation
      
      * add more docs and fix an assert
      
      * old data test is no longer valid. Keeping it around as reference for the new one
      
      * rough draft of data pipeline validation script
      
      * don't rely on shuffle default
      
      * tweaks and documentation
      
      * add separate eval batch size for performance
      
      * initial commit
      
      * terrible hacking
      
      * mini hacks
      
      * missed a bug
      
      * messing about trying to get TPU running
      
      * TFRecords based TPU attempt
      
      * bug fixes
      
      * don't log remotely
      
      * more bug fixes
      
      * TPU tweaks and bug fixes
      
      * more tweaks
      
      * more adjustments
      
      * rework model definition
      
      * tweak data pipeline
      
      * refactor async TFRecords generation
      
      * temp commit to run.sh
      
      * update log behavior
      
      * fix logging bug
      
      * add check for subprocess start to avoid cryptic hangs
      
      * unify deserialize and make it TPU compliant
      
      * delint
      
      * remove gRPC pipeline code
      
      * fix logging bug
      
      * delint and remove old test files
      
      * add unit tests for NCF pipeline
      
      * delint
      
      * clean up run.sh, and add run_tpu.sh
      
      * forgot the most important line
      
      * fix run.sh bugs
      
      * yet more bash debugging
      
      * small tweak to add keras summaries to model_fn
      
      * Clean up sixification issues
      
      * address PR comments
      
      * delinting is never over
      6518c1c7
  6. 12 Jul, 2018 1 commit
  7. 25 Jun, 2018 1 commit
  8. 20 Jun, 2018 1 commit
    • Taylor Robie's avatar
      Wide Deep refactor and deep movies (#4506) · 20070ca4
      Taylor Robie authored
      * begin branch
      
      * finish download script
      
      * rename download to dataset
      
      * intermediate commit
      
      * intermediate commit
      
      * misc tweaks
      
      * intermediate commit
      
      * intermediate commit
      
      * intermediate commit
      
      * delint and update census test.
      
      * add movie tests
      
      * delint
      
      * fix py2 issue
      
      * address PR comments
      
      * intermediate commit
      
      * intermediate commit
      
      * intermediate commit
      
      * finish wide deep transition to vanilla movielens
      
      * delint
      
      * intermediate commit
      
      * intermediate commit
      
      * intermediate commit
      
      * intermediate commit
      
      * fix import
      
      * add default ncf csv construction
      
      * change default on download_if_missing
      
      * shard and vectorize example serialization
      
      * fix import
      
      * update ncf data unittests
      
      * delint
      
      * delint
      
      * more delinting
      
      * fix wide-deep movielens serialization
      
      * address PR comments
      
      * add file_io tests
      
      * investigate wide-deep test failure
      
      * remove hard coded path and properly use flags.
      
      * address file_io test PR comments
      
      * missed a hash_bucked_size
      20070ca4
  9. 12 Jun, 2018 1 commit
    • Katherine Wu's avatar
      Transformer multi gpu, remove multi_gpu flag, distribution helper functions (#4457) · 29c9f985
      Katherine Wu authored
      * Add DistributionStrategy to transformer model
      
      * add num_gpu flag
      
      * Calculate per device batch size for transformer
      
      * remove reference to flags_core
      
      * Add synthetic data option to transformer
      
      * fix typo
      
      * add import back in
      
      * Use hierarchical copy
      
      * address PR comments
      
      * lint
      
      * fix spaces
      
      * group train op together to fix single GPU error
      
      * Fix translate bug (sorted_keys is a dict, not a list)
      
      * Change params to a default dict (translate.py was throwing errors because params didn't have the TPU parameters.)
      
      * Address PR comments. Removed multi gpu flag + more
      
      * fix lint
      
      * fix more lints
      
      * add todo for Synthetic dataset
      
      * Update docs
      29c9f985
  10. 08 Jun, 2018 1 commit
  11. 01 Jun, 2018 2 commits
    • Qianli Scott Zhu's avatar
      Add new test ID and test env info to the benchmark run. (#4426) · d2d6ab4c
      Qianli Scott Zhu authored
      * Add new test ID and test env info to the benchmark run.
      
      * Fix test.
      
      * Fix lint
      
      * Address review comment.
      d2d6ab4c
    • Qianli Scott Zhu's avatar
      Record the status for a benchmark run. (#4402) · 47c5642e
      Qianli Scott Zhu authored
      * Update benchmark logger to update the run status.
      
      This is important for streaming upload to bigquery so that the
      dashboard can ignore the 'running' benchmark at the moment since
      its not finished yet.
      
      * Move the run status into a separate table.
      
      Also update the run status in the benchmark uploader and
      BigqueryBenchmarkLogger.
      
      * Insert instead of update for the benchmark status for file logger.
      
      * Address review comments.
      
      Update the logger to have benchmark context, which will update the
      run status accordingly.
      
      * Fix broken tests.
      
      * Move the benchmark logger context to main function.
      
      * Fix tests.
      
      * Update the rest of the models to use the context in main.
      
      * Delint.
      47c5642e
  12. 25 May, 2018 1 commit
  13. 22 May, 2018 1 commit
    • Yanhui Liang's avatar
      Add recommendation model (#4175) · 81d77669
      Yanhui Liang authored
      * Add recommendation model
      
      * Fix pylints check error
      
      * Rename file
      
      * Address comments, update input pipeline, and add distribution strategy
      
      * Fix import error
      
      * Address more comments
      
      * Fix lints
      81d77669