- 11 May, 2019 1 commit
-
-
Toby Boyd authored
* Add FP16 and benchmarks. * add missing run and report. * Add loss_scale as option not included with dtype. * move loss_scale validation under dtype conditional. * add loss_scale to flags tested.
-
- 10 May, 2019 5 commits
-
-
Haoyu Zhang authored
* Fix trivial model to work properly with fp16 * Add comment on manual casting
-
Haoyu Zhang authored
Previously we had one dense layer in trivial model. The weight was [224*224*3, num_classes]. Using two dense layers, the weights are [224*224*3, 1] and [1, num_classes].
-
Haoyu Zhang authored
* Do not report metrics in performance benchmarks * Rename flag
-
Haoyu Zhang authored
-
Haoyu Zhang authored
* Modified tweaked tests to use tensor learning rate
-
- 09 May, 2019 2 commits
-
-
Toby Boyd authored
* Add first benchmark and return stats. * Remove print statements update training steps. * Revert print T: in print statement. * Remove print(stats) * add 2 gpu accuracy test for base. * Fixed total_batch_size when using gpu + gFile deprecations. * 8 GPU test name fix * Add 4 and 8 GPU tests. * typo fixes. * Clean up test names and methods. * bleu uncased. docstring format fix.
-
Haoyu Zhang authored
* Add learning rate tensor. This makes training slower * Improve LearningRateSchedule with better efficiency * Fix lint error * Replace constant definition with existing one
-
- 08 May, 2019 1 commit
-
-
Toby Boyd authored
-
- 07 May, 2019 2 commits
-
-
Haoyu Zhang authored
-
Toby Boyd authored
-
- 06 May, 2019 1 commit
-
-
Haoyu Zhang authored
-
- 04 May, 2019 1 commit
-
-
Haoyu Zhang authored
* Enable CuDNN BatchNorm spatial persistent by default; Remove 2nd zero padding layer * Apply scale=False and fused=True consistently to BatchNorm layers * Undo remove padding layer * Replace zero padding with padding attribute in max pooling for better performance * Resolve comments * Revert "Replace zero padding with padding attribute in max pooling for better performance" This reverts commit ad49db057c800ecac008eec1057005bd2c08ac73.
-
- 03 May, 2019 1 commit
-
-
Reed authored
-
- 02 May, 2019 1 commit
-
-
Reed authored
-
- 01 May, 2019 1 commit
-
-
Reed authored
This options allows the new tf.train.experimental.enable_mixed_precision_graph_rewrite() function to be used for fp16, instead of manual casts.
-
- 30 Apr, 2019 1 commit
-
-
Toby Boyd authored
-
- 29 Apr, 2019 5 commits
-
-
Igor authored
Replace per_device with per_replica and PerDevice with PerReplica, because the PerDevice concept was renamed and doesn't exist anymore. (#6693) * Replace per_device with per_replica and PerDevice with PerReplica, because the PerReplica concept was renamed and doesn't exist anymore.
-
Toby Boyd authored
* Add accuracy check. * Avoid double flag init, move data_dir to real data. * Comment on lower accuracy target.
-
Shining Sun authored
* bug fix * bug fix
-
Igor authored
* Add benchmarks with the --cloning flag to Resnet and NFC. * Renamed cloning to clone_model_in_keras_dist_strat. Dropped a few tests that aren't essential. * Fixed up the formatting after re-naming the flag to a much longer name. Thanks, lint. * Fixed the lint error in nfc_common.py
-
Songyi Blair Han authored
-
- 26 Apr, 2019 3 commits
-
-
Toby Boyd authored
* Combined imagenet and cifar-10 benchmarks * Comments and epochs_between_evals. * Added tuned tests and cleaned up benchmark flags * Fix names. * Return results and add images/sec hook. * updated doc strings for return values. * 128 to 256 batch for FP16 test * added more doc strings to fix lint.
-
Ayush Dubey authored
* Add num_packs flag for MirroredStrategy's cross device ops. * fix parens * Fix lint errors and make all_reduce_alg more robust. * Set default num_packs to 1
-
Gaurav Jain authored
tf.test.is_gpu_available() should not be called in flags since this is called before app.main() and the runtime has not yet been initialized.
-
- 25 Apr, 2019 2 commits
-
-
Ayush Dubey authored
* Remove contrib AllReduceCrossDeviceOps and update all_reduce_alg options with MirroredStrategy. * cleanup
-
Haoyu Zhang authored
Reason: test failures because contrib is not available in V2 This reverts commit 325dd761.
-
- 24 Apr, 2019 6 commits
-
-
Toby Boyd authored
-
Shining Sun authored
* Added none check for output_dir * Change double quote to single
-
Haoyu Zhang authored
-
Yuefeng Zhou authored
-
Haoyu Zhang authored
-
Haoyu Zhang authored
* Introduce a short sleep before ds.prefetch in tf.data. * Further limit dataset threads to reduce CPU contention * Tuned dataset sleep time * Rename dataset sleep flag; enable it only for Keras Graph mode
-
- 23 Apr, 2019 2 commits
-
-
Toby Boyd authored
* Small word tweak * Few more tweaks
-
Usama Muneeb authored
Added additional information on using the `SavedModel` for prediction purposes.
-
- 22 Apr, 2019 3 commits
-
-
Toby Boyd authored
* Use tf.image.resize_with_crop_or_pad * exp_per_second and hr_at_10
-
Shining Sun authored
-
Toby Boyd authored
-
- 20 Apr, 2019 2 commits
-
-
Shining Sun authored
-
Shining Sun authored
* Remove contrib imports, or move them inline * Use exposed API for FixedLenFeature * Replace tf.logging with absl logging * Change GFile to v2 APIs * replace tf.logging with absl loggin in movielens * Fixing an import bug * Change gfile to v2 APIs in code * Swap to keras optimizer v2 * Bug fix for optimizer * Change tf.log to tf.keras.backend.log * Change the loss function to keras loss * convert another loss to keras loss * Resolve comments and fix lint * Add a doc string * Fix existing tests and add new tests for DS * Added tests for multi-replica * Fix lint * resolve comments * make estimator run in tf2.0 * use compat v1 loss * fix lint issue
-