- 30 Mar, 2020 1 commit
-
-
Hongkun Yu authored
PiperOrigin-RevId: 303767122
-
- 25 Mar, 2020 1 commit
-
-
Reed Wanderman-Milne authored
The TF pip packages are always built with CUDA support, so tf.test.is_built_with_cuda() would return True even if the user had no GPU. PiperOrigin-RevId: 302928378
-
- 17 Mar, 2020 1 commit
-
-
ayushmankumar7 authored
-
- 09 Mar, 2020 2 commits
-
-
A. Unique TensorFlower authored
PiperOrigin-RevId: 299759804
-
A. Unique TensorFlower authored
PiperOrigin-RevId: 299757650
-
- 25 Feb, 2020 1 commit
-
-
Reed Wanderman-Milne authored
I plan on changing Keras to return Python floats instead of Numpy scalars, and this will break if .item() is called. Also, using the float constructor is cleaner, as it clearly communicates that a Python float is desired. PiperOrigin-RevId: 297192289
-
- 29 Jan, 2020 1 commit
-
-
Hongkun Yu authored
PiperOrigin-RevId: 292178802
-
- 27 Jan, 2020 1 commit
-
-
Yanhui Liang authored
PiperOrigin-RevId: 291810091
-
- 04 Jan, 2020 1 commit
-
-
A. Unique TensorFlower authored
PiperOrigin-RevId: 288064272
-
- 14 Dec, 2019 2 commits
-
-
Hongkun Yu authored
PiperOrigin-RevId: 285533511
-
A. Unique TensorFlower authored
PiperOrigin-RevId: 285503670
-
- 11 Oct, 2019 3 commits
-
-
Hongkun Yu authored
This reverts commit b4e560dc.
-
Hongkun Yu authored
* Revert "Update tf.contrib.data to tf.data.experimental. (#7650)" This reverts commit faf4bbb3. * revert research
-
Derek Murray authored
-
- 10 Oct, 2019 1 commit
-
-
A. Unique TensorFlower authored
change benchmark's log verbosity to logging.INFO. it seems to me that DEBUG map to ---v=1 internally, which is way to verbose for the purpose of benchmarking. PiperOrigin-RevId: 274040907
-
- 09 Sep, 2019 1 commit
-
-
Reed Wanderman-Milne authored
--stop_threshold, --num_gpu, --hooks, --export_dir, and --distribution_strategy have been unexposed from models which do not use them PiperOrigin-RevId: 268032080
-
- 04 Sep, 2019 1 commit
-
-
Reed Wanderman-Milne authored
--clean, --train_epochs, and --epochs_between_evals have been unexposed from models which do not use them PiperOrigin-RevId: 267065651
-
- 26 Aug, 2019 1 commit
-
-
Reed Wanderman-Milne authored
--synthetic_data, --dtype, --all_reduce_alg, and --num_packs have been unexposed from models which do not use them PiperOrigin-RevId: 265483564
-
- 23 Aug, 2019 1 commit
-
-
Reed Wanderman-Milne authored
--num_parallel_calls, --inter_op_parallelism_threads, and --intra_op_parallelism_threads have been unexposed from models which do not use them PiperOrigin-RevId: 264965788
-
- 20 Aug, 2019 1 commit
-
-
Hongkun Yu authored
PiperOrigin-RevId: 264300408
-
- 19 Aug, 2019 1 commit
-
-
Reed Wanderman-Milne authored
Only the V1 resnet model uses --max_train_steps. This unexposes the flag in the keras_application_models, mnist, keras resnet, CTL resnet Models. Before this change, such models allowed the flag to be specified, but ignored it. I also removed the "max_train" argument from the run_synthetic function, since this only had any meaning for the V1 resnet model. Instead, the V1 resnet model now directly passes --max_train_steps=1 to run_synthetic. PiperOrigin-RevId: 264269836
-
- 16 Aug, 2019 1 commit
-
-
Ayush Dubey authored
Also add `worker_hosts` and `task_index` flags. These flags enable running the model over multiple hosts by passing the cluster information via command line. Setting `TF_CONFIG` will continue to work. PiperOrigin-RevId: 263825245
-
- 01 Aug, 2019 1 commit
-
-
Haoyu Zhang authored
* Restructure resnet estimator code to under official/r1 * Continue moving resnet code... * Improved README.md
-