- 20 Mar, 2018 1 commit
-
-
Karmel Allison authored
* Glint everything * Adding rcfile and pylinting * Extra newline * Few last lints
-
- 24 Feb, 2018 1 commit
-
-
Neal Wu authored
-
- 08 Feb, 2018 1 commit
-
-
Neal Wu authored
-
- 31 Jan, 2018 2 commits
-
-
Mark Daoust authored
The `sparse` version is more efficient anyway. I'm returning the labels shape [1] instead of [] because tf.accuracy fails otherwise.
-
Neal Wu authored
-
- 26 Jan, 2018 1 commit
-
-
Mark Daoust authored
The `sparse` version is more efficient anyway. I'm returning the labels shape [1] instead of [] because tf.accuracy fails otherwise.
-
- 22 Jan, 2018 1 commit
-
-
Neal Wu authored
-
- 18 Jan, 2018 1 commit
-
-
Asim Shankar authored
-
- 12 Jan, 2018 1 commit
-
-
Mikalai Drabovich authored
Opening gzipped datasets in binary, read-only mode fixes the issue
-
- 06 Jan, 2018 1 commit
-
-
Asim Shankar authored
This is a step towards merging the example in https://github.com/tensorflow/tpu-demos/tree/master/cloud_tpu/models/mnist with this repository, so we have a single model definition for training across CPU/GPU/eager execution/TPU. The change to dataset.py is so that the raw data can be read from cloud storage systems (like GCS and S3).
-
- 03 Jan, 2018 1 commit
-
-
Asim Shankar authored
-
- 02 Jan, 2018 2 commits
-
-
Asim Shankar authored
-
Asim Shankar authored
- Prior to this change, the use of tf.data.Dataset essentially embedded the entire training/evaluation dataset into the graph as a constant, leading to unnecessarily humungous graphs (Fixes #3017) - Also, use batching on the evaluation dataset to allow evaluation on GPUs that cannot fit the entire evaluation dataset in memory (Fixes #3046)
-