- 03 Aug, 2018 1 commit
-
-
derekjchow authored
Bug fix: change dict.iteritems() to dict.items()
-
- 02 Aug, 2018 15 commits
-
-
Reed authored
-
Reed authored
The data_async_generation.py process would print to stderr, but the main process would redirect it's stderr to a pipe. The main process never read from the pipe, so when the pipe was full, data_async_generation.py would stall on a write to stderr. This change makes data_async_generation.py not write to stdout/stderr.
-
Mark Daoust authored
image segmentation post
-
Raymond Yuan authored
-
Raymond Yuan authored
-
Mark Daoust authored
-
Raymond Yuan authored
-
Mark Daoust authored
-
Mark Daoust authored
-
Mark Daoust authored
-
Mark Daoust authored
-
Mark Daoust authored
-
Mark Daoust authored
-
Abdullah Alrasheed authored
`iteritems()` was removed from python3. `items()` does the same functionality so changing it will work in both python2 and python3. The only difference as far as I know is `iteritems()` returns a generator where `items` returns a list. But for this this code it will not make any difference where we are just changing the key of the dict to a string.
-
Mark Daoust authored
-
- 01 Aug, 2018 6 commits
-
-
Raymond Yuan authored
-
Raymond Yuan authored
-
pkulzc authored
* Merged commit includes the following changes: 206852642 by Zhichao Lu: Build the balanced_positive_negative_sampler in the model builder for FasterRCNN. Also adds an option to use the static implementation of the sampler. -- 206803260 by Zhichao Lu: Fixes a misplaced argument in resnet fpn feature extractor. -- 206682736 by Zhichao Lu: This CL modifies the SSD meta architecture to support both Slim-based and Keras-based box predictors, and begins preparation for Keras box predictor support in the other meta architectures. Concretely, this CL adds a new `KerasBoxPredictor` base class and makes the meta architectures appropriately call whichever box predictors they are using. We can switch the non-ssd meta architectures to fully support Keras box predictors once the Keras Convolutional Box Predictor CL is submitted. -- 206669634 by Zhichao Lu: Adds an alternate m... -
Raymond Yuan authored
* nst colab * downloaded py filed * Removed text. Use gdoc for reviewing text, py for code * update ipynb * Removed google3 imports and added images * nst update images * final updates * add github and colab links * removed py file again
-
Reed authored
The output of an embeddding layer is already flattened, so the Flatten layers acted as no-ops.
-
Wolff Dobson authored
Update overfit_and_underfit.ipynb
-
- 31 Jul, 2018 14 commits
-
-
Alexandre Passos authored
Delete 4_Neural_Style_Transfer_with_Eager_Execution.py
-
Raymond Yuan authored
-
Raymond Yuan authored
* nst colab * downloaded py filed * Removed text. Use gdoc for reviewing text, py for code * update ipynb * Removed google3 imports and added images * nst update images * final updates
-
Taylor Robie authored
-
Reed authored
* Fix crash when Python interpreter not on PATH. * Fix lint error.
-
Reed authored
-
Raymond Yuan authored
-
Reed authored
-
Mark Daoust authored
Update noteboook title. Add link to MLCC text classification guide.
-
Taylor Robie authored
* add indirection file * remove unused imports * fix import
-
Reed authored
-
Reed authored
-
Reed authored
-
Alexandre Passos authored
Added readme
-
- 30 Jul, 2018 4 commits
-
-
Taylor Robie authored
* intermediate commit * ncf now working * reorder pipeline * allow batched decode for file backed dataset * fix bug * more tweaks * parallize false negative generation * shared pool hack * workers ignore sigint * intermediate commit * simplify buffer backed dataset creation to fixed length record approach only. (more cleanup needed) * more tweaks * simplify pipeline * fix misplaced cleanup() calls. (validation works\!) * more tweaks * sixify memoryview usage * more sixification * fix bug * add future imports * break up training input pipeline * more pipeline tuning * first pass at moving negative generation to async * refactor async pipeline to use files instead of ipc * refactor async pipeline * move expansion and concatenation from reduce worker to generation workers * abandon complete async due to interactions with the tensorflow threadpool * cleanup * remove performance_comparison.py * experiment with rough generator + interleave pipeline * yet more pipeline tuning * update on-the-fly pipeline * refactor preprocessing, and move train generation behind a GRPC server * fix leftover call * intermediate commit * intermediate commit * fix index error in data pipeline, and add logging to train data server * make sharding more robust to imbalance * correctly sample with replacement * file buffers are no longer needed for this branch * tweak sampling methods * add README for data pipeline * fix eval sampling, and vectorize eval metrics * add spillover and static training batch sizes * clean up cruft from earlier iterations * rough delint * delint 2 / n * add type annotations * update run script * make run.sh a bit nicer * change embedding initializer to match reference * rough pass at pure estimator model_fn * impose static shape hack (revisit later) * refinements * fix dir error in run.sh * add documentation * add more docs and fix an assert * old data test is no longer valid. Keeping it around as reference for the new one * rough draft of data pipeline validation script * don't rely on shuffle default * tweaks and documentation * add separate eval batch size for performance * initial commit * terrible hacking * mini hacks * missed a bug * messing about trying to get TPU running * TFRecords based TPU attempt * bug fixes * don't log remotely * more bug fixes * TPU tweaks and bug fixes * more tweaks * more adjustments * rework model definition * tweak data pipeline * refactor async TFRecords generation * temp commit to run.sh * update log behavior * fix logging bug * add check for subprocess start to avoid cryptic hangs * unify deserialize and make it TPU compliant * delint * remove gRPC pipeline code * fix logging bug * delint and remove old test files * add unit tests for NCF pipeline * delint * clean up run.sh, and add run_tpu.sh * forgot the most important line * fix run.sh bugs * yet more bash debugging * small tweak to add keras summaries to model_fn * Clean up sixification issues * address PR comments * delinting is never over
-
Raymond Yuan authored
-
Mark Daoust authored
-
Sundara Tejaswi Digumarti authored
Removed the conditional over distributed strategies when computing metrics. Metrics are now computed even when distributed strategies are used.
-