- 23 Jul, 2019 2 commits
-
-
Hongkun Yu authored
Only care about errors and output into an error file.
-
Hongkun Yu authored
259442882 by hongkuny<hongkuny@google.com>: Internal -- 259341546 by mrry<mrry@google.com>: Remove DEBUG-level logging from the BERT benchmark. This triggers graph serialization and other verbose logging in the TensorFlow runtime, which inflates the execution time. -- 259253185 by hongkuny<hongkuny@google.com>: Writes a separated checkpoint for the core model in pretraining. Clean up export utils to just take a model as argument. -- 258893811 by hongkuny<hongkuny@google.com>: Adds summaries for metrics, allowing metrics inside keras.model. -- 258881002 by hongkuny<hongkuny@google.com>: Fix lint. -- 258597234 by rxsang<rxsang@google.com>: Update all the TPUStrategy examples to use the new v2 APIs, i.e. make_dataset_iterator -> experimental_distribute_dataset, make_input_fn_iterator -> experimental_distribute_datasets_from_function, unwrap -> experimental_local_results, experimental_run -> experimental_run_v2 -- 258581998 by taylorrobie<taylorrobie@google.com>: Update keras v2 optimizers to reuse coefficients which are shared across all updates, which reduces the total number of ops created by between 5% (for simple optimizers such as SGD and Adagrad) and 25% (for complicated optimizers such as Adam and NAdam). Separate copies are made for each device and dtype. The effect of this change on run time is fairly minimal since Grappler is expected to consolidate most of these ops; however it does improve graph construction time. -- 258208153 by hongkuny<hongkuny@google.com>: Adds run_eagerly option for bert. -- 257883986 by hongkuny<hongkuny@google.com>: Adds tf.summary for bert training -- 256204636 by hongkuny<hongkuny@google.com>: Internal -- 256079834 by hongkuny<hongkuny@google.com>: Clean up: move common flags together for further refactoring Enable steps_per_loop option for all applications. -- 255493073 by hongkuny<hongkuny@google.com>: BERT initial OSS readme update. -- 255470372 by dmchen<dmchen@google.com>: Slightly expand expected range for F1 score in BERT SQuAD accuracy test -- 255109240 by hongkuny<hongkuny@google.com>: Update eval/predict batch sizes. -- 255010016 by hongkuny<hongkuny@google.com>: Internal -- 254874613 by hongkuny<hongkuny@google.com>: Update glue tasks enum to match directory name -- 254866171 by taylorrobie<taylorrobie@google.com>: Internal change 254785517 by zongweiz<zongweiz@google.com>: Use train_single_step for BERT GPU models to temporarily work around some performance bugs in GPU runs -- 254497647 by hongkuny<hongkuny@google.com>: Fix device placement for TPU export model. -- 254134531 by yuefengz<yuefengz@google.com>: Fix a typo in bert_benchmark.py -- 254069984 by hongkuny<hongkuny@google.com>: Automated rollback of changelist 254060732. 254061429 by hongkuny<hongkuny@google.com>: Use host while loop for training steps. -- 254060732 by yifeif<yifeif@google.com>: Automated rollback of changelist 254027750. 254027750 by hongkuny<hongkuny@google.com>: Internal change 253850824 by hongkuny<hongkuny@google.com>: Improve bert training utils. -- 253818191 by hongkuny<hongkuny@google.com>: Update savedmodel export to use new model.save() api. -- 253636854 by dmchen<dmchen@google.com>: Run only training in BERT SQuAD performance test -- 253118910 by hongkuny<hongkuny@google.com>: Internal change 253113801 by zongweiz<zongweiz@google.com>: Internal change 252697519 by dmchen<dmchen@google.com>: BERT SQuAD accuracy test -- 252663512 by A. Unique TensorFlower<gardener@tensorflow.org>: Internal change -- 252647871 by A. Unique TensorFlower<gardener@tensorflow.org>: Enable multi worker TPU training for BERT pretraining. -- 252522861 by hongkuny<hongkuny@google.com>: Remove export using trained model due to implementation error -- 252156812 by yuefengz<yuefengz@google.com>: Fix the callback method name in BERT: replaced on_batch_start with on_batch_begin. Without the fix, it won't work with Keras callbacks. -- 251782065 by dmchen<dmchen@google.com>: Internal change 251681245 by hongkuny<hongkuny@google.com>: Update bert to use the new tf.distribute APIs -- 251575972 by A. Unique TensorFlower<gardener@tensorflow.org>: Remove `steps_per_run` when instantiating TPUStrategy. -- 251325964 by hongkuny<hongkuny@google.com>: Improve flags -- 250942274 by tobyboyd<tobyboyd@google.com>: Internal change 250779087 by A. Unique TensorFlower<gardener@tensorflow.org>: Reduce BERT Perfzero benchmark test training steps. -- 250713045 by hongkuny<hongkuny@google.com>: TPU util -- 250606180 by A. Unique TensorFlower<gardener@tensorflow.org>: Fix BERT benchamrk test errors. -- 250589623 by A. Unique TensorFlower<gardener@tensorflow.org>: Change BERT benchmark test pretrained checkpoint url. -- 250587892 by A. Unique TensorFlower<gardener@tensorflow.org>: Fix error in BERT custom training loop checkpoint restoration. -- 250577163 by A. Unique TensorFlower<gardener@tensorflow.org>: Add logic to inject callback that measures performance in BERT custom training loop. -- 250529526 by hongkuny<hongkuny@google.com>: Internal clean up -- 250428976 by hongkuny<hongkuny@google.com>: Internal change 250415383 by A. Unique TensorFlower<gardener@tensorflow.org>: Add min/max value to BERT classifier benchmark test. -- 250376246 by A. Unique TensorFlower<gardener@tensorflow.org>: Add benchmark performance test to run BERT on multiple numbers of GPUs. -- 250347237 by A. Unique TensorFlower<gardener@tensorflow.org>: Fix linting errors in BERT benchmark test. -- 250326131 by A. Unique TensorFlower<gardener@tensorflow.org>: Internal change 250315593 by A. Unique TensorFlower<gardener@tensorflow.org>: Internal change 250303528 by haoyuzhang<haoyuzhang@google.com>: Add method docstring to fix lint error. -- 250009207 by A. Unique TensorFlower<gardener@tensorflow.org>: Add feature in BERT to write training metrics to a summary file. -- 249896208 by hongkuny<hongkuny@google.com>: Adds __init__.py -- 249883771 by hongkuny<hongkuny@google.com>: Creates a benchmark dir -- 249580533 by A. Unique TensorFlower<gardener@tensorflow.org>: Internal change 249566870 by A. Unique TensorFlower<gardener@tensorflow.org>: Set up BERT benchmark test. -- 249500988 by hongkuny<hongkuny@google.com>: Lints -- 249377254 by hongkuny<hongkuny@google.com>: Internal change 249373328 by hongkuny<hongkuny@google.com>: Clean up tf import -- 249333938 by hongkuny<hongkuny@google.com>: Fix tf1 import -- 249325089 by hongkuny<hongkuny@google.com>: BERT 2.0 -- 249173564 by hongkuny<hongkuny@google.com>: Internal change PiperOrigin-RevId: 259442882
-
- 22 Jul, 2019 1 commit
-
-
Hongkun Yu authored
* Update pylint.rcfile * Update pylint.rcfile * Update pylint.rcfile * add new sanity check script for lint to replace current lint script. * Revert "Update pylint.rcfile" This reverts commit f6036cd7e7c4b9e3eeb47bb56a63927a040a2761. * Revert "Update pylint.rcfile" This reverts commit e3af497342e26bbbbecfc8c8f79cb0e24a2ef960. * Revert "Update pylint.rcfile" This reverts commit 6136636eee6e90fd191ebbb4ccaa9fb89c0290f4. * update scripts * disable trailing-newlines
-
- 21 Jul, 2019 1 commit
-
-
Zongwei Zhou authored
-
- 20 Jul, 2019 3 commits
-
-
Zongwei Zhou authored
-
Toby Boyd authored
-
Toby Boyd authored
-
- 19 Jul, 2019 9 commits
-
-
Igor authored
259030078 by isaprykin<isaprykin@google.com>: Clean up the --clone_model_in_keras_dist_strat from Keras Resnet. The cloning flag has been removed. The current rule is that cloning is only done in graph mode. That resulted in duplicate benchmarks: eager+no-cloning vs eager+cloning. I removed eager+cloning ones. -- 259026454 by isaprykin<isaprykin@google.com>: Internal change PiperOrigin-RevId: 259030078 -
Jing Li authored
* Merged commit includes the following changes: 258867180 by jingli<jingli@google.com>: Add new folders for upcoming reorg in model garden. -- 258893811 by hongkuny<hongkuny@google.com>: Adds summaries for metrics, allowing metrics inside keras.model. -- 258893048 by isaprykin<isaprykin@google.com>: Remove the `cloning` argument to `compile()`. Keras models are distributed by cloning in graph mode and without cloning in eager mode as of the change # 258652546. -- 258881002 by hongkuny<hongkuny@google.com>: Fix lint. -- 258874998 by hongkuny<hongkuny@google.com>: Internal -- 258872662 by hongkuny<hongkuny@google.com>: Fix doc -- PiperOrigin-RevId: 258867180 * Create __init__.py * Update __init__.py * Update __init__.py * Update __init__.py -
guptapriya authored
-
guptapriya authored
-
guptapriya authored
This combination does not yet work. Fail early with an explicit message instead of throwing error later on.
-
Chris Mattmann authored
Fix for TF-models #7216: CIFAR-10 tutorial for multi-GPU fails because full shape isn't passed to prefetch_queue contributed by mattmann. (#7217)
-
guptapriya authored
The current approach checks for presence of contrib. Sometimes this is not sufficient (for e..g when testing TF 1 + enable_v2_behavior=True which is what internal tests currently do)
-
Hongkun Yu authored
258881002 by hongkuny<hongkuny@google.com>: Fix lint. -- 258874998 by hongkuny<hongkuny@google.com>: Internal -- 258872662 by hongkuny<hongkuny@google.com>: Fix doc -- 258871624 by hongkuny<hongkuny@google.com>: Internal change PiperOrigin-RevId: 258881002
-
- 18 Jul, 2019 7 commits
-
-
Hongkun Yu authored
258597234 by rxsang<rxsang@google.com>: Update all the TPUStrategy examples to use the new v2 APIs, i.e. make_dataset_iterator -> experimental_distribute_dataset, make_input_fn_iterator -> experimental_distribute_datasets_from_function, unwrap -> experimental_local_results, experimental_run -> experimental_run_v2 -- 258581998 by taylorrobie<taylorrobie@google.com>: Update keras v2 optimizers to reuse coefficients which are shared across all updates, which reduces the total number of ops created by between 5% (for simple optimizers such as SGD and Adagrad) and 25% (for complicated optimizers such as Adam and NAdam). Separate copies are made for each device and dtype. The effect of this change on run time is fairly minimal since Grappler is expected to consolidate most of these ops; however it does improve graph construction time. -- PiperOrigin-RevId: 258597234 -
Jing Li authored
-
Toby Boyd authored
* Added benchmarks and common flags. * Add cpu tests. * Add tracking epoch times. * fix transformer. * Add examples_per_second. * fix pylint
-
Chris Mattmann authored
Fix for #7225: CIFAR-10 eval fails with error TypeError: Input 'predictions' of 'InTopKV2' Op has type float16 that contributed by mattmann. (#7227)
-
Yongzhe Wang authored
* Merged commit includes the following changes: 257930561 by yongzhe: Mobile LSTD TfLite Client. -- 257928126 by yongzhe: Mobile SSD Tflite client. -- 257921181 by menglong: Fix discrepancy between pre_bottleneck = {true, false} -- 257561213 by yongzhe: File utils. -- 257449226 by yongzhe: Mobile SSD Client. -- 257264654 by yongzhe: SSD utils. -- 257235648 by yongzhe: Proto bazel build rules. -- 256437262 by Menglong Zhu: Fix check for FusedBatchNorm op to only verify it as a prefix. -- 256283755 by yongzhe: Bazel build and copybara changes. -- 251947295 by yinxiao: Add missing interleaved option in checkpoint restore. -- 251513479 by yongzhe: Conversion utils. -- 248783193 by yongzhe: Branch protos needed for the lstd client. -- 248200507 by menglong: Fix proto namespace in example config -- PiperOrigin-RevId: 257930561 * Delete BUILD * Merged commit includes the following changes: 258709909 by yongzhe: 1. Fix a bug that input wasn't copied. 2. Change the tensor indexing to support graph with postprocessing. 3. Fix a bug that the quantized lstm states weren't initialized. -- 258398095 by yongzhe: Internal change. -- PiperOrigin-RevId: 258709909 * Adding myself as the code owner -
Haoyu Zhang authored
* Config threadpool, cuDNN persistent BN, and grappler layout optimizer properly for ResNet56 * Add tweaked tests for Resnet56 * Avoid triggering the last partial batch overhead by explicitly dropping remainder
-
Yongzhe Wang authored
* Merged commit includes the following changes: 257930561 by yongzhe: Mobile LSTD TfLite Client. -- 257928126 by yongzhe: Mobile SSD Tflite client. -- 257921181 by menglong: Fix discrepancy between pre_bottleneck = {true, false} -- 257561213 by yongzhe: File utils. -- 257449226 by yongzhe: Mobile SSD Client. -- 257264654 by yongzhe: SSD utils. -- 257235648 by yongzhe: Proto bazel build rules. -- 256437262 by Menglong Zhu: Fix check for FusedBatchNorm op to only verify it as a prefix. -- 256283755 by yongzhe: Bazel build and copybara changes. -- 251947295 by yinxiao: Add missing interleaved option in checkpoint restore. -- 251513479 by yongzhe: Conversion utils. -- 248783193 by yongzhe: Branch protos needed for the lstd client. -- 248200507 by menglong: Fix proto namespace in example config -- P...
-
- 16 Jul, 2019 3 commits
-
-
Hongkun Yu authored
258208153 by hongkuny<hongkuny@google.com>: Adds run_eagerly option for bert. -- PiperOrigin-RevId: 258208153 -
nnigania authored
* Ncf perf changes 1)exclude metric layer from CTL train step 2)dataset optimization to fix size of the sample_weights, preventing a costly broadcast during loss calculation for multi-gpu case
-
yongzhe2160 authored
* Merged commit includes the following changes: 257930561 by yongzhe: Mobile LSTD TfLite Client. -- 257928126 by yongzhe: Mobile SSD Tflite client. -- 257921181 by menglong: Fix discrepancy between pre_bottleneck = {true, false} -- 257561213 by yongzhe: File utils. -- 257449226 by yongzhe: Mobile SSD Client. -- 257264654 by yongzhe: SSD utils. -- 257235648 by yongzhe: Proto bazel build rules. -- 256437262 by Menglong Zhu: Fix check for FusedBatchNorm op to only verify it as a prefix. -- 256283755 by yongzhe: Bazel build and copybara changes. -- 251947295 by yinxiao: Add missing interleaved option in checkpoint restore. -- 251513479 by yongzhe: Conversion utils. -- 248783193 by yongzhe: Branch protos needed for the lstd client. -- 248200507 by menglong: Fix proto namespace in example config -- PiperOrigin-RevId: 257930561 * Delete BUILD
-
- 15 Jul, 2019 3 commits
-
-
Bruce Fontaine authored
* Initial implementation of Shakespeare character LSTM. * Fix import order
-
Hongkun Yu authored
257883986 by hongkuny<hongkuny@google.com>: Adds tf.summary for bert training -- PiperOrigin-RevId: 257883986 -
pkulzc authored
257914648 by lzc: Internal changes -- 257525973 by Zhichao Lu: Fixes bug that silently prevents checkpoints from loading when training w/ eager + functions. Also sets up scripts to run training. -- 257296614 by Zhichao Lu: Adding detection_features to model outputs -- 257234565 by Zhichao Lu: Fix wrong order of `classes_with_max_scores` in class-agnostic NMS caused by sorting in partitioned-NMS. -- 257232002 by ronnyvotel: Supporting `filter_nonoverlapping` option in np_box_list_ops.clip_to_window(). -- 257198282 by Zhichao Lu: Adding the focal loss and l1 loss from the Objects as Points paper. -- 257089535 by Zhichao Lu: Create Keras based ssd + resnetv1 + fpn. -- 257087407 by Zhichao Lu: Make object_detection/data_decoders Python3-compatible. -- 257004582 by Zhichao Lu: Updates _decode_raw_data_into_masks_and_boxes to the latest binary masks-to-string encoding format. -- 257002124 by Zhichao Lu: Make object_detection/utils Python3-compatible, except json_utils. The patching trick used in json_utils is not going to work in Python 3. -- 256795056 by lzc: Add a detection_anchor_indices field to detection outputs. -- 256477542 by Zhichao Lu: Make object_detection/core Python3-compatible. -- 256387593 by Zhichao Lu: Edit class_id_function_approximations builder to skip class ids not present in label map. -- 256259039 by Zhichao Lu: Move NMS to TPU for FasterRCNN. -- 256071360 by rathodv: When multiclass_scores is empty, add one-hot encoding of groundtruth_classes as multiclass scores so that data_augmentation ops that expect the presence of multiclass_scores don't have to individually handle this case. Also copy input tensor_dict to out_tensor_dict first to avoid inplace modification. -- 256023645 by Zhichao Lu: Adds the first WIP iterations of TensorFlow v2 eager + functions style custom training & evaluation loops. -- 255980623 by Zhichao Lu: Adds a new data augmentation operation "remap_labels" which remaps a set of labels to a new label. -- 255753259 by Zhichao Lu: Announcement of the released evaluation tutorial for Open Images Challenge 2019. -- 255698776 by lzc: Fix rewrite_nn_resize_op function which was broken by tf forward compatibility movement. -- 255623150 by Zhichao Lu: Add Keras-based ResnetV1 models. -- 255504992 by Zhichao Lu: Fixing the typo in specifying label expansion for ground truth segmentation file. -- 255470768 by Zhichao Lu: 1. Fixing Python bug with parsed arguments. 2. Adding capability to parse relevant columns from CSV header. 3. Fixing bug with duplicated labels expansion. -- 255462432 by Zhichao Lu: Adds a new data augmentation operation "drop_label_probabilistically" which drops a given label with the given probability. This supports experiments on training in the presence of label noise. -- 255441632 by rathodv: Fallback on groundtruth classes when multiclass_scores tensor is empty. -- 255434899 by Zhichao Lu: Ensuring evaluation binary can run even with big files by synchronizing processing of ground truth and predictions: in this way, ground truth is not stored but immediatly used for evaluation. In case gt of object masks, this allows to run evaluations on relatively large sets. -- 255337855 by lzc: Internal change. -- 255308908 by Zhichao Lu: Add comment to clarify usage of calibration parameters proto. -- 255266371 by Zhichao Lu: Ensuring correct processing of the case, when no groundtruth masks are provided for an image. -- 255236648 by Zhichao Lu: Refactor model_builder in faster_rcnn.py to a util_map, so that it's possible to be overwritten. -- 255093285 by Zhichao Lu: Updating capability to subsample data during evaluation -- 255081222 by rathodv: Convert groundtruth masks to be of type float32 before its used in the loss function. When using mixed precision training, masks are represented using bfloat16 tensors in the input pipeline for performance reasons. We need to convert them to float32 before using it in the loss function. -- 254788436 by Zhichao Lu: Add forward_compatible to non_max_suppression_with_scores to make it is compatible with older tensorflow version. -- 254442362 by Zhichao Lu: Add num_layer field to ssd feature extractor proto. -- 253911582 by jonathanhuang: Plumbs Soft-NMS options (using the new tf.image.non_max_suppression_with_scores op) into the TF Object Detection API. It adds a `soft_nms_sigma` field to the postprocessing proto file and plumbs this through to both the multiclass and class_agnostic versions of NMS. Note that there is no effect on behavior of NMS when soft_nms_sigma=0 (which it is set to by default). See also "Soft-NMS -- Improving Object Detection With One Line of Code" by Bodla et al (https://arxiv.org/abs/1704.04503) -- 253703949 by Zhichao Lu: Internal test fixes. -- 253151266 by Zhichao Lu: Fix the op type check for FusedBatchNorm, given that we introduced FusedBatchNormV3 in a previous change. -- 252718956 by Zhichao Lu: Customize activation function to enable relu6 instead of relu for saliency prediction model seastarization -- 252158593 by Zhichao Lu: Make object_detection/core Python3-compatible. -- 252150717 by Zhichao Lu: Make object_detection/core Python3-compatible. -- 251967048 by Zhichao Lu: Make GraphRewriter proto extensible. -- 251950039 by Zhichao Lu: Remove experimental_export_device_assignment from TPUEstimator.export_savedmodel(), so as to remove rewrite_for_inference(). As a replacement, export_savedmodel() V2 API supports device_assignment where user call tpu.rewrite in model_fn and pass in device_assigment there. -- 251890697 by rathodv: Updated docstring to include new output nodes. -- 251662894 by Zhichao Lu: Add autoaugment augmentation option to objection detection api codebase. This is an available option in preprocessor.py. The intended usage of autoaugment is to be done along with random flipping and cropping for best results. -- 251532908 by Zhichao Lu: Add TrainingDataType enum to track whether class-specific or agnostic data was used to fit the calibration function. This is useful, since classes with few observations may require a calibration function fit on all classes. -- 251511339 by Zhichao Lu: Add multiclass isotonic regression to the calibration builder. -- 251317769 by pengchong: Internal Change. -- 250729989 by Zhichao Lu: Fixing bug in gt statistics count in case of mask and box annotations. -- 250729627 by Zhichao Lu: Label expansion for segmentation. -- 250724905 by Zhichao Lu: Fix use_depthwise in fpn and test it with fpnlite on ssd + mobilenet v2. -- 250670379 by Zhichao Lu: Internal change 250630364 by lzc: Fix detection_model_zoo footnotes -- 250560654 by Zhichao Lu: Fix static shape issue in matmul_crop_and_resize. -- 250534857 by Zhichao Lu: Edit class agnostic calibration function docstring to more accurately describe the function's outputs. -- 250533277 by Zhichao Lu: Edit the multiclass messages to use class ids instead of labels. -- PiperOrigin-RevId: 257914648
-
- 11 Jul, 2019 5 commits
-
-
Toby Boyd authored
-
Toby Boyd authored
* Record highest uncased bleu found. * change to bleu_best_score_iteration
-
Toby Boyd authored
-
saberkun authored
257314238 by hongkuny<hongkuny@google.com>: Creates transformer v2 README. Remove contents that are not implemented. -- PiperOrigin-RevId: 257314238 -
Toby Boyd authored
* Move to global_step. * Hook to use global_step. * fix comment start step 1 not step 0. * remove hack used for testing. * Add docstring.
-
- 10 Jul, 2019 1 commit
-
-
Rahul Nimbal authored
* Update research/maskgan/README.md Co-Authored-By:Andrew M Dai <andy.dai@gmail.com>
-
- 09 Jul, 2019 2 commits
-
-
Haoyu Zhang authored
* Improve performance for Cifar ResNet benchmarks * Revert batch size changes to benchmarks
-
David Andersen authored
Update to tf 1.14 syntax, fix bug #7125 (needed additional expand for conv2d). Suppress compat warnings by moving to compat.v1 versions of some functions. Note that this code is not 2.0 compatible yet - that will be a future push. (#7177)
-
- 08 Jul, 2019 3 commits
-
-
Toby Boyd authored
-
Toby Boyd authored
* reduce iterations from 20 to 12. * add fp16 dynamic batch accuracy check. * fix existing lint issue.
-
Yang Liu authored
* Bug fix of cifar10_eval.py line 120: changed from eval_data = FLAGS.eval_data == 'test' to: eval_data = FLAGS.eval_data comment: the original code assigns 'true' to eval_data, when the script is being used to evaluate networks on the evaluation set, it DOES NOT load the evaluation set as intended, but actually loads the trainning set (line 105 of cifar10_input.py). * Remove one line that was an artifact of issue.
-