- 02 Jun, 2020 5 commits
-
-
A. Unique TensorFlower authored
PiperOrigin-RevId: 314417394
-
Maxim Neumann authored
PiperOrigin-RevId: 314412294
-
Hongkun Yu authored
PiperOrigin-RevId: 314394246
-
Chen Chen authored
PiperOrigin-RevId: 314373769
-
A. Unique TensorFlower authored
PiperOrigin-RevId: 314270569
-
- 01 Jun, 2020 1 commit
-
-
Hongkun Yu authored
PiperOrigin-RevId: 314177321
-
- 30 May, 2020 1 commit
-
-
Hongkun Yu authored
PiperOrigin-RevId: 313906815
-
- 29 May, 2020 2 commits
-
-
Hongkun Yu authored
Proposes the full functionality of MultiHeadAttention layer. This change first goes to model garden NLP library. PiperOrigin-RevId: 313847485
-
Chen Chen authored
PiperOrigin-RevId: 313812017
-
- 28 May, 2020 3 commits
-
-
Hongkun Yu authored
Deprecate old customized training loop for run_classifier.py as compile/fit fully satisfy needs/performance. PiperOrigin-RevId: 313660745
-
Reed Wanderman-Milne authored
Float32 is used if the model uses mixed precision with bfloat16. Float16 activation are unchanged. The motivation is that BERT with the LAMB optimizer with a gelu activation has an unstable loss when gelu is in bfloat16. Unfortunately, it is not easy to check if the LAMB optimizer and gelu is used, and perhaps there are other cases that work better with float32 activations instead of bfloat16 activations, so we always do the activation in float32 instead of bfloat16. PiperOrigin-RevId: 313618322
-
A. Unique TensorFlower authored
PiperOrigin-RevId: 313536026
-
- 26 May, 2020 2 commits
-
-
Maxim Neumann authored
PiperOrigin-RevId: 313148142
-
André Susano Pinto authored
This allows one to finetune a BERT model into a task before using it for another task. E.g. SQuAD before finetune another QA type of tasks. PiperOrigin-RevId: 313145768
-
- 24 May, 2020 2 commits
-
-
Hongkun Yu authored
PiperOrigin-RevId: 312988874
-
Hongkun Yu authored
PiperOrigin-RevId: 312939899
-
- 23 May, 2020 2 commits
-
-
Hongkun Yu authored
PiperOrigin-RevId: 312923051
-
Hongkun Yu authored
PiperOrigin-RevId: 312889153
-
- 21 May, 2020 3 commits
-
-
A. Unique TensorFlower authored
PiperOrigin-RevId: 312754139
-
A. Unique TensorFlower authored
PiperOrigin-RevId: 312751112
-
Hongkun Yu authored
Transformer Encoder: when embedding width differs from hidden size, add a projection to hidden size. PiperOrigin-RevId: 312708922
-
- 20 May, 2020 2 commits
-
-
Hongkun Yu authored
PiperOrigin-RevId: 312515585
-
Xiaoquan Kong authored
-
- 19 May, 2020 7 commits
-
-
Chen Chen authored
PiperOrigin-RevId: 312366167
-
Hongkun Yu authored
Remove is_v2_0 PiperOrigin-RevId: 312336907
-
Chen Chen authored
PiperOrigin-RevId: 312313738
-
André Susano Pinto authored
A default of 1 all the times is bad for TPU users which end up not using the device effectively. A larger default all the times is bad for GPU users. So compromise and make this dependent on the devices available. PiperOrigin-RevId: 312230371
-
A. Unique TensorFlower authored
PiperOrigin-RevId: 312209760
-
Ruoxin Sang authored
PiperOrigin-RevId: 312194218
-
Saurabh Saxena authored
PiperOrigin-RevId: 312193729
-
- 18 May, 2020 3 commits
-
-
Chen Chen authored
PiperOrigin-RevId: 312116965
-
Hongkun Yu authored
PiperOrigin-RevId: 312099777
-
A. Unique TensorFlower authored
PiperOrigin-RevId: 312096778
-
- 17 May, 2020 1 commit
-
-
A. Unique TensorFlower authored
Update nlp.modeling.layers.ReZeroTransformer, to have the same interface with nlp.modeling.layers.Transformer PiperOrigin-RevId: 311937563
-
- 15 May, 2020 1 commit
-
-
Hongkun Yu authored
PiperOrigin-RevId: 311773503
-
- 14 May, 2020 5 commits
-
-
nuka137 authored
-
Hongkun Yu authored
PiperOrigin-RevId: 311602262
-
Chen Chen authored
PiperOrigin-RevId: 311597242
-
Chen Chen authored
PiperOrigin-RevId: 311577663
-
Jeremiah Harmsen authored
Add network and BERT model to perform per-token classification (e.g., for named entity recognition tasks). PiperOrigin-RevId: 311480326
-