1. 21 Dec, 2021 1 commit
  2. 23 Nov, 2021 1 commit
  3. 10 Mar, 2021 2 commits
  4. 17 Feb, 2021 2 commits
  5. 04 Nov, 2020 2 commits
  6. 26 Oct, 2020 2 commits
  7. 24 Aug, 2020 2 commits
  8. 12 Aug, 2020 2 commits
  9. 11 Aug, 2020 1 commit
  10. 08 Aug, 2020 2 commits
  11. 17 Jul, 2020 2 commits
  12. 08 Jul, 2020 2 commits
  13. 03 Jun, 2020 4 commits
    • Hongkun Yu's avatar
      a6c0e677
    • Hongkun Yu's avatar
      4bb13e61
    • xinliupitt's avatar
      Add relative positional embedding to KerasBERT (#8617) · c3c2386c
      xinliupitt authored
      * root dir
      
      * zone updated
      
      * print mask
      
      * preview emb
      
      * tf print
      
      * input only
      
      * emb
      
      * tf print
      
      * emb after mask
      
      * masked_softmax print
      
      * print scores
      
      * multi folder
      
      * first pos emb
      
      * check input shape
      
      * add test temp
      
      * import math
      
      * two classes
      
      * prints
      
      * all get_pos replace
      
      * make time scale private
      
      * pos emb comments
      
      * print input
      
      * embedding_inputs
      
      * tf shape
      
      * dimention list
      
      * tf_util
      
      * print tf_util
      
      * concise
      
      * transformer pos change to layer
      
      * keep length var
      
      * length as input
      
      * None as input
      
      * print time signal
      
      * print time signal
      
      * remove print
      
      * test input shape
      
      * double check shape
      
      * double check shape
      
      * double check shape
      
      * more test
      
      * shape check
      
      * shape check
      
      * print 97 info
      
      * print 97 info new
      
      * test if sam
      
      * assert same
      
      * remove assert
      
      * tf print same
      
      * tf print diff
      
      * output example
      
      * output example
      
      * output example
      
      * formal test
      
      * formal test length
      
      * raise valurerror
      
      * test valurerror
      
      * double check
      
      * comments
      
      * remove prints
      
      * rename relative
      
      * delet naive test
      
      * delete docs in xinliu branch
      
      * code reformat
      
      * import order
      
      * indentation fix
      
      * more files
      
      * adjust char number
      
      * disable not callable
      
      * comment to length
      
      * error of length unequal to input_shape
      
      * root dir
      
      * zone updated
      
      * print mask
      
      * preview emb
      
      * tf print
      
      * input only
      
      * emb
      
      * tf print
      
      * emb after mask
      
      * masked_softmax print
      
      * print scores
      
      * multi folder
      
      * remove docs
      
      * remove prints
      
      * root dir
      
      * zone updated
      
      * print mask
      
      * preview emb
      
      * tf print
      
      * input only
      
      * emb
      
      * tf print
      
      * emb after mask
      
      * masked_softmax print
      
      * print scores
      
      * multi folder
      
      * remove docs
      
      * apply revised 3 files
      
      * rm prints
      c3c2386c
    • Tianqi Liu's avatar
      Internal change · 20897493
      Tianqi Liu authored
      PiperOrigin-RevId: 314451720
      20897493
  14. 02 Jun, 2020 1 commit
    • xinliupitt's avatar
      Add relative positional embedding to KerasBERT (#8606) · 2db2501b
      xinliupitt authored
      * root dir
      
      * zone updated
      
      * print mask
      
      * preview emb
      
      * tf print
      
      * input only
      
      * emb
      
      * tf print
      
      * emb after mask
      
      * masked_softmax print
      
      * print scores
      
      * multi folder
      
      * first pos emb
      
      * check input shape
      
      * add test temp
      
      * import math
      
      * two classes
      
      * prints
      
      * all get_pos replace
      
      * make time scale private
      
      * pos emb comments
      
      * print input
      
      * embedding_inputs
      
      * tf shape
      
      * dimention list
      
      * tf_util
      
      * print tf_util
      
      * concise
      
      * transformer pos change to layer
      
      * keep length var
      
      * length as input
      
      * None as input
      
      * print time signal
      
      * print time signal
      
      * remove print
      
      * test input shape
      
      * double check shape
      
      * double check shape
      
      * double check shape
      
      * more test
      
      * shape check
      
      * shape check
      
      * print 97 info
      
      * print 97 info new
      
      * test if sam
      
      * assert same
      
      * remove assert
      
      * tf print same
      
      * tf print diff
      
      * output example
      
      * output example
      
      * output example
      
      * formal test
      
      * formal test length
      
      * raise valurerror
      
      * test valurerror
      
      * double check
      
      * comments
      
      * remove prints
      
      * rename relative
      
      * delet naive test
      
      * delete docs in xinliu branch
      
      * code reformat
      
      * import order
      
      * indentation fix
      
      * more files
      
      * adjust char number
      
      * disable not callable
      
      * comment to length
      
      * error of length unequal to input_shape
      2db2501b
  15. 13 Feb, 2020 1 commit
  16. 19 Dec, 2019 1 commit
  17. 02 Dec, 2019 1 commit
  18. 22 Nov, 2019 1 commit
  19. 09 Oct, 2019 1 commit
    • Reed Wanderman-Milne's avatar
      Simply LayerNorm mixed precision logic. · 0257b276
      Reed Wanderman-Milne authored
      Instead of needing to ensure variables are float32, casting inputs to float32, etc, instead dtype="float32" is passed to the layer constructor, which will do all that logic automatically.
      
      The only difference is the output of LayerNorm is now float32 instead of float16, so an extra cast is needed elsewhere.
      
      PiperOrigin-RevId: 273833286
      0257b276
  20. 07 Oct, 2019 1 commit
  21. 05 Sep, 2019 1 commit
  22. 22 Aug, 2019 1 commit
  23. 21 Aug, 2019 1 commit
  24. 20 Aug, 2019 1 commit
  25. 08 Aug, 2019 1 commit
    • Reed's avatar
      Fix fp16 Transformer model. (#7402) · 58340818
      Reed authored
      Also, do Transformer inference in fp16, as well as training, when --dtype=fp16. In TF 2, layers now cannot run in multiple different dtypes, so we must use the same dtype for training and inference.
      58340818
  26. 24 Jul, 2019 1 commit
  27. 21 Jun, 2019 1 commit
  28. 19 Jun, 2019 1 commit