"tests/git@developer.sourcefind.cn:OpenDAS/mmcv.git" did not exist on "a03774d6dbb24113be51e2f0aa2256586c5befe4"
  1. 07 May, 2019 1 commit
    • Kartikay Khandelwal's avatar
      Mask out embeddings associated with padding (#710) · 8d9063fe
      Kartikay Khandelwal authored
      Summary:
      Pull Request resolved: https://github.com/pytorch/fairseq/pull/710
      
      Previously there was a bug in how we dealt with padding when computing the input representation from the segment and position embedding. D15144912 fixed this by adding an offset based on the padding id. However this makes assumptions about the padding id which may not hold true for vocabularies built outside of pyText and fairseq. Based on a discussion with barlaso, this diff 0's out all the embeddings associated with the padding.
      
      Reviewed By: borguz
      
      Differential Revision: D15209395
      
      fbshipit-source-id: 5573020e610f5466e673fe3845c3ed34ebb5c44d
      8d9063fe
  2. 06 May, 2019 5 commits
  3. 05 May, 2019 3 commits
  4. 04 May, 2019 4 commits
  5. 03 May, 2019 2 commits
  6. 02 May, 2019 5 commits
    • Peng-Jen Chen's avatar
      Make learned positional embedding optional · 39264559
      Peng-Jen Chen authored
      Summary:
      - Add learned positional embedding binary flag to masked LM model.
      - Add base arch config for masked LM model which sets all the binary parameters to False. Otherwise some of the binary flag parameters will always be override by config in `xlm_architecture` (e.g. encoder_learned_pos)
      
      Reviewed By: liezl200
      
      Differential Revision: D15054487
      
      fbshipit-source-id: d78827f352b9160a89c9dc4f45b9fce15a2f234d
      39264559
    • Myle Ott's avatar
      Move distributed_init into DistributedFairseqModel (#687) · 34726d56
      Myle Ott authored
      Summary:
      This should make rendezvous happen as lazily as possible.
      Pull Request resolved: https://github.com/pytorch/fairseq/pull/687
      
      Differential Revision: D15151145
      
      Pulled By: myleott
      
      fbshipit-source-id: d70816a85414c5d509a6b12e2b339b4736db2c88
      34726d56
    • Myle Ott's avatar
      Validate on all sets based on --save-interval-updates · fb18be00
      Myle Ott authored
      Summary: Pull Request resolved: https://github.com/pytorch/fairseq/pull/693
      
      Differential Revision: D15174831
      
      fbshipit-source-id: 98688b1269ead5694e5116659ff64507d3c0d1c0
      fb18be00
    • Myle Ott's avatar
      Fix inconsistent gradient check · 4a30a5f6
      Myle Ott authored
      Summary: Pull Request resolved: https://github.com/pytorch/fairseq/pull/692
      
      Differential Revision: D15174954
      
      fbshipit-source-id: 1a7bff9aeed3e2cc658577be9d79e8c9f72314c2
      4a30a5f6
    • Kritika Singh's avatar
      Make CTC work with more encoder-only models · ffc9c8cc
      Kritika Singh authored
      Summary:
      Changes include:
      1. Added get_normalized_probabilities to the encoder-only base class FairseqEncoderModel
      2. Made CTCCriterion work for both batch_first (LSTMSubsampleEncoderModel) and batch_second (LSTMEncoderOnly) encoder types
      3. Added tests for different encoder and CTC combinations.
      
      TODO:
      CTC still doesn't work for VGGLSTMEncoderModel so I have disabled that. Will debug and send out fix in another diff.
      
      Reviewed By: jay-mahadeokar
      
      Differential Revision: D15158818
      
      fbshipit-source-id: acb484bad705c937d676d2c3dcde3e3562d68ed9
      ffc9c8cc
  7. 01 May, 2019 5 commits
  8. 30 Apr, 2019 6 commits
  9. 29 Apr, 2019 2 commits
  10. 27 Apr, 2019 2 commits
  11. 26 Apr, 2019 1 commit
  12. 25 Apr, 2019 4 commits