1. 03 Oct, 2019 2 commits
    • Simon Layton's avatar
      Fix test fails and warnings · 89988364
      Simon Layton authored
      Attention output was in bnij ordering instead of ijbn which everything
      else will expect. This was an oversight on my part, and keeps the
      attention inputs/outputs identical to the original code.
      
      Also moved back from tensor slicing to index_select in rel_shift_bnij to
      make the tracer happy.
      89988364
    • Simon Layton's avatar
      Fix missed head transpose · 9ffda216
      Simon Layton authored
      9ffda216
  2. 02 Oct, 2019 1 commit
    • Simon Layton's avatar
      Re-order attention head outputs for better perf · d51b5894
      Simon Layton authored
      Significant performance boost over the original orderings
      on an already somewhat optimised branch this gave me > 2x end-to-end
      throughput on a squad xlnet fine-tuning task (batch 8, seq-length 612,
      fp16)
      d51b5894
  3. 01 Oct, 2019 6 commits
  4. 30 Sep, 2019 1 commit
    • Santosh Gupta's avatar
      Update README.md · 5c3b32d4
      Santosh Gupta authored
      Lines 183 - 200, fixed indentation. Line 198, replaced `tokenizer_class` with `BertTokenizer`, since `tokenizer_class` is not defined in the loop it belongs to.
      5c3b32d4
  5. 29 Sep, 2019 1 commit
  6. 28 Sep, 2019 2 commits
  7. 27 Sep, 2019 18 commits
  8. 26 Sep, 2019 9 commits