- 15 Jun, 2018 28 commits
-
-
Alexei Baevski authored
remove completed sentences from batch and allow batching uneven lengths (with fixes to make padded sequences work correctly in all models)
-
Myle Ott authored
-
Myle Ott authored
-
Myle Ott authored
This reverts commit b2e119c209363e6ff6d2878a69c7d1a507a2e9be.
-
Myle Ott authored
-
Myle Ott authored
-
Myle Ott authored
-
Myle Ott authored
-
Myle Ott authored
Improve dataloader speed and deprecate concept of batch_offset (use --sample-without-replacement instead)
-
Sergey Edunov authored
-
Myle Ott authored
-
Sergey Edunov authored
-
Myle Ott authored
-
Myle Ott authored
-
alexeib authored
-
Alexei Baevski authored
- process only last time step during generation - cache keys and values - dont apply masking during generation
-
Myle Ott authored
-
Myle Ott authored
-
Myle Ott authored
-
Myle Ott authored
-
Myle Ott authored
-
Myle Ott authored
-
alexeib authored
-
alexeib authored
-
Myle Ott authored
-
Myle Ott authored
-
Myle Ott authored
-
Myle Ott authored
-
- 24 May, 2018 1 commit
-
-
Myle Ott authored
-
- 22 May, 2018 1 commit
-
-
- 21 May, 2018 1 commit
-
-
Myle Ott authored
-
- 09 May, 2018 3 commits
- 01 May, 2018 3 commits
- 02 Apr, 2018 1 commit
-
-
Myle Ott authored
Changes: - 7d19e36: Add `--sampling` flag to generate.py to sample instead of doing beam search - c777340: Add `scripts/average_checkpoints.py` to average multiple checkpoints into a combined model - 3ea882c: Add `--max-update` option to train.py to stop training after a given number of updates - small bugfixes for distributed training, LSTM, inverse square root LR scheduler
-
- 28 Mar, 2018 2 commits
-
-
Sergey Edunov authored
Update training commands
-
Runqi Yang authored
specify a single GPU setup for IWSLT14
-