- 20 Jun, 2019 5 commits
-
-
Myle Ott authored
Summary: Pull Request resolved: https://github.com/pytorch/fairseq/pull/818 Differential Revision: D15916265 Pulled By: myleott fbshipit-source-id: c66c0bd988d3472c4150226952f34ee8d4c3db86
-
davidecaroselli authored
Summary: I have made an upgrade to my previous implementation of MMapIndexedDataset, now: - It uses up to **4 times less memory and disk space** - Words per second is slightly improved thanks to less memory access Pull Request resolved: https://github.com/pytorch/fairseq/pull/816 Differential Revision: D15899848 Pulled By: myleott fbshipit-source-id: 9ddeb4809729ef69cc6b0867b33ee71184d845e6
-
Peng-Jen Chen authored
Summary: In https://github.com/pytorch/fairseq/issues/656, people are often confused about how to set multilingual translation parameters at inference time. This diff add more checks to ensure the arguments (`--lang-pairs`, `--encoder-langtok`, `--decoder-langtok`) load from checkpoint are consistent with arguments specified in generate/interactive command line. We also add a section in example page to explain how to set the arguments Reviewed By: myleott Differential Revision: D15682169 fbshipit-source-id: 64e6db94cd72ea7ce2d0aa1067c9c2dcd3b8a2ac
-
alexeib authored
Summary: Merging wav2vec to master. Includes renames (Cpc -> wav2vec) and some light example files. Pull Request resolved: https://github.com/fairinternal/fairseq-py/pull/654 Differential Revision: D15913409 Pulled By: alexeib fbshipit-source-id: f723e6f211706cd9431c7d76dc12c4e80c9cfc80
-
Myle Ott authored
Summary: Notable (possibly breaking) changes: - d45db804: Remove checkpoint utility functions from utils.py into checkpoint_utils.py - f2563c21: Move LM definitions into separate files - dffb1674: Updates to model API: - `FairseqModel` -> `FairseqEncoderDecoderModel` - add `FairseqDecoder.extract_features` and `FairseqDecoder.output_layer` - `encoder_out_dict` -> `encoder_out` - rm unused `remove_head` functions - 34726d56: Move `distributed_init` into `DistributedFairseqModel` - cf17068a: Simplify distributed launch by automatically launching multiprocessing on each node for all visible GPUs (allows launching just one job per node instead of one per GPU) - d45db804: Change default LR scheduler from `reduce_lr_on_plateau` to `fixed` - 96ac28d3: Rename `--sampling-temperature` -> `--temperature` - fc1a19a3: Deprecate dummy batches - a1c997bd: Add memory mapped datasets - 0add50c2: Allow cycling over multiple datasets, where each one becomes an "epoch" Plus many additional features and bugfixes Pull Request resolved: https://github.com/pytorch/fairseq/pull/817 Differential Revision: D15913844 Pulled By: myleott fbshipit-source-id: d5b5d678efdd9dd3e4d7ca848ddcf1ec2b21bf6b
-
- 19 Jun, 2019 4 commits
-
-
Michael Wu authored
Summary: add flags to freeze embedding parameters and transformer layer parameters in `TransformerSentenceEncoder`. Reviewed By: myleott Differential Revision: D15866135 fbshipit-source-id: e634d7adfd5e81eacccf2b9cf6bc15bad30bd1fe
-
Myle Ott authored
Summary: Pull Request resolved: https://github.com/pytorch/fairseq/pull/811 Differential Revision: D15880880 Pulled By: myleott fbshipit-source-id: c47e09a90c945aca82b26edb4a8af93e063d5b00
-
freewym authored
Summary: …rch.distributed.ReduceOp Pull Request resolved: https://github.com/pytorch/fairseq/pull/804 Differential Revision: D15877033 Pulled By: myleott fbshipit-source-id: 58e7c39a88b67345a55b761fee4d9f211a5ee82c
-
Arya McCarthy authored
Summary: Pull Request resolved: https://github.com/pytorch/fairseq/pull/813 Pull Request resolved: https://github.com/fairinternal/fairseq-py/pull/663 Pull Request resolved: https://github.com/fairinternal/fairspeq/pull/4 Introduce new training for speech models which accept additional training data. Reviewed By: liezl200 Differential Revision: D15846661 fbshipit-source-id: 8b2cbfd56a86cf03c0b34c4a025bebdd5db7204e
-
- 15 Jun, 2019 1 commit
-
-
Myle Ott authored
Summary: Pull Request resolved: https://github.com/fairinternal/fairseq-py/pull/655 Differential Revision: D15816573 fbshipit-source-id: ac0118a1d407dc132cc7d82e029eac6c8ec76d2a
-
- 13 Jun, 2019 1 commit
-
-
Myle Ott authored
Summary: It's so much faster to extract (3 minutes instead of 20). Pull Request resolved: https://github.com/pytorch/fairseq/pull/803 Differential Revision: D15795810 Pulled By: myleott fbshipit-source-id: 3b2ae8bd7924a77ac8e795f5e1a7da0c4ae27374
-
- 12 Jun, 2019 3 commits
-
-
Nayan Singhal authored
Summary: Implemented model averaging for fairseq. Removed the ddp wrapper if global optimizer is provided. Syncing all the models based on the iteration provide in the input TODO: 1) Fix throughput and wps meter. Need to check other meters too. 2) Replace Model average code with BMUF algorithm implementation. Reviewed By: myleott Differential Revision: D15711044 fbshipit-source-id: 58a4af74db2a61d06762597b95836cbeb1ed82cc
-
Myle Ott authored
Summary: Pull Request resolved: https://github.com/pytorch/fairseq/pull/801 Differential Revision: D15781975 Pulled By: myleott fbshipit-source-id: b86276cd3a40138c09494637c43ce52a56c4aced
-
Myle Ott authored
Summary: Pull Request resolved: https://github.com/pytorch/fairseq/pull/799 Differential Revision: D15773932 Pulled By: myleott fbshipit-source-id: 650c0621bedb3b7ecebc0654d8e10d7692c50994
-
- 11 Jun, 2019 7 commits
-
-
Myle Ott authored
Summary: Pull Request resolved: https://github.com/pytorch/fairseq/pull/793 Differential Revision: D15758755 Pulled By: myleott fbshipit-source-id: b93e4ac11bde36a0b59b4d6d1c84d31c3124d767
-
Myle Ott authored
Summary: Pull Request resolved: https://github.com/pytorch/fairseq/pull/797 Differential Revision: D15761071 Pulled By: myleott fbshipit-source-id: 257d4a2297e83da7e59baed154dbafd6bfe614bf
-
Myle Ott authored
Summary: This is a temporary workaround to support sampling after https://github.com/pytorch/fairseq/issues/713. We'll need to revisit this to support sampling and beam more generally. Pull Request resolved: https://github.com/pytorch/fairseq/pull/796 Differential Revision: D15760808 Pulled By: myleott fbshipit-source-id: ecaf4f161b0c30de037f32007e4610a559a49230
-
Bairen Yi authored
Summary: See #467. Ping myleott to review. This is a work-related contribution. Ping lark to review. Pull Request resolved: https://github.com/pytorch/fairseq/pull/794 Differential Revision: D15756816 Pulled By: myleott fbshipit-source-id: 6dce3ff3a713bf5f60e5782bc260b2ca9d2c0a9b
-
Myle Ott authored
Summary: Pull Request resolved: https://github.com/pytorch/fairseq/pull/792 Differential Revision: D15741781 Pulled By: myleott fbshipit-source-id: c256c7900c307d485904e69b1526b9acbe08fec9
-
yilinyang7 authored
when given prefix_tokens, sequence generator would generate (exactly) same finished candidates (#713) Summary: https://github.com/pytorch/fairseq/issues/712 Pull Request resolved: https://github.com/pytorch/fairseq/pull/713 Differential Revision: D15242432 Pulled By: myleott fbshipit-source-id: a230ee48f4bf891c805609c428d7233a0ad21179
-
Sergey Edunov authored
Summary: Multi-Head attention is currently not TPU-friendly, specifically .data_ptr() is not supported and should not be used. Also there are potential issues with correctness of existing code (e.g. data_ptr() can point to the same storage for different tensors). Rather than rely on data_ptr() we should explicitly set self_attention or encoder_decoder_attention flags. Pull Request resolved: https://github.com/fairinternal/fairseq-py/pull/636 Reviewed By: myleott Differential Revision: D15709898 Pulled By: edunov fbshipit-source-id: f931713193c51be848a5de20da730ac3a3ce0187
-
- 10 Jun, 2019 2 commits
-
-
Myle Ott authored
Summary: - make it possible to load file_utils.py without the dependencies - add some more demo features Pull Request resolved: https://github.com/pytorch/fairseq/pull/791 Differential Revision: D15739950 Pulled By: myleott fbshipit-source-id: 38df5209973a6fe2e3651575b97134e096aaf5bf
-
freewym authored
Summary: In the current progress bar, the counter for log_interval will always start from 0, which is not correct if reloading from a checkpoint in the middle of an epoch. This fix obtains the offset from the iterator to set the counter correctly. Pull Request resolved: https://github.com/pytorch/fairseq/pull/778 Differential Revision: D15739953 Pulled By: myleott fbshipit-source-id: a1d13403ec5783b22e01d7cb63874fd8dea7f8b0
-
- 07 Jun, 2019 1 commit
-
-
Ning Dong authored
Summary: Pull Request resolved: https://github.com/pytorch/fairseq/pull/770 Without this change comment here https://fburl.com/w1cejgw9 is inconsistent with the implementation. Reviewed By: xianxl Differential Revision: D15582826 fbshipit-source-id: 16d8368560153b251beed8b290f51fcdd8a8faee
-
- 06 Jun, 2019 1 commit
-
-
Matt Le authored
Reviewed By: pipibjc Differential Revision: D15635402 fbshipit-source-id: e92fab914de40775d7bad851420355240d822bde
-
- 04 Jun, 2019 4 commits
-
-
Matt Le authored
Summary: We never actually load the model parameters from an XLM model when using tranformer_from_pretrained_xlm. Also, change encoder_learned_pos from True -> False Reviewed By: liezl200 Differential Revision: D15629061 fbshipit-source-id: 759eadc88041eae94505477960de57dd78a99dcb
-
lematt1991 authored
Summary: Resolves #762 Pull Request resolved: https://github.com/pytorch/fairseq/pull/776 Differential Revision: D15631503 Pulled By: lematt1991 fbshipit-source-id: 103f77d553476917b8b0f8001767217fb311d920
-
lematt1991 authored
Summary: Resolves #768 Pull Request resolved: https://github.com/pytorch/fairseq/pull/769 Differential Revision: D15621841 Pulled By: lematt1991 fbshipit-source-id: 694effe3788ff7d04864217d673608ec31da589e
-
Biao Lu authored
Summary: Pull Request resolved: https://github.com/fairinternal/fairseq-py/pull/630 Pull Request resolved: https://github.com/fairinternal/fairseq-py/pull/629 Pull Request resolved: https://github.com/pytorch/translate/pull/562 Pull Request resolved: https://github.com/pytorch/fairseq/pull/774 forked masked_lm_dictionary from fairseq changed import in pytorch_translate to use the new masked_lm_dictionary registered cooresponding tasks Reviewed By: liezl200 Differential Revision: D15410352 fbshipit-source-id: 06516caabdd4dc5cdee9ad1d8025978f4eea6c4b
-
- 03 Jun, 2019 2 commits
-
-
Haoran Li authored
Summary: lm_output_learned_bias doesn't exist when loading the model for fine-tuning Reviewed By: jingfeidu Differential Revision: D15579190 fbshipit-source-id: 45e8e193399943c89b77cc553d3d6d49b056e55a
-
Nathan Ng authored
Summary: Pull Request resolved: https://github.com/fairinternal/fairseq-py/pull/621 Differential Revision: D15571435 Pulled By: myleott fbshipit-source-id: 67d25b00c8c1bc69dbffd8521da56f7cc14eb75e
-
- 02 Jun, 2019 2 commits
-
-
Myle Ott authored
Summary: Pull Request resolved: https://github.com/fairinternal/fairseq-py/pull/625 Differential Revision: D15595787 Pulled By: myleott fbshipit-source-id: ba6edf305ed41be392194f492e034dd66d1743fe
-
Myle Ott authored
Summary: Pull Request resolved: https://github.com/fairinternal/fairseq-py/pull/624 Differential Revision: D15595746 Pulled By: myleott fbshipit-source-id: b79e489de9ff37ee7cbf939092a6e5ec0dbebbf5
-
- 01 Jun, 2019 1 commit
-
-
Myle Ott authored
Summary: Pull Request resolved: https://github.com/fairinternal/fairseq-py/pull/622 Differential Revision: D15572555 Pulled By: myleott fbshipit-source-id: 2b81f22207b4c894ffe645af0b45c70ac0a80612
-
- 31 May, 2019 1 commit
-
-
Myle Ott authored
Summary: Pull Request resolved: https://github.com/fairinternal/fairseq-py/pull/620 Differential Revision: D15569440 Pulled By: myleott fbshipit-source-id: c4681f1c72467c04cd2654e87bc724c94b76e3fb
-
- 30 May, 2019 5 commits
-
-
Myle Ott authored
Summary: Pull Request resolved: https://github.com/fairinternal/fairseq-py/pull/617 Differential Revision: D15555328 Pulled By: myleott fbshipit-source-id: 35d1f329f887cb0b867c7a22f17a16f3c9c66815
-
Myle Ott authored
Summary: Pull Request resolved: https://github.com/fairinternal/fairseq-py/pull/619 Differential Revision: D15562983 Pulled By: myleott fbshipit-source-id: 9240f56f18c87120b7d38e0db374d24a55999395
-
Khoa Ho authored
Summary: Change the wording to avoid confusion. Mixed precision ensures both higher arithmetic throughput and numerical stability, not exactly synonymous to pure half-precision/FP16 training. Also add mentioning of tensor cores since older generation GPUs without tensor cores don't support true mixed precision training. Pull Request resolved: https://github.com/pytorch/fairseq/pull/766 Differential Revision: D15559565 Pulled By: myleott fbshipit-source-id: c71e720772657bb3e8ad330b58bf69e23beb614e
-
Myle Ott authored
Summary: Pull Request resolved: https://github.com/fairinternal/fairseq-py/pull/613 Differential Revision: D15541384 Pulled By: myleott fbshipit-source-id: ef2c0b0a51cdf37af2ccff0546f524d49f87e65d
-
Myle Ott authored
Summary: Pull Request resolved: https://github.com/fairinternal/fairseq-py/pull/618 Differential Revision: D15552599 Pulled By: myleott fbshipit-source-id: 2192a30a9c5af31b954a3a1716166dd6ba27b23a
-