- 15 May, 2019 4 commits
-
-
Myle Ott authored
Summary: - `FairseqModel` -> `FairseqEncoderDecoderModel` - add `FairseqDecoder.extract_features` and `FairseqDecoder.output_layer` - `encoder_out_dict` -> `encoder_out` - rm unused `remove_head` functions - update docs Pull Request resolved: https://github.com/fairinternal/fairseq-py/pull/561 Differential Revision: D15271142 Pulled By: myleott fbshipit-source-id: 8e8864e399336020f0271c780598e968ff51a264
-
Myle Ott authored
Summary: Pull Request resolved: https://github.com/fairinternal/fairseq-py/pull/578 Differential Revision: D15352060 Pulled By: myleott fbshipit-source-id: 7dc2fceca37ec96c89356662831b0d82f28bef6f
-
Myle Ott authored
Summary: Pull Request resolved: https://github.com/fairinternal/fairseq-py/pull/579 Differential Revision: D15352058 Pulled By: myleott fbshipit-source-id: cebef02edcfcb203ef2e32c64f7f28e08c4e46b0
-
Myle Ott authored
Summary: Various fixes for Masked LM - use --activation-fn instead of --gelu - use --dataset-impl instead of --lazy-load - add embed_scale option to TransformerSentenceEncoder - fix encoder_normalize_before to include a final layer norm - delete BertLayerNorm Pull Request resolved: https://github.com/fairinternal/fairseq-py/pull/573 Reviewed By: borguz Differential Revision: D15317933 Pulled By: myleott fbshipit-source-id: 8ecb46556ad43e76e92d41ed8f5a62e8516fd375
-
- 14 May, 2019 3 commits
-
-
Myle Ott authored
Summary: Pull Request resolved: https://github.com/fairinternal/fairseq-py/pull/575 Differential Revision: D15318004 Pulled By: myleott fbshipit-source-id: ad918d71b1bd8074decf5ec3463dd9bc9487bbe9
-
Nayan Singhal authored
Summary: 1. Define a EpochMinibatchIterator which extends the EpochBatchIterator. It has same functionality as EpochBatchIterator except two major changes: use static batching and use MiniBatchIterator for getting the indices. 2. SplitSeqCollater is used instead of Seq2SeqCollater. 3. LSTM_subsample started storing the previous states and reset it once the sample is over. Reviewed By: jay-mahadeokar Differential Revision: D15209023 fbshipit-source-id: 900b8bd1f25159ffc77f8106e26729a3e7422a1f
-
Dmytro Okhonko authored
Summary: Move `load_checkpoint`, `save_checkpoint` and `reload_train` from train.py to checkpoint_utils.py Move `get_perplexity` from train.py to utils.py. This will make train.py lighter and allow us to reuse all this utils functionality when fairseq is used as external library. Reviewed By: myleott Differential Revision: D15289607 fbshipit-source-id: 4b7c95225ac22e402bcda3497811361809110df1
-
- 13 May, 2019 4 commits
-
-
Myle Ott authored
Summary: Pull Request resolved: https://github.com/fairinternal/fairseq-py/pull/576 Differential Revision: D15318086 Pulled By: myleott fbshipit-source-id: c6587737ca7b97edc97ad4aef5c5c9ac7e92b2f2
-
Myle Ott authored
Summary: This was named gelu_fast after the original implementation: https://github.com/hendrycks/GELUs/blob/master/mnist_ae.py#L62-L63 But in practice it's actually slower and uses more memory. Rename to gelu_accurate. Pull Request resolved: https://github.com/fairinternal/fairseq-py/pull/571 Differential Revision: D15317874 Pulled By: myleott fbshipit-source-id: c96fbc89bf91b27ced1ab8d5b25a8f23f922ec24
-
Myle Ott authored
Summary: Pull Request resolved: https://github.com/fairinternal/fairseq-py/pull/574 Differential Revision: D15317984 Pulled By: myleott fbshipit-source-id: 09a66229cc6b4c95678ca1ca13c9e0da25b203de
-
Myle Ott authored
Summary: Pull Request resolved: https://github.com/fairinternal/fairseq-py/pull/572 Differential Revision: D15317928 Pulled By: myleott fbshipit-source-id: b3f0e9229737a63b49937e7c5b918470f18ddc45
-
- 12 May, 2019 2 commits
-
-
zhiqiang authored
Summary: `--output-format` -> `--dataset-impl` in Tutorial: Classifying Names with a Character-Level RNN Pull Request resolved: https://github.com/pytorch/fairseq/pull/735 Differential Revision: D15314625 Pulled By: myleott fbshipit-source-id: 65b8efd1a367ca754e5b9dca088aefbc648864dd
-
Myle Ott authored
Summary: Pull Request resolved: https://github.com/pytorch/fairseq/pull/736 Differential Revision: D15314626 Pulled By: myleott fbshipit-source-id: 1e0c32529afee57e43fe5d6c7991cd13eb8a52c4
-
- 11 May, 2019 2 commits
-
-
Naman Goyal authored
Summary: Pull Request resolved: https://github.com/fairinternal/fairseq-py/pull/568 Differential Revision: D15308483 Pulled By: myleott fbshipit-source-id: 9d898ce523e46e6b6fb444274f478da0b577b603
-
Myle Ott authored
Summary: Pull Request resolved: https://github.com/fairinternal/fairseq-py/pull/560 Differential Revision: D15260838 Pulled By: myleott fbshipit-source-id: 5f80dd82775c10ce46a3e1c451ccaf0ef55bfa31
-
- 10 May, 2019 2 commits
-
-
Jay Mahadeokar authored
Summary: As in title. Reviewed By: skritika Differential Revision: D15299135 fbshipit-source-id: 2fd513b32c0ab41911cdf0b0186f6c3bb5256285
-
myleott authored
-
- 09 May, 2019 5 commits
-
-
Myle Ott authored
Set initial learning rate in LR schedulers by calling step_update(0) at init
-
Myle Ott authored
-
Myle Ott authored
Summary: Pull Request resolved: https://github.com/fairinternal/fairseq-py/pull/564 Differential Revision: D15278017 Pulled By: myleott fbshipit-source-id: b6fba1b62145ea533b40f5eb9b134e6aa122e546
-
Jingfei Du authored
Summary: the old no_bias_kv argument for masked_lm models are not used. Split it into 2 arguments and expose them. Reviewed By: myleott Differential Revision: D15266154 fbshipit-source-id: 60b041f8370ca1d8869ed3402fb9a67d1cd8e0e8
-
- 08 May, 2019 7 commits
-
-
Myle Ott authored
Reviewed By: jmp84 Differential Revision: D15264847 fbshipit-source-id: 4ba9224d1b35c3de0d26c9b4c1ee6d641d3d8535
-
Myle Ott authored
Summary: Pull Request resolved: https://github.com/pytorch/fairseq/pull/723 Differential Revision: D15260870 Pulled By: myleott fbshipit-source-id: 73d9b138b9ab44f96824076258f1a6319193d0f7
-
Naman Goyal authored
Summary: 1) Made the model compatible with using either `masked_lm_dataset` or `monolingual_dataset`. 2) fixed default args setting task. (`bert` vs `masked_lm`) myleott should we keep both? 3) bug in setting default value of `sentence_class_num` 4) bug for padding mask in `fp16`. Pull Request resolved: https://github.com/pytorch/fairseq/pull/721 Differential Revision: D15259885 fbshipit-source-id: 9dbf7fb8192992c1251670287bed719e41c08fcc
-
Myle Ott authored
Summary: Pull Request resolved: https://github.com/pytorch/fairseq/pull/720 Differential Revision: D15259091 Pulled By: myleott fbshipit-source-id: 06a35996c06ccddb49fdc9e01e348ff3c9da334e
-
Myle Ott authored
Summary: Pull Request resolved: https://github.com/pytorch/fairseq/pull/719 Differential Revision: D15258483 Pulled By: myleott fbshipit-source-id: dd00daa6f1c87264c1196a77dfffc8c876ebde7f
-
Myle Ott authored
Summary: Pull Request resolved: https://github.com/pytorch/fairseq/pull/717 Differential Revision: D15254560 Pulled By: myleott fbshipit-source-id: 2a07614e8d294636f706939e60f0091c73115494
-
Jay Mahadeokar authored
Summary: D15214049 introduced a bug such that if a tasks args does not contain data, then it will give error ``` File "/data/users/jaym/fbsource/fbcode/buck-out/dev/gen/deeplearning/projects/fairspeq/train#link-tree/train.py", line 119, in reload_train if len(args.data.split(":")) == 1: AttributeError: 'Namespace' object has no attribute 'data' ``` This diff checks if data is in args to avoid above error. Reviewed By: myleott, jmp84 Differential Revision: D15253373 fbshipit-source-id: 14fb9ad878ee50f1b7583349bb17e29c03c40815
-
- 07 May, 2019 5 commits
-
-
Naman Goyal authored
Summary: Pull Request resolved: https://github.com/pytorch/fairseq/pull/715 Differential Revision: D15240723 fbshipit-source-id: 11d7280cb187d68f107902822e878f2a04b840c7
-
taineleau authored
Summary: Pull Request resolved: https://github.com/pytorch/fairseq/pull/711 Differential Revision: D15239618 Pulled By: myleott fbshipit-source-id: 82f3f79501a13a967324b8a66281cd134bf1ef23
-
Davide Caroselli authored
Summary: Following discussion in https://github.com/pytorch/fairseq/issues/574: - Implemented MMapIndexedDataset and MMapIndexedDatasetBuilder compatible with IndexedDataset/IndexedDatasetBuilder - Update scripts/read_binarized.py to support new MMapIndexedDataset - Option '--raw-text' and '--lazy-load' replaced with '--dataset-impl' and moved the option definition custom task args to more high-level options.add_dataset_args() (more appropriate) - Implemented also utils functions in indexed_dataset: make_dataset(), dataset_exists() Pull Request resolved: https://github.com/pytorch/fairseq/pull/589 Differential Revision: D14597128 Pulled By: myleott fbshipit-source-id: 4e92d99920cbaa52cfe5a0f1f5d9ae5c92d4268e
-
Myle Ott authored
Summary: Pull Request resolved: https://github.com/pytorch/fairseq/pull/704 Differential Revision: D15221549 Pulled By: myleott fbshipit-source-id: b0021acdc2d7792ce51421f1432e1f2bd8218f7b
-
Kartikay Khandelwal authored
Summary: Pull Request resolved: https://github.com/pytorch/fairseq/pull/710 Previously there was a bug in how we dealt with padding when computing the input representation from the segment and position embedding. D15144912 fixed this by adding an offset based on the padding id. However this makes assumptions about the padding id which may not hold true for vocabularies built outside of pyText and fairseq. Based on a discussion with barlaso, this diff 0's out all the embeddings associated with the padding. Reviewed By: borguz Differential Revision: D15209395 fbshipit-source-id: 5573020e610f5466e673fe3845c3ed34ebb5c44d
-
- 06 May, 2019 5 commits
-
-
Naman Goyal authored
Summary: Co-authored-by:
myleott <myleott@fb.com> Changing `data` to be `str` with colon separated list for loading sharded datasets. This change is useful for loading large datasets that cannot fit into, memory. The large dataset can be sharded and then each shard is loaded in one epoch in roudrobin manner. For example, if there are `5` shards of data and `10` epochs then the shards will be iterated upon `[0, 1, 2, 3, 4, 0, 1, 2, 3, 4]`. myleott We need to look into `translation.py` as it currently already expects a list and then concats the datasets. Pull Request resolved: https://github.com/pytorch/fairseq/pull/696 Differential Revision: D15214049 fbshipit-source-id: 03e43a7b69c7aefada2ca668abf1eac1969fe013
-
Myle Ott authored
Summary: Pull Request resolved: https://github.com/pytorch/fairseq/pull/707 Differential Revision: D15219014 Pulled By: myleott fbshipit-source-id: f38f2cf817d05e0871ff9084a810d109848e827c
-
Naman Goyal authored
Summary: Co-authored-by:
jingfeidu <jingfeidu@fb.com> 1) Adding `masked_lm` task for BERT like training. Code mostly taken from jingfeidu 's implementation. 2) Added `has_eos` option to `block_pair_dataset` for working with dataset that has been preprocessed with having `eos`. Depends on: https://github.com/pytorch/fairseq/pull/696 Pull Request resolved: https://github.com/pytorch/fairseq/pull/697 Differential Revision: D15214050 fbshipit-source-id: c179ce2d70e59d2ddc941b13ceda99d929878931
-
Maksym Del authored
Summary: Pass required "sample_key" argument to forward-backward call in semi-supervised task. Pull Request resolved: https://github.com/pytorch/fairseq/pull/706 Differential Revision: D15217957 Pulled By: pipibjc fbshipit-source-id: bf943d566c5caa67682dfb16ff8b7c432323cdba
-
Liezl Puzon authored
Summary: Pull Request resolved: https://github.com/pytorch/fairseq/pull/705 This adds functionality in fairseq to load a pretrained encoder or decoder from another pretrained model into the current model. Reviewed By: jmp84 Differential Revision: D15207084 fbshipit-source-id: 32a710ff77389928e20793c71d312863df9dd8ae
-
- 05 May, 2019 1 commit
-
-
Myle Ott authored
Reviewed By: chenyangyu1988 Differential Revision: D14784219 fbshipit-source-id: 273888d6e3d22a01d5e7edfbc786195e7b78efef
-