"torchvision/csrc/ops/autograd/roi_pool_kernel.cpp" did not exist on "7f7e7663e1bd5b8ae4e6f4a0b561319777769bdf"
- 21 Nov, 2019 2 commits
-
-
ngoyal2707 authored
Summary: Pull Request resolved: https://github.com/fairinternal/fairseq-py/pull/922 Differential Revision: D18617322 fbshipit-source-id: 50645197cb7f075b5f878818a97358653077c3e0
-
Alex Xiao authored
Summary: Modifying number of shards internally to disable data sharding for batch iteration is dangerous because the caller of these tasks is not limited to fairspeq/train. So therefore we should put the onus of data sharding properly on the caller rather than the task itself. Reviewed By: myleott Differential Revision: D18456424 fbshipit-source-id: d46be16c441c50082f9a768d0b259e6c28a4b67b
-
- 20 Nov, 2019 1 commit
-
-
Jiatao Gu authored
Summary: Clean up the original NAT loss and make it more general to adapt new losses used in NAT models. Pull Request resolved: https://github.com/fairinternal/fairseq-py/pull/921 Differential Revision: D18610145 Pulled By: MultiPath fbshipit-source-id: d04dd0fc4047b5f8e332cfe66b1e28cbf39494af
-
- 19 Nov, 2019 3 commits
-
-
Naman Goyal authored
Summary: Pull Request resolved: https://github.com/fairinternal/fairseq-py/pull/920 Differential Revision: D18593088 fbshipit-source-id: d4479ee8dae2ca623e62e12bd145165a116fb70a
-
freewym authored
Summary: …r to correctly recover the training from a "non-shuffle" checkpoint Pull Request resolved: https://github.com/pytorch/fairseq/pull/1375 Differential Revision: D18566535 Pulled By: myleott fbshipit-source-id: ff7b1a6ead708801f537ec7885e30e37168cd34b
-
ngoyal2707 authored
Summary: Pull Request resolved: https://github.com/fairinternal/fairseq-py/pull/915 Differential Revision: D18580996 fbshipit-source-id: 9505a81892ba8ad997c03465d6a2d488c379c762
-
- 18 Nov, 2019 3 commits
-
-
alexeib authored
Summary: recent layerdrop related changes break existing models because they assume presence of certain args Pull Request resolved: https://github.com/fairinternal/fairseq-py/pull/918 Reviewed By: huihuifan Differential Revision: D18578572 Pulled By: alexeib fbshipit-source-id: 368c2d5b3add55864bf59516820807303aac6001
-
Myle Ott authored
Summary: Fixes https://github.com/pytorch/fairseq/issues/1376 Pull Request resolved: https://github.com/pytorch/fairseq/pull/1386 Differential Revision: D18566839 Pulled By: myleott fbshipit-source-id: 71805f58fab90f53f757bf4ef69eb914195af38a
-
Louis Martin authored
Summary: Pull Request resolved: https://github.com/fairinternal/fairseq-py/pull/913 Differential Revision: D18565866 Pulled By: myleott fbshipit-source-id: e845759dafe915805c2e38f53c6835cbcef5db2f
-
- 17 Nov, 2019 1 commit
-
-
Angela Fan authored
Summary: Pull Request resolved: https://github.com/pytorch/fairseq/pull/1385 Differential Revision: D18565188 Pulled By: huihuifan fbshipit-source-id: 9580663b208f286a249bbfa2bacd71f34a01ca9f
-
- 15 Nov, 2019 1 commit
-
-
Ilia Cherniavskii authored
Summary: Pull Request resolved: https://github.com/fairinternal/fairseq-py/pull/887 Pull Request resolved: https://github.com/facebookresearch/pytext/pull/1052 Pull Request resolved: https://github.com/pytorch/fairseq/pull/1250 Adding config parameter "use_torchscript" that enables use of TS for BERT training Reviewed By: chenyangyu1988 Differential Revision: D17872083 fbshipit-source-id: 00ac4b04e7f26aa56fe84fe9feaded676d6deb71
-
- 14 Nov, 2019 4 commits
-
-
Myle Ott authored
Summary: Pull Request resolved: https://github.com/fairinternal/fairseq-py/pull/911 Differential Revision: D18511627 Pulled By: myleott fbshipit-source-id: 37d7606ae629f9acf84715dbc9045fb683075db4
-
freewym authored
Summary: If the training stopped in the middle of the last epoch, and then it was resumed from checkpoint, it will not continue the training because `epoch_itr.epoch < max_epoch` is not satisfied. This PR fixed the issue. Pull Request resolved: https://github.com/pytorch/fairseq/pull/1275 Differential Revision: D18483945 Pulled By: myleott fbshipit-source-id: 80df6f73fa17606a79a28e8328bb4c577f504683
-
Abhimanyu Sharma authored
Summary: Pull Request resolved: https://github.com/fairinternal/fairseq-py/pull/910 Pull Request resolved: https://github.com/facebookresearch/pytext/pull/1124 Pull Request resolved: https://github.com/pytorch/fairseq/pull/1362 Split the Fariseq MemoryEfficientFP16Optimizer class into 2 classes so that it can be easily imported in pytext through a wrapper class. Iter 2 - fixed some issues to ensure code runs correctly on fblearner. Iter 3 - fixed review comments, incorrect import and lints. Iter 4 - fixed pytext test breaks. Iter 5 - fix pytext test breaks. Iter 6 - fix comments and refactor based on conversation with chenyang. Reviewed By: chenyangyu1988 Differential Revision: D18410916 fbshipit-source-id: 5238ee553cd2811ed0573825e1c29000980cc489
-
Jiatao Gu authored
Summary: (1) Enable to print the iterative refinement history for all NAT models by setting --retain-iter-history during decoding; (2) Fix a small bug in the decoding process in Levenshtein Transformer. Pull Request resolved: https://github.com/fairinternal/fairseq-py/pull/908 Differential Revision: D18493234 Pulled By: MultiPath fbshipit-source-id: 9e7702adcea49f39d3c10b5349b5a9ae66399a24
-
- 13 Nov, 2019 5 commits
-
-
Myle Ott authored
Summary: Pull Request resolved: https://github.com/fairinternal/fairseq-py/pull/896 Differential Revision: D18250948 Pulled By: myleott fbshipit-source-id: 7a515311e18795670b29f5e24eeba7619a625da7
-
zheng authored
Summary: As their names suggest, the parameters `embedding_dim`, `ffn_embedding_dim`, and `num_attention_heads` should have type `int`, not `float`. Also validated by https://github.com/pytorch/fairseq/blob/b5f41f828b0ec9b67fa60aceb0778073d1b368b2/fairseq/modules/sparse_transformer_sentence_encoder.py#L22#L24. Pull Request resolved: https://github.com/pytorch/fairseq/pull/1268 Differential Revision: D18372518 Pulled By: myleott fbshipit-source-id: 666739b6270a975536785886068a975e07312bb0
-
Zhanghao Wu authored
Summary: Originally, the 'ppl' is calculated but returned as a string, which will not be printed to the tensorboard. Pull Request resolved: https://github.com/pytorch/fairseq/pull/1212 Differential Revision: D18339553 Pulled By: myleott fbshipit-source-id: 52e64d5d173bfd79836a72ee103cb25c8bb2a4c2
-
Myle Ott authored
Summary: Pull Request resolved: https://github.com/fairinternal/fairseq-py/pull/907 Differential Revision: D18480215 Pulled By: myleott fbshipit-source-id: b02002f631f6d47380f309d4f464bd135d623280
-
Myle Ott authored
Summary: Pull Request resolved: https://github.com/fairinternal/fairseq-py/pull/899 Differential Revision: D18373060 Pulled By: myleott fbshipit-source-id: bb5510ec15799a0a10a7c0669e76d8200e1ba479
-
- 12 Nov, 2019 1 commit
-
-
Spencer Poff authored
Summary: Using PyTorch IterableDataset for streaming iterators. Such that there is a clean differentiation in interface between datasets that are streaming data and those that support indexed access. Reviewed By: myleott Differential Revision: D18438694 fbshipit-source-id: 482857d8357091ea2a6bf819535b09ba7f1a5b7d
-
- 10 Nov, 2019 1 commit
-
-
Louis Martin authored
Summary: Check locally that everything works fine. Model is uploaded to fbaipublicfiles. I fixed a few inconsistencies in the bpe encoding along the way, e.g. related to https://github.com/pytorch/fairseq/issues/1306.. Pull Request resolved: https://github.com/fairinternal/fairseq-py/pull/904 Reviewed By: ngoyal2707 Differential Revision: D18418345 Pulled By: louismartin fbshipit-source-id: 53acb4d021581968d70430ee9babee07d6573c17
-
- 09 Nov, 2019 1 commit
-
-
Naman Goyal authored
Summary: This is the first version of BART code / model release. It still requires lot of clean up, instructions, making sure results are reproducible before we can release it. Pull Request resolved: https://github.com/fairinternal/fairseq-py/pull/902 Differential Revision: D18389535 fbshipit-source-id: 77f16800307ce831bd29538fdd34800793210f46
-
- 08 Nov, 2019 2 commits
-
-
Myle Ott authored
Summary: Pull Request resolved: https://github.com/fairinternal/fairseq-py/pull/903 Reviewed By: sujitoc Differential Revision: D18327653 fbshipit-source-id: 739ddbaf54862acdf7b4f1bc3ad538bde5ae00fd
-
Xian Li authored
Summary: To avoid the case where can_ins_mask has all False so max_lengths has size [0, 1] which failed expand_as operator. Move it back into the skipping branch in script. The same for deletion and ins_word. Reviewed By: kahne Differential Revision: D18365340 fbshipit-source-id: 509ac21d7d6fd9083d0710697288203977314c52
-
- 07 Nov, 2019 4 commits
-
-
Kevin authored
Summary: Solves https://github.com/pytorch/fairseq/issues/1218. Pull Request resolved: https://github.com/pytorch/fairseq/pull/1219 Differential Revision: D18339541 Pulled By: myleott fbshipit-source-id: 6d5bd7b60fa7fd30c038fdad54591343a01f228b
-
Louis MARTIN authored
Summary: Models seem to train fine with this modification. I checked that the mask for beginning of words is correct but didn't check if the actual masking worked correctly. Pull Request resolved: https://github.com/pytorch/fairseq/pull/1292 Differential Revision: D18338307 Pulled By: myleott fbshipit-source-id: eae9e29d6ab648e768d70921694a898554496704
-
freewym authored
Summary: …all set_epoch() for each sub dataset Pull Request resolved: https://github.com/pytorch/fairseq/pull/1272 Differential Revision: D18338300 Pulled By: myleott fbshipit-source-id: 973d57f52c5cf4ad40122d4a625942281c7983b7
-
Liam authored
Summary: "pytorch.fairseq" -> "pytorch/fairseq" to avoid following error: ``` ValueError: not enough values to unpack (expected 2, got 1) Pull Request resolved: https://github.com/pytorch/fairseq/pull/1310 Differential Revision: D18338223 Pulled By: myleott fbshipit-source-id: c95fcc3bb814c7f980a22996dc7923d6d487810b
-
- 06 Nov, 2019 2 commits
-
-
Naman Goyal authored
Summary: Pull Request resolved: https://github.com/fairinternal/fairseq-py/pull/901 Differential Revision: D18349686 fbshipit-source-id: ba0a378e3fb98a35b3ef2e2103c2f921c4729e40
-
Jerry Ma authored
Summary: - Adds memory summary logging to validation and optimization steps. - Clarifies in the logging that optimization OOMs are not recoverable. Pull Request resolved: https://github.com/fairinternal/fairseq-py/pull/893 Differential Revision: D18110763 Pulled By: jma127 fbshipit-source-id: 49340e611169c606ab9c991265167a79f51846e6
-
- 05 Nov, 2019 2 commits
-
-
ngoyal2707 authored
Summary: TODO: 1) Need to update bibtex entry 2) Need to upload models, spm_vocab and dict.txt to public s3 location. For Future: 1) I will probably add instructions to finetune on XNLI and NER, POS etc. but currently no timeline for that. Pull Request resolved: https://github.com/fairinternal/fairseq-py/pull/900 Reviewed By: myleott Differential Revision: D18333076 Pulled By: myleott fbshipit-source-id: 3f3d3716fcc41c78d2dd4525f60b519abbd0459c
-
Spencer Poff authored
Summary: https://github.com/pytorch/fairseq/pull/1097 added key padding mask history in TransformerDecoderLayer, but during an edge case where only the current or only the previous key_padding_mask exists, the resulting key_padding_mask is the wrong size. This diff adds empty columns in such a case to ensure key_padding_mask is a usable size. Reviewed By: myleott Differential Revision: D18224313 fbshipit-source-id: c9fb7266baf0a2d79a66704e00a5ea8bd2987ff6
-
- 02 Nov, 2019 1 commit
-
-
Myle Ott authored
Summary: Pull Request resolved: https://github.com/pytorch/fairseq/pull/1340 Differential Revision: D18289455 Pulled By: myleott fbshipit-source-id: a1c8163a35273b6c646d300142701e8a317d7378
-
- 01 Nov, 2019 2 commits
-
-
Chau Tran authored
Summary: Fix integration test Reviewed By: xianxl Differential Revision: D18040440 fbshipit-source-id: 98c8ab7970d081f17deb54c69aa35669de12d767
-
Halil Akin authored
Summary: Pull Request resolved: https://github.com/fairinternal/fairseq-py/pull/898 Pull Request resolved: https://github.com/pytorch/fairseq/pull/1333 Pull Request resolved: https://github.com/fairinternal/fairspeq/pull/11 This in_proj_weight and in_proj_bias properties are not the right way of providing backward compatibility, and it's causing other incompatibilities with the new Dynamic Quantization API. So, let's remove this, and properly fix the failing tests. Reviewed By: myleott Differential Revision: D18264129 fbshipit-source-id: fc1838657a60d914ca83c4e0f6add5ed8206ac54
-
- 31 Oct, 2019 2 commits
-
-
Myle Ott authored
Summary: Pull Request resolved: https://github.com/fairinternal/fairseq-py/pull/897 Differential Revision: D18250587 Pulled By: myleott fbshipit-source-id: b9cef376bc014b68766229aab7b6e454480757d3
-
Myle Ott authored
Summary: Pull Request resolved: https://github.com/fairinternal/fairseq-py/pull/895 Reviewed By: akinh Differential Revision: D18246479 Pulled By: myleott fbshipit-source-id: a610f1e4943619d32a523601a572fb09cdc5638d
-
- 30 Oct, 2019 1 commit
-
-
Xian Li authored
Summary: This diff enables layer drop in transformer decoder in production training pipeline (ptt_transformer). It builds on top of the fairseq implementation D18094657 added by Angela Fan, and added additional logic to handle corresponding dropping layers at test time in exported model. Reviewed By: jhcross Differential Revision: D18165586 fbshipit-source-id: 373ac00268a25fa9e412edcb483becdfe792d992
-
- 28 Oct, 2019 1 commit
-
-
Ning Dong authored
Summary: Revert the interface change for iterative_refinement_generator Reviewed By: kahne Differential Revision: D18165103 fbshipit-source-id: 075c276746eb90d7c359b6ad92e1ef25e8452bcc
-