Commit 57b6a6db authored by Liezl Puzon's avatar Liezl Puzon Committed by Facebook Github Bot
Browse files

Fix fairseq unittest timeouts (#667)

Summary:
Pull Request resolved: https://github.com/pytorch/fairseq/pull/667

Use smaller models so that unittests won't timeout

Reviewed By: pipibjc

Differential Revision: D15056894

fbshipit-source-id: af9fbda6ea6e56cf82d52555620121b189e2f013
parent 5008fd4e
...@@ -114,6 +114,9 @@ class TestTranslation(unittest.TestCase): ...@@ -114,6 +114,9 @@ class TestTranslation(unittest.TestCase):
train_translation_model(data_dir, 'lstm_wiseman_iwslt_de_en', [ train_translation_model(data_dir, 'lstm_wiseman_iwslt_de_en', [
'--encoder-layers', '2', '--encoder-layers', '2',
'--decoder-layers', '2', '--decoder-layers', '2',
'--encoder-embed-dim', '8',
'--decoder-embed-dim', '8',
'--decoder-out-embed-dim', '8',
]) ])
generate_main(data_dir) generate_main(data_dir)
...@@ -125,7 +128,10 @@ class TestTranslation(unittest.TestCase): ...@@ -125,7 +128,10 @@ class TestTranslation(unittest.TestCase):
train_translation_model(data_dir, 'lstm', [ train_translation_model(data_dir, 'lstm', [
'--encoder-layers', '2', '--encoder-layers', '2',
'--encoder-bidirectional', '--encoder-bidirectional',
'--encoder-hidden-size', '256', '--encoder-hidden-size', '16',
'--encoder-embed-dim', '8',
'--decoder-embed-dim', '8',
'--decoder-out-embed-dim', '8',
'--decoder-layers', '2', '--decoder-layers', '2',
]) ])
generate_main(data_dir) generate_main(data_dir)
...@@ -135,7 +141,12 @@ class TestTranslation(unittest.TestCase): ...@@ -135,7 +141,12 @@ class TestTranslation(unittest.TestCase):
with tempfile.TemporaryDirectory('test_transformer') as data_dir: with tempfile.TemporaryDirectory('test_transformer') as data_dir:
create_dummy_data(data_dir) create_dummy_data(data_dir)
preprocess_translation_data(data_dir) preprocess_translation_data(data_dir)
train_translation_model(data_dir, 'transformer_iwslt_de_en') train_translation_model(data_dir, 'transformer_iwslt_de_en', [
'--encoder-layers', '2',
'--decoder-layers', '2',
'--encoder-embed-dim', '8',
'--decoder-embed-dim', '8',
])
generate_main(data_dir) generate_main(data_dir)
def test_lightconv(self): def test_lightconv(self):
...@@ -146,6 +157,8 @@ class TestTranslation(unittest.TestCase): ...@@ -146,6 +157,8 @@ class TestTranslation(unittest.TestCase):
train_translation_model(data_dir, 'lightconv_iwslt_de_en', [ train_translation_model(data_dir, 'lightconv_iwslt_de_en', [
'--encoder-conv-type', 'lightweight', '--encoder-conv-type', 'lightweight',
'--decoder-conv-type', 'lightweight', '--decoder-conv-type', 'lightweight',
'--encoder-embed-dim', '8',
'--decoder-embed-dim', '8',
]) ])
generate_main(data_dir) generate_main(data_dir)
...@@ -157,6 +170,8 @@ class TestTranslation(unittest.TestCase): ...@@ -157,6 +170,8 @@ class TestTranslation(unittest.TestCase):
train_translation_model(data_dir, 'lightconv_iwslt_de_en', [ train_translation_model(data_dir, 'lightconv_iwslt_de_en', [
'--encoder-conv-type', 'dynamic', '--encoder-conv-type', 'dynamic',
'--decoder-conv-type', 'dynamic', '--decoder-conv-type', 'dynamic',
'--encoder-embed-dim', '8',
'--decoder-embed-dim', '8',
]) ])
generate_main(data_dir) generate_main(data_dir)
...@@ -170,6 +185,10 @@ class TestTranslation(unittest.TestCase): ...@@ -170,6 +185,10 @@ class TestTranslation(unittest.TestCase):
'--method', 'hMoElp', '--method', 'hMoElp',
'--mean-pool-gating-network', '--mean-pool-gating-network',
'--num-experts', '3', '--num-experts', '3',
'--encoder-layers', '2',
'--decoder-layers', '2',
'--encoder-embed-dim', '8',
'--decoder-embed-dim', '8',
]) ])
generate_main(data_dir, [ generate_main(data_dir, [
'--task', 'translation_moe', '--task', 'translation_moe',
...@@ -188,13 +207,17 @@ class TestStories(unittest.TestCase): ...@@ -188,13 +207,17 @@ class TestStories(unittest.TestCase):
create_dummy_data(data_dir) create_dummy_data(data_dir)
preprocess_translation_data(data_dir) preprocess_translation_data(data_dir)
config = [ config = [
'--encoder-layers', '[(512, 3)] * 2', '--encoder-layers', '[(128, 3)] * 2',
'--decoder-layers', '[(512, 3)] * 2', '--decoder-layers', '[(128, 3)] * 2',
'--decoder-attention', 'True', '--decoder-attention', 'True',
'--encoder-attention', 'False', '--encoder-attention', 'False',
'--gated-attention', 'True', '--gated-attention', 'True',
'--self-attention', 'True', '--self-attention', 'True',
'--project-input', 'True', '--project-input', 'True',
'--encoder-embed-dim', '8',
'--decoder-embed-dim', '8',
'--decoder-out-embed-dim', '8',
'--multihead-self-attention-nheads', '2'
] ]
train_translation_model(data_dir, 'fconv_self_att_wp', config) train_translation_model(data_dir, 'fconv_self_att_wp', config)
generate_main(data_dir) generate_main(data_dir)
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment