Unverified Commit 4f6e5257 authored by Stas Bekman's avatar Stas Bekman Committed by GitHub
Browse files

model card improvements (#7221)

parent eb074af7
...@@ -8,7 +8,7 @@ tags: ...@@ -8,7 +8,7 @@ tags:
- translation - translation
- wmt16 - wmt16
- allenai - allenai
license: Apache 2.0 license: apache-2.0
datasets: datasets:
- wmt16 - wmt16
metrics: metrics:
...@@ -29,16 +29,6 @@ All 3 models are available: ...@@ -29,16 +29,6 @@ All 3 models are available:
* [wmt16-en-de-dist-6-1](https://huggingface.co/allenai/wmt16-en-de-dist-6-1) * [wmt16-en-de-dist-6-1](https://huggingface.co/allenai/wmt16-en-de-dist-6-1)
* [wmt16-en-de-12-1](https://huggingface.co/allenai/wmt16-en-de-12-1) * [wmt16-en-de-12-1](https://huggingface.co/allenai/wmt16-en-de-12-1)
```
@misc{kasai2020deep,
title={Deep Encoder, Shallow Decoder: Reevaluating the Speed-Quality Tradeoff in Machine Translation},
author={Jungo Kasai and Nikolaos Pappas and Hao Peng and James Cross and Noah A. Smith},
year={2020},
eprint={2006.10369},
archivePrefix={arXiv},
primaryClass={cs.CL}
}
```
## Intended uses & limitations ## Intended uses & limitations
...@@ -97,3 +87,18 @@ PYTHONPATH="src:examples/seq2seq" python examples/seq2seq/run_eval.py allenai/wm ...@@ -97,3 +87,18 @@ PYTHONPATH="src:examples/seq2seq" python examples/seq2seq/run_eval.py allenai/wm
- [training, etc.](http://www.statmt.org/wmt16/) - [training, etc.](http://www.statmt.org/wmt16/)
- [test set](http://matrix.statmt.org/test_sets/newstest2016.tgz?1504722372) - [test set](http://matrix.statmt.org/test_sets/newstest2016.tgz?1504722372)
### BibTeX entry and citation info
```
@misc{kasai2020deep,
title={Deep Encoder, Shallow Decoder: Reevaluating the Speed-Quality Tradeoff in Machine Translation},
author={Jungo Kasai and Nikolaos Pappas and Hao Peng and James Cross and Noah A. Smith},
year={2020},
eprint={2006.10369},
archivePrefix={arXiv},
primaryClass={cs.CL}
}
```
...@@ -8,7 +8,7 @@ tags: ...@@ -8,7 +8,7 @@ tags:
- translation - translation
- wmt16 - wmt16
- allenai - allenai
license: Apache 2.0 license: apache-2.0
datasets: datasets:
- wmt16 - wmt16
metrics: metrics:
...@@ -29,16 +29,6 @@ All 3 models are available: ...@@ -29,16 +29,6 @@ All 3 models are available:
* [wmt16-en-de-dist-6-1](https://huggingface.co/allenai/wmt16-en-de-dist-6-1) * [wmt16-en-de-dist-6-1](https://huggingface.co/allenai/wmt16-en-de-dist-6-1)
* [wmt16-en-de-12-1](https://huggingface.co/allenai/wmt16-en-de-12-1) * [wmt16-en-de-12-1](https://huggingface.co/allenai/wmt16-en-de-12-1)
```
@misc{kasai2020deep,
title={Deep Encoder, Shallow Decoder: Reevaluating the Speed-Quality Tradeoff in Machine Translation},
author={Jungo Kasai and Nikolaos Pappas and Hao Peng and James Cross and Noah A. Smith},
year={2020},
eprint={2006.10369},
archivePrefix={arXiv},
primaryClass={cs.CL}
}
```
## Intended uses & limitations ## Intended uses & limitations
...@@ -97,3 +87,18 @@ PYTHONPATH="src:examples/seq2seq" python examples/seq2seq/run_eval.py allenai/wm ...@@ -97,3 +87,18 @@ PYTHONPATH="src:examples/seq2seq" python examples/seq2seq/run_eval.py allenai/wm
- [training, etc.](http://www.statmt.org/wmt16/) - [training, etc.](http://www.statmt.org/wmt16/)
- [test set](http://matrix.statmt.org/test_sets/newstest2016.tgz?1504722372) - [test set](http://matrix.statmt.org/test_sets/newstest2016.tgz?1504722372)
### BibTeX entry and citation info
```
@misc{kasai2020deep,
title={Deep Encoder, Shallow Decoder: Reevaluating the Speed-Quality Tradeoff in Machine Translation},
author={Jungo Kasai and Nikolaos Pappas and Hao Peng and James Cross and Noah A. Smith},
year={2020},
eprint={2006.10369},
archivePrefix={arXiv},
primaryClass={cs.CL}
}
```
...@@ -8,7 +8,7 @@ tags: ...@@ -8,7 +8,7 @@ tags:
- translation - translation
- wmt16 - wmt16
- allenai - allenai
license: Apache 2.0 license: apache-2.0
datasets: datasets:
- wmt16 - wmt16
metrics: metrics:
...@@ -29,16 +29,6 @@ All 3 models are available: ...@@ -29,16 +29,6 @@ All 3 models are available:
* [wmt16-en-de-dist-6-1](https://huggingface.co/allenai/wmt16-en-de-dist-6-1) * [wmt16-en-de-dist-6-1](https://huggingface.co/allenai/wmt16-en-de-dist-6-1)
* [wmt16-en-de-12-1](https://huggingface.co/allenai/wmt16-en-de-12-1) * [wmt16-en-de-12-1](https://huggingface.co/allenai/wmt16-en-de-12-1)
```
@misc{kasai2020deep,
title={Deep Encoder, Shallow Decoder: Reevaluating the Speed-Quality Tradeoff in Machine Translation},
author={Jungo Kasai and Nikolaos Pappas and Hao Peng and James Cross and Noah A. Smith},
year={2020},
eprint={2006.10369},
archivePrefix={arXiv},
primaryClass={cs.CL}
}
```
## Intended uses & limitations ## Intended uses & limitations
...@@ -97,3 +87,18 @@ PYTHONPATH="src:examples/seq2seq" python examples/seq2seq/run_eval.py allenai/wm ...@@ -97,3 +87,18 @@ PYTHONPATH="src:examples/seq2seq" python examples/seq2seq/run_eval.py allenai/wm
- [training, etc.](http://www.statmt.org/wmt16/) - [training, etc.](http://www.statmt.org/wmt16/)
- [test set](http://matrix.statmt.org/test_sets/newstest2016.tgz?1504722372) - [test set](http://matrix.statmt.org/test_sets/newstest2016.tgz?1504722372)
### BibTeX entry and citation info
```
@misc{kasai2020deep,
title={Deep Encoder, Shallow Decoder: Reevaluating the Speed-Quality Tradeoff in Machine Translation},
author={Jungo Kasai and Nikolaos Pappas and Hao Peng and James Cross and Noah A. Smith},
year={2020},
eprint={2006.10369},
archivePrefix={arXiv},
primaryClass={cs.CL}
}
```
...@@ -9,7 +9,7 @@ tags: ...@@ -9,7 +9,7 @@ tags:
- translation - translation
- wmt19 - wmt19
- allenai - allenai
license: Apache 2.0 license: apache-2.0
datasets: datasets:
- wmt19 - wmt19
metrics: metrics:
...@@ -29,16 +29,6 @@ For more details, please, see [Deep Encoder, Shallow Decoder: Reevaluating the S ...@@ -29,16 +29,6 @@ For more details, please, see [Deep Encoder, Shallow Decoder: Reevaluating the S
* [wmt19-de-en-6-6-big](https://huggingface.co/allenai/wmt19-de-en-6-6-big) * [wmt19-de-en-6-6-big](https://huggingface.co/allenai/wmt19-de-en-6-6-big)
* [wmt19-de-en-6-6-base](https://huggingface.co/allenai/wmt19-de-en-6-6-base) * [wmt19-de-en-6-6-base](https://huggingface.co/allenai/wmt19-de-en-6-6-base)
```
@misc{kasai2020deep,
title={Deep Encoder, Shallow Decoder: Reevaluating the Speed-Quality Tradeoff in Machine Translation},
author={Jungo Kasai and Nikolaos Pappas and Hao Peng and James Cross and Noah A. Smith},
year={2020},
eprint={2006.10369},
archivePrefix={arXiv},
primaryClass={cs.CL}
}
```
## Intended uses & limitations ## Intended uses & limitations
...@@ -95,3 +85,18 @@ PYTHONPATH="src:examples/seq2seq" python examples/seq2seq/run_eval.py allenai/wm ...@@ -95,3 +85,18 @@ PYTHONPATH="src:examples/seq2seq" python examples/seq2seq/run_eval.py allenai/wm
- [training, etc.](http://www.statmt.org/wmt19/) - [training, etc.](http://www.statmt.org/wmt19/)
- [test set](http://matrix.statmt.org/test_sets/newstest2019.tgz?1556572561) - [test set](http://matrix.statmt.org/test_sets/newstest2019.tgz?1556572561)
### BibTeX entry and citation info
```
@misc{kasai2020deep,
title={Deep Encoder, Shallow Decoder: Reevaluating the Speed-Quality Tradeoff in Machine Translation},
author={Jungo Kasai and Nikolaos Pappas and Hao Peng and James Cross and Noah A. Smith},
year={2020},
eprint={2006.10369},
archivePrefix={arXiv},
primaryClass={cs.CL}
}
```
...@@ -9,7 +9,7 @@ tags: ...@@ -9,7 +9,7 @@ tags:
- translation - translation
- wmt19 - wmt19
- allenai - allenai
license: Apache 2.0 license: apache-2.0
datasets: datasets:
- wmt19 - wmt19
metrics: metrics:
...@@ -29,16 +29,6 @@ For more details, please, see [Deep Encoder, Shallow Decoder: Reevaluating the S ...@@ -29,16 +29,6 @@ For more details, please, see [Deep Encoder, Shallow Decoder: Reevaluating the S
* [wmt19-de-en-6-6-big](https://huggingface.co/allenai/wmt19-de-en-6-6-big) * [wmt19-de-en-6-6-big](https://huggingface.co/allenai/wmt19-de-en-6-6-big)
* [wmt19-de-en-6-6-base](https://huggingface.co/allenai/wmt19-de-en-6-6-base) * [wmt19-de-en-6-6-base](https://huggingface.co/allenai/wmt19-de-en-6-6-base)
```
@misc{kasai2020deep,
title={Deep Encoder, Shallow Decoder: Reevaluating the Speed-Quality Tradeoff in Machine Translation},
author={Jungo Kasai and Nikolaos Pappas and Hao Peng and James Cross and Noah A. Smith},
year={2020},
eprint={2006.10369},
archivePrefix={arXiv},
primaryClass={cs.CL}
}
```
## Intended uses & limitations ## Intended uses & limitations
...@@ -95,3 +85,18 @@ PYTHONPATH="src:examples/seq2seq" python examples/seq2seq/run_eval.py allenai/wm ...@@ -95,3 +85,18 @@ PYTHONPATH="src:examples/seq2seq" python examples/seq2seq/run_eval.py allenai/wm
- [training, etc.](http://www.statmt.org/wmt19/) - [training, etc.](http://www.statmt.org/wmt19/)
- [test set](http://matrix.statmt.org/test_sets/newstest2019.tgz?1556572561) - [test set](http://matrix.statmt.org/test_sets/newstest2019.tgz?1556572561)
### BibTeX entry and citation info
```
@misc{kasai2020deep,
title={Deep Encoder, Shallow Decoder: Reevaluating the Speed-Quality Tradeoff in Machine Translation},
author={Jungo Kasai and Nikolaos Pappas and Hao Peng and James Cross and Noah A. Smith},
year={2020},
eprint={2006.10369},
archivePrefix={arXiv},
primaryClass={cs.CL}
}
```
...@@ -8,7 +8,7 @@ tags: ...@@ -8,7 +8,7 @@ tags:
- translation - translation
- wmt19 - wmt19
- facebook - facebook
license: Apache 2.0 license: apache-2.0
datasets: datasets:
- wmt19 - wmt19
metrics: metrics:
...@@ -92,6 +92,19 @@ note: fairseq reports using a beam of 50, so you should get a slightly higher sc ...@@ -92,6 +92,19 @@ note: fairseq reports using a beam of 50, so you should get a slightly higher sc
- [training, etc.](http://www.statmt.org/wmt19/) - [training, etc.](http://www.statmt.org/wmt19/)
- [test set](http://matrix.statmt.org/test_sets/newstest2019.tgz?1556572561) - [test set](http://matrix.statmt.org/test_sets/newstest2019.tgz?1556572561)
### BibTeX entry and citation info
```bibtex
@inproceedings{...,
year={2020},
title={Facebook FAIR's WMT19 News Translation Task Submission},
author={Ng, Nathan and Yee, Kyra and Baevski, Alexei and Ott, Myle and Auli, Michael and Edunov, Sergey},
booktitle={Proc. of WMT},
}
```
## TODO ## TODO
- port model ensemble (fairseq uses 4 model checkpoints) - port model ensemble (fairseq uses 4 model checkpoints)
......
...@@ -8,7 +8,7 @@ tags: ...@@ -8,7 +8,7 @@ tags:
- translation - translation
- wmt19 - wmt19
- facebook - facebook
license: Apache 2.0 license: apache-2.0
datasets: datasets:
- wmt19 - wmt19
metrics: metrics:
...@@ -92,6 +92,19 @@ note: fairseq reports using a beam of 50, so you should get a slightly higher sc ...@@ -92,6 +92,19 @@ note: fairseq reports using a beam of 50, so you should get a slightly higher sc
- [training, etc.](http://www.statmt.org/wmt19/) - [training, etc.](http://www.statmt.org/wmt19/)
- [test set](http://matrix.statmt.org/test_sets/newstest2019.tgz?1556572561) - [test set](http://matrix.statmt.org/test_sets/newstest2019.tgz?1556572561)
### BibTeX entry and citation info
```bibtex
@inproceedings{...,
year={2020},
title={Facebook FAIR's WMT19 News Translation Task Submission},
author={Ng, Nathan and Yee, Kyra and Baevski, Alexei and Ott, Myle and Auli, Michael and Edunov, Sergey},
booktitle={Proc. of WMT},
}
```
## TODO ## TODO
- port model ensemble (fairseq uses 4 model checkpoints) - port model ensemble (fairseq uses 4 model checkpoints)
......
...@@ -8,7 +8,7 @@ tags: ...@@ -8,7 +8,7 @@ tags:
- translation - translation
- wmt19 - wmt19
- facebook - facebook
license: Apache 2.0 license: apache-2.0
datasets: datasets:
- wmt19 - wmt19
metrics: metrics:
...@@ -92,6 +92,19 @@ note: fairseq reports using a beam of 50, so you should get a slightly higher sc ...@@ -92,6 +92,19 @@ note: fairseq reports using a beam of 50, so you should get a slightly higher sc
- [training, etc.](http://www.statmt.org/wmt19/) - [training, etc.](http://www.statmt.org/wmt19/)
- [test set](http://matrix.statmt.org/test_sets/newstest2019.tgz?1556572561) - [test set](http://matrix.statmt.org/test_sets/newstest2019.tgz?1556572561)
### BibTeX entry and citation info
```bibtex
@inproceedings{...,
year={2020},
title={Facebook FAIR's WMT19 News Translation Task Submission},
author={Ng, Nathan and Yee, Kyra and Baevski, Alexei and Ott, Myle and Auli, Michael and Edunov, Sergey},
booktitle={Proc. of WMT},
}
```
## TODO ## TODO
- port model ensemble (fairseq uses 4 model checkpoints) - port model ensemble (fairseq uses 4 model checkpoints)
......
...@@ -8,7 +8,7 @@ tags: ...@@ -8,7 +8,7 @@ tags:
- translation - translation
- wmt19 - wmt19
- facebook - facebook
license: Apache 2.0 license: apache-2.0
datasets: datasets:
- wmt19 - wmt19
metrics: metrics:
...@@ -92,6 +92,19 @@ note: fairseq reports using a beam of 50, so you should get a slightly higher sc ...@@ -92,6 +92,19 @@ note: fairseq reports using a beam of 50, so you should get a slightly higher sc
- [training, etc.](http://www.statmt.org/wmt19/) - [training, etc.](http://www.statmt.org/wmt19/)
- [test set](http://matrix.statmt.org/test_sets/newstest2019.tgz?1556572561) - [test set](http://matrix.statmt.org/test_sets/newstest2019.tgz?1556572561)
### BibTeX entry and citation info
```bibtex
@inproceedings{...,
year={2020},
title={Facebook FAIR's WMT19 News Translation Task Submission},
author={Ng, Nathan and Yee, Kyra and Baevski, Alexei and Ott, Myle and Auli, Michael and Edunov, Sergey},
booktitle={Proc. of WMT},
}
```
## TODO ## TODO
- port model ensemble (fairseq uses 4 model checkpoints) - port model ensemble (fairseq uses 4 model checkpoints)
......
...@@ -33,7 +33,7 @@ tags: ...@@ -33,7 +33,7 @@ tags:
- translation - translation
- wmt16 - wmt16
- allenai - allenai
license: Apache 2.0 license: apache-2.0
datasets: datasets:
- wmt16 - wmt16
metrics: metrics:
...@@ -54,16 +54,6 @@ All 3 models are available: ...@@ -54,16 +54,6 @@ All 3 models are available:
* [wmt16-en-de-dist-6-1](https://huggingface.co/allenai/wmt16-en-de-dist-6-1) * [wmt16-en-de-dist-6-1](https://huggingface.co/allenai/wmt16-en-de-dist-6-1)
* [wmt16-en-de-12-1](https://huggingface.co/allenai/wmt16-en-de-12-1) * [wmt16-en-de-12-1](https://huggingface.co/allenai/wmt16-en-de-12-1)
```
@misc{{kasai2020deep,
title={{Deep Encoder, Shallow Decoder: Reevaluating the Speed-Quality Tradeoff in Machine Translation}},
author={{Jungo Kasai and Nikolaos Pappas and Hao Peng and James Cross and Noah A. Smith}},
year={{2020}},
eprint={{2006.10369}},
archivePrefix={{arXiv}},
primaryClass={{cs.CL}}
}}
```
## Intended uses & limitations ## Intended uses & limitations
...@@ -122,6 +112,21 @@ PYTHONPATH="src:examples/seq2seq" python examples/seq2seq/run_eval.py allenai/{m ...@@ -122,6 +112,21 @@ PYTHONPATH="src:examples/seq2seq" python examples/seq2seq/run_eval.py allenai/{m
- [training, etc.](http://www.statmt.org/wmt16/) - [training, etc.](http://www.statmt.org/wmt16/)
- [test set](http://matrix.statmt.org/test_sets/newstest2016.tgz?1504722372) - [test set](http://matrix.statmt.org/test_sets/newstest2016.tgz?1504722372)
### BibTeX entry and citation info
```
@misc{{kasai2020deep,
title={{Deep Encoder, Shallow Decoder: Reevaluating the Speed-Quality Tradeoff in Machine Translation}},
author={{Jungo Kasai and Nikolaos Pappas and Hao Peng and James Cross and Noah A. Smith}},
year={{2020}},
eprint={{2006.10369}},
archivePrefix={{arXiv}},
primaryClass={{cs.CL}}
}}
```
""" """
model_card_dir.mkdir(parents=True, exist_ok=True) model_card_dir.mkdir(parents=True, exist_ok=True)
path = os.path.join(model_card_dir, "README.md") path = os.path.join(model_card_dir, "README.md")
......
...@@ -33,7 +33,7 @@ tags: ...@@ -33,7 +33,7 @@ tags:
- translation - translation
- wmt19 - wmt19
- allenai - allenai
license: Apache 2.0 license: apache-2.0
datasets: datasets:
- wmt19 - wmt19
metrics: metrics:
...@@ -53,16 +53,6 @@ For more details, please, see [Deep Encoder, Shallow Decoder: Reevaluating the S ...@@ -53,16 +53,6 @@ For more details, please, see [Deep Encoder, Shallow Decoder: Reevaluating the S
* [wmt19-de-en-6-6-big](https://huggingface.co/allenai/wmt19-de-en-6-6-big) * [wmt19-de-en-6-6-big](https://huggingface.co/allenai/wmt19-de-en-6-6-big)
* [wmt19-de-en-6-6-base](https://huggingface.co/allenai/wmt19-de-en-6-6-base) * [wmt19-de-en-6-6-base](https://huggingface.co/allenai/wmt19-de-en-6-6-base)
```
@misc{{kasai2020deep,
title={{Deep Encoder, Shallow Decoder: Reevaluating the Speed-Quality Tradeoff in Machine Translation}},
author={{Jungo Kasai and Nikolaos Pappas and Hao Peng and James Cross and Noah A. Smith}},
year={{2020}},
eprint={{2006.10369}},
archivePrefix={{arXiv}},
primaryClass={{cs.CL}}
}}
```
## Intended uses & limitations ## Intended uses & limitations
...@@ -119,6 +109,21 @@ PYTHONPATH="src:examples/seq2seq" python examples/seq2seq/run_eval.py allenai/{m ...@@ -119,6 +109,21 @@ PYTHONPATH="src:examples/seq2seq" python examples/seq2seq/run_eval.py allenai/{m
- [training, etc.](http://www.statmt.org/wmt19/) - [training, etc.](http://www.statmt.org/wmt19/)
- [test set](http://matrix.statmt.org/test_sets/newstest2019.tgz?1556572561) - [test set](http://matrix.statmt.org/test_sets/newstest2019.tgz?1556572561)
### BibTeX entry and citation info
```
@misc{{kasai2020deep,
title={{Deep Encoder, Shallow Decoder: Reevaluating the Speed-Quality Tradeoff in Machine Translation}},
author={{Jungo Kasai and Nikolaos Pappas and Hao Peng and James Cross and Noah A. Smith}},
year={{2020}},
eprint={{2006.10369}},
archivePrefix={{arXiv}},
primaryClass={{cs.CL}}
}}
```
""" """
model_card_dir.mkdir(parents=True, exist_ok=True) model_card_dir.mkdir(parents=True, exist_ok=True)
path = os.path.join(model_card_dir, "README.md") path = os.path.join(model_card_dir, "README.md")
......
...@@ -34,7 +34,7 @@ tags: ...@@ -34,7 +34,7 @@ tags:
- translation - translation
- wmt19 - wmt19
- facebook - facebook
license: Apache 2.0 license: apache-2.0
datasets: datasets:
- wmt19 - wmt19
metrics: metrics:
...@@ -118,6 +118,19 @@ note: fairseq reports using a beam of 50, so you should get a slightly higher sc ...@@ -118,6 +118,19 @@ note: fairseq reports using a beam of 50, so you should get a slightly higher sc
- [training, etc.](http://www.statmt.org/wmt19/) - [training, etc.](http://www.statmt.org/wmt19/)
- [test set](http://matrix.statmt.org/test_sets/newstest2019.tgz?1556572561) - [test set](http://matrix.statmt.org/test_sets/newstest2019.tgz?1556572561)
### BibTeX entry and citation info
```bibtex
@inproceedings{{...,
year={{2020}},
title={{Facebook FAIR's WMT19 News Translation Task Submission}},
author={{Ng, Nathan and Yee, Kyra and Baevski, Alexei and Ott, Myle and Auli, Michael and Edunov, Sergey}},
booktitle={{Proc. of WMT}},
}}
```
## TODO ## TODO
- port model ensemble (fairseq uses 4 model checkpoints) - port model ensemble (fairseq uses 4 model checkpoints)
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment