# Wikitext ### Paper Pointer Sentinel Mixture Models https://arxiv.org/pdf/1609.07843.pdf The WikiText language modeling dataset is a collection of over 100 million tokens extracted from the set of verified Good and Featured articles on Wikipedia. NOTE: This `Task` is based on WikiText-2. Homepage: https://www.salesforce.com/products/einstein/ai-research/the-wikitext-dependency-language-modeling-dataset/ ### Citation ``` @misc{merity2016pointer, title={Pointer Sentinel Mixture Models}, author={Stephen Merity and Caiming Xiong and James Bradbury and Richard Socher}, year={2016}, eprint={1609.07843}, archivePrefix={arXiv}, primaryClass={cs.CL} } ``` ### Subtasks ### Checklist - [x] Is in Eval-harness v1.0 ? - [x] Has been checked for regression from v1.0? - [ ] Has been checked for equivalence with original paper methodology? - [ ] "Main" checked variant clearly denoted?