Wikitext-103 and LAMBADA datasets

The dataset used in the paper is not explicitly mentioned, but it is mentioned that the authors trained a GPT2 transformer language model on the Wikitext-103 and LAMBADA datasets.

Data and Resources

Cite this as

Sheng Shen, Pete Walsh, Kurt Keutzer, Jesse Dodge, Matthew Peters, Iz Beltagy (2024). Dataset: Wikitext-103 and LAMBADA datasets. https://doi.org/10.57702/h8pd5o1u

DOI retrieved: December 16, 2024

Additional Info

Field Value
Created December 16, 2024
Last update December 16, 2024
Defined In https://doi.org/10.48550/arXiv.2203.06211
Author Sheng Shen
More Authors
Pete Walsh
Kurt Keutzer
Jesse Dodge
Matthew Peters
Iz Beltagy
Homepage https://huggingface.co/datasets/wikitext-103