GPT-2 XL

The GPT-2 dataset is a large language model, specifically the GPT-2 XL model, trained on the Common Crawl dataset.

Data and Resources

Cite this as

Steven Bill, Alec Radford, Iishita Gupta, Jason Weston, Stephen Roller, Jamie Ray, Mike Thorne, Yiming Yang, Zhilin Yang, Wen-tau Hsieh, Kaito Kawaguchi, Keunwoo Kim, Jasa Gupta, Anirudh Anantharaman, Amanpreet Singh, Edwin Chen, Suriya Sikdar, Omar Bilello, Yiming Yang, Zhilin Yang (2024). Dataset: GPT-2 XL. https://doi.org/10.57702/borltms7

DOI retrieved: December 17, 2024

Additional Info

Field Value
Created December 17, 2024
Last update December 17, 2024
Defined In https://doi.org/10.48550/arXiv.2401.17975
Author Steven Bill
More Authors
Alec Radford
Iishita Gupta
Jason Weston
Stephen Roller
Jamie Ray
Mike Thorne
Yiming Yang
Zhilin Yang
Wen-tau Hsieh
Kaito Kawaguchi
Keunwoo Kim
Jasa Gupta
Anirudh Anantharaman
Amanpreet Singh
Edwin Chen
Suriya Sikdar
Omar Bilello
Yiming Yang
Zhilin Yang
Homepage https://huggingface.co/models/cardinal-research/gpt-2-xl