General Language Understanding Evaluation (GLUE) benchmark

GLUE is a multi-task benchmark that contains a diverse set of natural language understanding tasks including sentiment analysis, natural language inference, and textual similarity evaluation.

Data and Resources

Cite this as

Alex Wang, Amanpreet Singh, Julian Michael, Felix Hill, Omer Levy, Samuel R. Bowman (2024). Dataset: General Language Understanding Evaluation (GLUE) benchmark. https://doi.org/10.57702/3iiy7g13

DOI retrieved: November 25, 2024

Additional Info

Field Value
Created November 25, 2024
Last update November 25, 2024
Defined In https://doi.org/10.48550/arXiv.2002.08307
Author Alex Wang
More Authors
Amanpreet Singh
Julian Michael
Felix Hill
Omer Levy
Samuel R. Bowman
Homepage https://gluebenchmark.com/