-
MNLI, QQP, and SST-2
The dataset used in this paper consists of three tasks: Multi-Genre Natural Language Inference (MNLI), Quora Question Pairs (QQP), and Stanford Sentiment Treebank (SST-2). -
Are Larger Pretrained Language Models Uniformly Better? Comparing Performance...
Larger language models have higher accu- racy on average, but are they better on ev- ery single instance (datapoint)? -
Learning to summarize with human feedback
The paper presents a study on the impact of synthetic data on large language models (LLMs) and proposes a method to steer LLMs towards desirable non-differentiable attributes. -
Reward Model Ensembles
The authors used three datasets: TL;DR, HELPFULNESS, and XSUM/NLI. -
STAMP 4 NLP
STAMP 4 NLP is an instantiable, iterative, and incremental process model for developing natural language processing applications with a focus on quality, business value, and... -
Detecting Hallucinated Content in Conditional Neural Sequence Generation
Neural sequence models can generate highly fluent sentences, but recent studies have also shown that they are also prone to hallucinate additional content not supported by the... -
A general theoretical paradigm to understand learning from human preferences
The paper proposes a novel approach to aligning language models with human preferences, focusing on the use of preference optimization in reward-free RLHF. -
Llama: Open and efficient foundation language models
The LLaMA dataset is a large language model dataset used in the paper. -
Mixtral of Experts
The dataset used in the paper for instruction following task -
Toward an Architecture for Never-ending Language Learning
Toward an Architecture for Never-ending Language Learning. -
MISMATCH: Fine-grained Evaluation of Machine-generated Text
The dataset used in the paper for fine-grained evaluation of machine-generated text with mismatch error types. -
BERT: Pre-training of deep bidirectional transformers for language understanding
This paper proposes BERT, a pre-trained deep bidirectional transformer for language understanding. -
Training Dataset
The training dataset is a collection of the publicly available Arabic corpora listed below: The unshuffled OSCAR corpus (Ortiz Su´arez et al., 2020). The Arabic Wikipedia dump... -
Orca: Progressive Learning from Complex Explanation Traces
The Orca approach involves leveraging explanation tuning to generate detailed responses from a large language model. -
Evol-Instruct: A Pipeline for Automatically Evolving Instruction Datasets
The Evol-Instruct pipeline involves automatically evolving instruction datasets using large language models. -
Various Datasets
The datasets used in the paper are described as follows: WikiMIA, BookMIA, Temporal Wiki, Temporal arXiv, ArXiv-1 month, Multi-Webdata, LAION-MI, Gutenberg.