-
Llama: Open and efficient foundation language models
The LLaMA dataset is a large language model dataset used in the paper. -
BERT: Pre-training of deep bidirectional transformers for language understanding
This paper proposes BERT, a pre-trained deep bidirectional transformer for language understanding. -
Training Language Models to Perform Tasks
A dataset for training language models to perform tasks such as question answering and text classification. -
Interpreting Learned Feedback Patterns in Large Language Models
The dataset used in the paper is not explicitly described, but it is mentioned that the authors used a condensed representation of LLM activations obtained from sparse...