-
APPS: A Dataset for Code Generation Evaluation
The APPS dataset is a collection of programming problems used to evaluate the performance of code generation models. -
Evaluating large language models trained on code
The paper presents the results of the OpenAI Codex evaluation on generating Python code. -
Execution-based Evaluation for NL2Bash
A set of 50 prompts to evaluate execution-based evaluation for NL2Bash task -
CodeUltraFeedback
CodeUltraFeedback is a preference dataset of 10,000 complex instructions to tune and align LLMs to coding preferences through AI feedback. -
Evol-Instruct-Code-80k
Evol-Instruct-Code-80k is a dataset for evaluating the performance of code generation models.