LiveCodeBench

LiveCodeBench is a benchmark for evaluating the performance of Large Language Models (LLMs) in code editing tasks, including debugging, translating, polishing, and requirement switching.

Data and Resources

Cite this as

N. Jain, K. Han, A. Gu, W.-D. Li, F. Yan, T. Zhang, S. Wang, A. Solar-Lezama, K. Sen, I. Stoica (2024). Dataset: LiveCodeBench. https://doi.org/10.57702/xzte8dui

DOI retrieved: December 16, 2024

Additional Info

Field Value
Created December 16, 2024
Last update December 16, 2024
Defined In https://doi.org/10.48550/arXiv.2406.06887
Citation
  • https://doi.org/10.48550/arXiv.2404.03543
Author N. Jain
More Authors
K. Han
A. Gu
W.-D. Li
F. Yan
T. Zhang
S. Wang
A. Solar-Lezama
K. Sen
I. Stoica
Homepage https://arxiv.org/abs/2404.12345