HyperAIHyperAI

Command Palette

Search for a command to run...

5 months ago

Measuring Coding Challenge Competence With APPS

Dan Hendrycks; Steven Basart; Saurav Kadavath; Mantas Mazeika; Akul Arora; Ethan Guo; Collin Burns; Samir Puranik; Horace He; Dawn Song; Jacob Steinhardt

Measuring Coding Challenge Competence With APPS

Abstract

While programming is one of the most broadly applicable skills in modern society, modern machine learning models still cannot code solutions to basic problems. Despite its importance, there has been surprisingly little work on evaluating code generation, and it can be difficult to accurately assess code generation performance rigorously. To meet this challenge, we introduce APPS, a benchmark for code generation. Unlike prior work in more restricted settings, our benchmark measures the ability of models to take an arbitrary natural language specification and generate satisfactory Python code. Similar to how companies assess candidate software developers, we then evaluate models by checking their generated code on test cases. Our benchmark includes 10,000 problems, which range from having simple one-line solutions to being substantial algorithmic challenges. We fine-tune large language models on both GitHub and our training set, and we find that the prevalence of syntax errors is decreasing exponentially as models improve. Recent models such as GPT-Neo can pass approximately 20% of the test cases of introductory problems, so we find that machine learning models are now beginning to learn how to code. As the social significance of automatic code generation increases over the coming years, our benchmark can provide an important measure for tracking advancements.

Code Repositories

ncoop57/gpt-code-clippy
jax
Mentioned in GitHub
codedotal/gpt-code-clippy
jax
Mentioned in GitHub
hendrycks/apps
Official
pytorch
Mentioned in GitHub

Benchmarks

BenchmarkMethodologyMetrics
code-generation-on-appsGPT-Neo 2.7B
Competition Pass@1: 0.00%
Competition Pass@1000: 11.40%
Competition Pass@5: 0.00%
Competition Pass@any: 11.40%
Interview Pass@1: 0.57%
Interview Pass@1000: 9.83%
Interview Pass@5: 0.80%
Interview Pass@any: 9.83%
Introductory Pass@1: 3.90%
Introductory Pass@1000: 27.90%
Introductory Pass@5: 5.50%
Introductory Pass@any: 27.90%

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp