Dataset for measuring coding challenge competence (NeurIPS 2021)
Top 64.9% on sourcepulse
This repository provides the APPS dataset and code for measuring coding challenge competence, targeting researchers and developers in AI and natural language processing. It enables the evaluation of large language models on their ability to solve programming problems, offering a standardized benchmark for progress in automated programming.
How It Works
The APPS dataset consists of programming problems sourced from competitive programming platforms. The associated code allows for fine-tuning and evaluating transformer-based language models on their ability to generate correct code solutions. This approach standardizes the assessment of coding competence, moving beyond simple code completion tasks.
Quick Start & Requirements
train/README
and eval/README
respectively.datasets
library.Highlighted Details
Maintenance & Community
The project is associated with authors from prominent institutions and presented at NeurIPS 2021. Further community engagement details are not explicitly provided in the README.
Licensing & Compatibility
The repository's license is not explicitly stated in the README. Compatibility for commercial use or closed-source linking would require clarification of the licensing terms.
Limitations & Caveats
The README does not specify the exact license, which may impact commercial adoption. Detailed setup and execution instructions are referenced in separate README files within subdirectories.
1 year ago
1 day