Research paper for self-alignment in code generation
Top 88.0% on sourcepulse
SelfCodeAlign offers a fully open and transparent pipeline for enhancing code generation models without human annotations or proprietary distilled data. This approach enables the creation of self-aligned code models, such as StarCoder2-Instruct, achieving state-of-the-art performance on coding tasks. The project targets researchers and developers seeking to build and improve code LLMs with reproducible and accessible methods.
How It Works
The SelfCodeAlign pipeline leverages a self-improvement loop. It uses an existing code model (StarCoder2-15B) to generate instruction-response pairs. These synthetic data are then used to fine-tune the same model, effectively aligning it with desired coding behaviors without external human feedback or reliance on data from larger, closed models. This method promotes transparency and reproducibility in LLM alignment.
Quick Start & Requirements
pip install -e .
(from the cloned repository)bigcode/starcoder2-15b
.bigcode/self-oss-instruct-sc2-exec-filter-50k
.README-SC2INST.md
for detailed instructions.Highlighted Details
Maintenance & Community
The project is associated with the BigCode community, a collaboration focused on responsible development of large language models for code. Key contributors include researchers from institutions like Hugging Face and ServiceNow.
Licensing & Compatibility
The project and its output model (StarCoder2-Instruct) are released under permissive licenses, facilitating commercial use and integration into closed-source projects. Specific license details are available with the model and code.
Limitations & Caveats
The fine-tuning process is computationally intensive, requiring substantial GPU resources. While the pipeline is designed for transparency, replicating the exact results may depend on specific hardware configurations and hyperparameter tuning.
5 months ago
1 day