Program synthesis research release (ICLR 2023)
Top 95.5% on sourcepulse
CodeGen2 provides official research releases of large language models (LLMs) for program synthesis, specifically addressing the challenges of training LLMs on both programming and natural languages. It targets researchers and developers working on code generation, autocompletion, and other program synthesis tasks, offering models ranging from 1 billion to 16 billion parameters.
How It Works
CodeGen2 utilizes an auto-regressive sampling approach for program synthesis. The models are trained on a diverse dataset encompassing both natural language and programming languages, enabling them to generate code based on natural language descriptions or to fill in code snippets. This dual-language training is a key differentiator, allowing for more versatile and context-aware code generation.
Quick Start & Requirements
Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
The README does not detail specific limitations, performance benchmarks, or known issues. The significant hardware requirements for larger models may be a barrier to entry for some users.
2 years ago
1 day