Open-source model family for program synthesis
Top 9.9% on sourcepulse
CodeGen is a family of open-source large language models for program synthesis, offering competitive performance against models like OpenAI Codex. It is designed for researchers and developers working on code generation, autocompletion, and other code-related AI tasks.
How It Works
CodeGen models are trained using the Jaxformer library, a framework for data pre-processing, training, and fine-tuning. The models leverage transformer architectures and are trained on a large corpus of code and natural language, enabling them to generate code snippets from natural language descriptions and perform multi-turn program synthesis.
Quick Start & Requirements
AutoTokenizer
and AutoModelForCausalLM
from the transformers
library.transformers
library. Specific model versions may have varying hardware requirements (e.g., GPU recommended for larger models).Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
6 months ago
Inactive