Code LLM for code completion and generation
Top 1.9% on sourcepulse
DeepSeek Coder is a suite of open-source code language models trained from scratch on 2 trillion tokens, comprising 87% code and 13% natural language. It offers models ranging from 1B to 33B parameters, designed for project-level code completion and infilling with a 16K context window. The models achieve state-of-the-art performance on various coding benchmarks, making them suitable for developers and researchers seeking advanced code generation capabilities.
How It Works
The models are pre-trained on a massive dataset of code and natural language, with a focus on project-level context. This is achieved through a 16K context window and a fill-in-the-blank task, enabling the models to understand and generate code across entire projects. Instruction-tuned variants are also available for conversational coding assistance.
Quick Start & Requirements
pip install -r requirements.txt
Highlighted Details
Maintenance & Community
awesome-deepseek-coder
.Licensing & Compatibility
Limitations & Caveats
eos_token_id
adjustment.1 year ago
1 week