Qwen3-Coder  by QwenLM

Code LLM for code completion, generation, and assistant use cases

created 1 year ago
10,387 stars

Top 4.9% on sourcepulse

GitHubView on GitHub
Project Summary

Qwen2.5-Coder is a series of open-source large language models specifically designed for code generation and understanding. It aims to provide powerful, diverse, and practical coding assistance across various programming languages and use cases, targeting developers and researchers in the AI and software engineering fields.

How It Works

The Qwen2.5-Coder series builds upon the Qwen2.5 base models, fine-tuning them for coding tasks. It supports a context length of up to 128K tokens, leveraging techniques like YaRN for efficient handling of long sequences. The models are trained on 92 coding languages and include specialized tokens for fill-in-the-middle (FIM) and repository-level code completion, enabling sophisticated code insertion and context-aware generation across multiple files.

Quick Start & Requirements

  • Installation: pip install -r requirements.txt (requires transformers>4.37.0).
  • Prerequisites: Python >= 3.9.
  • Usage: Models can be loaded and used via the Hugging Face transformers library. Examples for chat, code completion, and fill-in-the-middle are provided.
  • Resources: Various model sizes (0.5B to 32B) are available, including quantized versions (AWQ, GGUF, GPTQ) for reduced resource requirements.
  • Links: Hugging Face Organization, ModelScope Organization, Documentation, Chatbot Demo, Artifacts Demo.

Highlighted Details

  • SOTA Performance: The 32B-Instruct model claims to match GPT-4o's coding capabilities and is recognized as a State-of-the-Art open-source code model.
  • Broad Language Support: Covers 92 programming languages.
  • Long Context: Supports up to 128K token context length.
  • Advanced Features: Includes specialized tokens and methods for fill-in-the-middle and repository-level code completion.

Maintenance & Community

  • Developed by Alibaba Cloud's Qwen team.
  • Active community support via Discord and WeChat groups.

Licensing & Compatibility

  • The specific license is not explicitly stated in the README, but the models are open-sourced by Alibaba Cloud. Compatibility for commercial use should be verified.

Limitations & Caveats

  • The README does not explicitly detail licensing restrictions for commercial use. The performance claims for the 32B model matching GPT-4o are self-reported and should be validated through independent benchmarks.
Health Check
Last commit

2 days ago

Responsiveness

1 day

Pull Requests (30d)
11
Issues (30d)
47
Star History
6,005 stars in the last 90 days

Explore Similar Projects

Starred by Chip Huyen Chip Huyen(Author of AI Engineering, Designing Machine Learning Systems), Travis Fischer Travis Fischer(Founder of Agentic), and
6 more.

codellama by meta-llama

0.1%
16k
Inference code for CodeLlama models
created 1 year ago
updated 11 months ago
Feedback? Help us improve.