Qwen3-Coder  by QwenLM

Code LLM for code completion, generation, and assistant use cases

Created 1 year ago
13,458 stars

Top 3.7% on SourcePulse

GitHubView on GitHub
Project Summary

Qwen2.5-Coder is a series of open-source large language models specifically designed for code generation and understanding. It aims to provide powerful, diverse, and practical coding assistance across various programming languages and use cases, targeting developers and researchers in the AI and software engineering fields.

How It Works

The Qwen2.5-Coder series builds upon the Qwen2.5 base models, fine-tuning them for coding tasks. It supports a context length of up to 128K tokens, leveraging techniques like YaRN for efficient handling of long sequences. The models are trained on 92 coding languages and include specialized tokens for fill-in-the-middle (FIM) and repository-level code completion, enabling sophisticated code insertion and context-aware generation across multiple files.

Quick Start & Requirements

  • Installation: pip install -r requirements.txt (requires transformers>4.37.0).
  • Prerequisites: Python >= 3.9.
  • Usage: Models can be loaded and used via the Hugging Face transformers library. Examples for chat, code completion, and fill-in-the-middle are provided.
  • Resources: Various model sizes (0.5B to 32B) are available, including quantized versions (AWQ, GGUF, GPTQ) for reduced resource requirements.
  • Links: Hugging Face Organization, ModelScope Organization, Documentation, Chatbot Demo, Artifacts Demo.

Highlighted Details

  • SOTA Performance: The 32B-Instruct model claims to match GPT-4o's coding capabilities and is recognized as a State-of-the-Art open-source code model.
  • Broad Language Support: Covers 92 programming languages.
  • Long Context: Supports up to 128K token context length.
  • Advanced Features: Includes specialized tokens and methods for fill-in-the-middle and repository-level code completion.

Maintenance & Community

  • Developed by Alibaba Cloud's Qwen team.
  • Active community support via Discord and WeChat groups.

Licensing & Compatibility

  • The specific license is not explicitly stated in the README, but the models are open-sourced by Alibaba Cloud. Compatibility for commercial use should be verified.

Limitations & Caveats

  • The README does not explicitly detail licensing restrictions for commercial use. The performance claims for the 32B model matching GPT-4o are self-reported and should be validated through independent benchmarks.
Health Check
Last Commit

1 month ago

Responsiveness

1 day

Pull Requests (30d)
1
Issues (30d)
18
Star History
1,213 stars in the last 30 days

Explore Similar Projects

Starred by Didier Lopes Didier Lopes(Founder of OpenBB), Chip Huyen Chip Huyen(Author of "AI Engineering", "Designing Machine Learning Systems"), and
3 more.

DeepSeek-Coder-V2 by deepseek-ai

0.3%
6k
Open-source code language model comparable to GPT4-Turbo
Created 1 year ago
Updated 11 months ago
Starred by Chip Huyen Chip Huyen(Author of "AI Engineering", "Designing Machine Learning Systems"), Vincent Weisser Vincent Weisser(Cofounder of Prime Intellect), and
15 more.

codellama by meta-llama

0.0%
16k
Inference code for CodeLlama models
Created 2 years ago
Updated 1 year ago
Feedback? Help us improve.