Seed-Coder  by ByteDance-Seed

Code LLM for code generation, completion, and reasoning tasks

Created 5 months ago
549 stars

Top 58.2% on SourcePulse

GitHubView on GitHub
Project Summary

Seed-Coder is a family of 8B parameter code LLMs (base, instruct, reasoning) from ByteDance Seed, designed to enhance coding capabilities by using LLMs to curate their own training data, minimizing human effort. It targets developers and researchers seeking powerful, lightweight, open-source code intelligence solutions.

How It Works

Seed-Coder employs a "model-centric" data pipeline, leveraging LLMs for data filtering and curation from sources like GitHub, commits, and web data. This approach aims to reduce manual effort in pretraining data construction while achieving state-of-the-art performance for its size.

Quick Start & Requirements

  • Install/Run: Deployable via Hugging Face transformers or vLLM.
  • Prerequisites: torch (bfloat16 recommended), transformers, vLLM for advanced deployment.
  • Resources: 8B parameter models; vLLM supports multi-GPU and tensor parallelism for long contexts (up to 32K tokens).
  • Links: Homepage, Hugging Face

Highlighted Details

  • State-of-the-art performance among open-source models at the 8B scale across various coding tasks.
  • Models include Base, Instruct (for user intent alignment), and Reasoning (RL-trained).
  • Supports long context windows up to 64K tokens for the Reasoning model.
  • Fully compatible with vLLM for efficient inference and distributed serving.

Maintenance & Community

  • Developed by the ByteDance Seed Team, founded in 2023.
  • Models are publicly available on Hugging Face.

Licensing & Compatibility

  • MIT License. Permissive for commercial use and closed-source linking.

Limitations & Caveats

  • Evaluation results for BigCodeBench were updated due to an inconsistent setting; users should refer to the latest reported benchmarks.
Health Check
Last Commit

3 months ago

Responsiveness

1 day

Pull Requests (30d)
0
Issues (30d)
1
Star History
7 stars in the last 30 days

Explore Similar Projects

Starred by Shizhe Diao Shizhe Diao(Author of LMFlow; Research Scientist at NVIDIA), Yineng Zhang Yineng Zhang(Inference Lead at SGLang; Research Scientist at Together AI), and
8 more.

EAGLE by SafeAILab

10.6%
2k
Speculative decoding research paper for faster LLM inference
Created 1 year ago
Updated 1 week ago
Starred by Chip Huyen Chip Huyen(Author of "AI Engineering", "Designing Machine Learning Systems"), Vincent Weisser Vincent Weisser(Cofounder of Prime Intellect), and
15 more.

codellama by meta-llama

0.0%
16k
Inference code for CodeLlama models
Created 2 years ago
Updated 1 year ago
Feedback? Help us improve.