OpenSeek  by FlagAI-Open

Open-source platform for collaborative LLM development and next-generation model innovation

Created 1 year ago
253 stars

Top 99.3% on SourcePulse

GitHubView on GitHub
Project Summary

OpenSeek is an open-source initiative by BAAI aiming to foster global collaborative innovation in algorithms, data, and systems for next-generation large language models, targeting researchers and developers. It addresses the gap in complete code, computational resources, and data support for academic LLM breakthroughs, with the goal of developing models that surpass DeepSeek and promoting independent technological advancement.

How It Works

The project champions a collaborative ecosystem, inspired by initiatives like Bigscience and OPT, to build an independent open-source algorithmic innovation system. Its core approach involves exploring advanced data construction mechanisms, open-sourcing the entire LLM training pipeline, and developing innovative training and inference code. A key differentiator is the explicit goal to support various AI chips beyond Nvidia, reducing hardware dependency and enhancing model universality.

Quick Start & Requirements

Installation is recommended via Docker (docker pull openseek2025/openseek:flagscale-20250527) or from source by cloning the FlagScale repository and running ./install/install-requirements.sh --env train. Prerequisites include a Python environment and the FlagScale dependencies. Users must also prepare the OpenSeek-Pretrain-100B dataset. Detailed setup and configuration are outlined in the README and linked FlagScale documentation.

Highlighted Details

  • Models: Features OpenSeek-Small V1 (1.4B parameters, 720B tokens) and a baseline version (1.4B parameters, 100B tokens).
  • Datasets: Includes CCI4.0-M2-V1, a large-scale bilingual pre-training dataset (5.2TB Chinese, 22TB English, CoT data), and OpenSeek-Pretrain-100B (a 100B token subset).
  • Hardware Agnosticism: Aims to support diverse AI chips beyond Nvidia, promoting broader accessibility and adaptability.
  • Community-Driven: Leverages community contributions for algorithmic, data, and system improvements.

Maintenance & Community

Initiated by the Beijing Academy of Artificial Intelligence (BAAI) and supported by the FlagScale team. The project actively shares news on data and model releases (e.g., CCI4.0-M2-V1 and OpenSeek-Small V1 on 05/06/2025) and hosts online meetups. A Discord channel is available for community interaction.

Licensing & Compatibility

The project is licensed under the Apache 2.0 license, which is permissive for commercial use and integration into closed-source projects.

Limitations & Caveats

The README does not explicitly detail limitations. However, the project's ambitious scope—reproducing and surpassing DeepSeek, developing a full training pipeline, and supporting diverse hardware—suggests a complex setup and significant resource requirements. Potential licensing concerns for specific data components are hinted at with the release of CCI4.0-M2-Extra data.

Health Check
Last Commit

2 months ago

Responsiveness

1 week

Pull Requests (30d)
1
Issues (30d)
10
Star History
6 stars in the last 30 days

Explore Similar Projects

Feedback? Help us improve.