Open-source bilingual LLMs trained from scratch
Top 6.8% on sourcepulse
The Yi series models are open-source large language models developed by 01.AI, trained from scratch on a 3T multilingual corpus. They are designed for strong language understanding, reasoning, and comprehension, targeting researchers, developers, and businesses seeking high-performing bilingual LLMs.
How It Works
Yi models are built upon the Transformer architecture, similar to Llama, but are not derivatives. This foundation provides stability and compatibility within the AI ecosystem. The key differentiators are 01.AI's proprietary training datasets, efficient pipelines, and robust infrastructure, which contribute to Yi models' competitive performance against leading LLMs.
Quick Start & Requirements
pip
(Python 3.10+), Docker, conda-lock
, and llama.cpp
for quantized models.git-lfs
.Highlighted Details
Maintenance & Community
The project is actively maintained by 01.AI. Community engagement is encouraged via Discord and WeChat. Recent updates include the Yi-1.5 series and the Yi Cookbook.
Licensing & Compatibility
The Yi series models are distributed under the Apache 2.0 license, permitting personal, academic, and commercial use. Derivative works require attribution.
Limitations & Caveats
The chat models' increased response diversity, while beneficial for creative tasks, may lead to higher instances of hallucination or non-determinism. Adjusting generation parameters like temperature is recommended for more coherent outputs.
8 months ago
1 day