Open-source LLM for chat, instruction-following, and general language tasks
Top 53.5% on sourcepulse
YuLan-Chat offers open-source chat-based large language models, primarily targeting Chinese and English language users. Developed by Renmin University of China, it aims to provide helpful, honest, and harmless AI assistants, with recent versions trained from scratch and featuring enhanced Chinese language support and longer context windows.
How It Works
YuLan-Chat models are built through large-scale pre-training on English, Chinese, and multilingual data. They are then fine-tuned using a curriculum learning strategy with high-quality instructions and human preference data. This approach enhances their helpfulness, honesty, and harmlessness. Specific versions have expanded vocabularies and context lengths (up to 4k) to better support Chinese inputs and outputs.
Quick Start & Requirements
conda create -n yulan python=3.10 -y
, conda activate yulan
), install PyTorch and bitsandbytes (versions 1.13 and 0.39.0 recommended), then pip install -r requirements.txt
.Highlighted Details
Maintenance & Community
The project is actively developed by researchers from Renmin University of China. Specific contributors are listed for pre-training and fine-tuning roles. No community links (Discord, Slack) are provided in the README.
Licensing & Compatibility
The project uses the MIT License. However, all data and code are restricted to academic purposes only, which may limit commercial use or integration into closed-source projects.
Limitations & Caveats
While efforts are made to mitigate harmful outputs, the models are probabilistic and may generate biased, discriminatory, or otherwise harmful content. The project disclaims responsibility for consequences arising from the dissemination of such information.
6 months ago
1 day