Large language model for research, fine-tuning, and deployment
Top 50.3% on sourcepulse
Yuan 2.0 is a family of open-source large language models developed by Inspur Information, offering models in 2B, 51B, and 102B parameter sizes. It aims to provide enhanced capabilities in semantics, mathematics, reasoning, and code generation, targeting researchers and developers for further development and application building.
How It Works
Yuan 2.0 builds upon its predecessor by utilizing a more diverse and high-quality dataset for pre-training and instruction fine-tuning. This approach aims to improve the model's understanding across various domains, including language, math, reasoning, and code. The project provides scripts for pre-training, fine-tuning, and inference services, enabling users to adapt and deploy the models for specific applications.
Quick Start & Requirements
docker pull yuanmodel/yuan2.0:V1-base
docker run --gpus all --privileged --ulimit stack=68719476736 --shm-size=1000G -itd -v /path/to/yuan_2.0:/workspace/yuan_2.0 -v /path/to/dataset:/workspace/dataset -v /path/to/checkpoints:/workspace/checkpoints --name your_name yuanmodel/yuan2.0:V1-base
docker exec -it your_name bash
Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
1 year ago
1 day