Large language model for NLP tasks
Top 55.8% on sourcepulse
Yuan 1.0 is a large-scale pre-trained language model designed for zero-shot and few-shot learning tasks in Natural Language Processing. It targets researchers and developers seeking high-performance language models, offering state-of-the-art results on various NLP tasks with a focus on Chinese text.
How It Works
Yuan 1.0 is a 246 billion parameter singleton language model trained on a custom 5TB Chinese text corpus, the largest of its kind. The model architecture design incorporates insights from large-scale distributed training. It also utilizes a calibration and label expansion method to enhance zero-shot and few-shot performance, aiming to produce human-indistinguishable text.
Quick Start & Requirements
pip install requests hashlib json
Highlighted Details
Maintenance & Community
The project is associated with Inspur. Further community or maintenance details are not explicitly provided in the README.
Licensing & Compatibility
The README does not specify a license. API access is controlled via application and key generation.
Limitations & Caveats
Access to the model's full capabilities is gated behind an application process. The README does not detail specific hardware requirements for running inference locally, implying API usage is the primary access method.
8 months ago
1 week