SkyText-Chinese-GPT3  by SkyWorkAIGC

Chinese GPT3 pre-trained language model

created 2 years ago
405 stars

Top 72.8% on sourcepulse

GitHubView on GitHub
Project Summary

SkyText is a large Chinese GPT-3 pre-trained language model developed by Singularity-AI. It addresses the need for advanced Chinese NLP capabilities, offering functionalities beyond basic chat and Q&A, including content generation, translation, and creative writing. The model is suitable for researchers and developers working with Chinese language AI applications.

How It Works

SkyText leverages a novel Chinese encoding method optimized for the language's unique characteristics, differing from English-centric approaches. This innovation, combined with a rigorous 30+ step data cleaning process for its training corpus, aims to enhance the model's comprehension and performance on Chinese text.

Quick Start & Requirements

  • Install via transformers library.
  • Requires transformers>=4.18.0.
  • Example usage involves loading GPT2LMHeadModel and AutoTokenizer from "SkyWork/SkyText" or "SkyWork/SkyTextTiny".
  • GPU is recommended for optimal performance (example uses device=0).
  • Official Hugging Face model pages: SkyText (13B), SkyTextTiny (2.6B).

Highlighted Details

  • Offers a 13 billion parameter model (currently closed-source, with a new 10 billion parameter model upcoming) and a 2.6 billion parameter "Tiny" version.
  • Supports diverse tasks: article continuation, dialogue, Chinese-English translation, content style generation, reasoning, poetry, recipe generation, and more.
  • Emphasizes a data cleaning pipeline with over 30 steps for improved model quality.

Maintenance & Community

  • Community access via WeChat group (QR code in README).
  • Encourages starring the repository.

Licensing & Compatibility

  • Licensed under the MIT License.
  • Permissive for commercial use and integration into closed-source projects.

Limitations & Caveats

The 13 billion parameter model is currently closed-source, with users directed to await the release of a new 10 billion parameter model.

Health Check
Last commit

2 years ago

Responsiveness

Inactive

Pull Requests (30d)
0
Issues (30d)
0
Star History
0 stars in the last 90 days

Explore Similar Projects

Feedback? Help us improve.