LLM for multilingual tasks, creative writing, math, and multimodal applications
Top 29.4% on sourcepulse
The Skywork project provides a series of 13B parameter large language models (LLMs) trained on 3.2TB of multilingual and code data, aiming to offer strong performance across general tasks, creative writing, and mathematical reasoning. It targets researchers and developers seeking high-quality, open-source bilingual (Chinese/English) models with commercial use potential.
How It Works
Skywork models are built on a thinner, deeper architecture (52 layers) compared to Llama-2-13B, with a larger vocabulary size (65,536) achieved via BPE tokenization. Training involves a two-stage process: initial pre-training on general corpora, followed by a second stage incorporating STEM data to boost reasoning and mathematical abilities. The project also releases Skypile-150B, a 600GB Chinese dataset, and offers quantized versions for consumer GPU deployment.
Quick Start & Requirements
pip install -r requirements.txt
Highlighted Details
Maintenance & Community
The project is developed by the Kunlun Group · Skywork team. Integration with Huawei's MindFormers suite on Ascend hardware is available.
Licensing & Compatibility
The models are available under the "Skywork Community License" and support commercial use, provided terms are followed. Usage is restricted from activities threatening national/social security or unlawful actions.
Limitations & Caveats
The SkyPile-150B dataset, while filtered, may still contain sensitive information. The project disclaims responsibility for risks arising from model misuse or unforeseen issues. Some model variants (Chat, MM) are listed as "coming soon" on certain platforms.
4 months ago
1 day