GPT-2 for multiple languages, including pretrained models
Top 25.4% on sourcepulse
This repository provides a GPT-2 implementation optimized for multilingual support, specifically featuring a 1.5 billion parameter Chinese pretrained model. It is designed for researchers and developers working with large-scale language models for Chinese text generation and analysis.
How It Works
The project adapts Grover's training scripts for GPT-2, incorporating a ported BERT tokenizer compatible with multilingual corpora. It leverages Cloud TPUs for efficient training, enabling the creation of large, high-performance models like the 1.5B parameter Chinese version.
Quick Start & Requirements
Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
The project is designated for academic research and does not offer conclusive remarks. The specific license for commercial use or closed-source linking is not detailed.
2 years ago
Inactive