Korean GPT-2 model for text generation
Top 58.7% on sourcepulse
KoGPT2 is a Korean language model designed to overcome the limitations of existing Korean NLP models, offering improved performance for text generation, classification, and regression tasks. It is targeted at researchers and developers working with Korean text data.
How It Works
KoGPT2 is a decoder-only transformer model, similar to GPT-2, trained on over 40GB of Korean text data. It utilizes a Character BPE tokenizer with a vocabulary size of 51,200, augmented with common emojis and emoticons to enhance recognition. The model also includes unused tokens for custom task-specific definitions.
Quick Start & Requirements
transformers
library: pip install transformers
from transformers import GPT2LMHeadModel, PreTrainedTokenizerFast
Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
The CC-BY-NC-SA 4.0 license restricts commercial use and requires derivative works to be shared under the same license, which may not be suitable for all commercial applications.
10 months ago
1 day