Korean BERT for language tasks
Top 30.0% on sourcepulse
KoBERT is a Korean BERT model pre-trained on a large Korean corpus, offering improved performance over multilingual BERT for Korean NLP tasks. It is designed for researchers and developers working with Korean language processing, providing a strong foundation for fine-tuning on specific downstream tasks like sentiment analysis and named entity recognition.
How It Works
KoBERT is based on the BERT architecture, featuring 12 layers, 768 hidden units, and 12 attention heads. It uses a SentencePiece tokenizer trained on Korean Wikipedia, resulting in a vocabulary size of 8,002 tokens. The model is trained on approximately 54 million sentences from Korean Wikipedia, aiming for better Korean language understanding and efficiency with fewer parameters than BERT base.
Quick Start & Requirements
pip install git+https://git@github.com/SKTBrain/KoBERT.git@master
Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
The README mentions that the model is returned in eval()
mode by default, requiring explicit switching to train()
mode for fine-tuning.
1 month ago
Inactive