Discover and explore top open-source AI tools and projects—updated daily.
kakaobrainKorean generative pre-trained transformer for classifying, searching, summarizing, or generating Korean texts
Top 36.9% on SourcePulse
KoGPT is a large language model developed by KakaoBrain, specifically designed for Korean text generation. It offers a 6B parameter model, KoGPT6B-ryan1.5b, leveraging Rotary Position Embeddings (RoPE) for enhanced performance. This model is suitable for researchers and developers working with Korean NLP tasks like classification, summarization, and generation.
How It Works
KoGPT is a Transformer-based causal language model. It utilizes Rotary Position Embeddings (RoPE) for positional encoding, a design choice that can improve performance in sequence modeling tasks by injecting relative positional information. The model is trained on a large corpus of Korean text, making it highly proficient in understanding and generating Korean language nuances.
Quick Start & Requirements
pip install transformers torch.float16 or 32GB VRAM for float32 versions. CUDA is recommended.python -m kogpttransformers.Highlighted Details
float32 and float16 versions for flexibility in hardware requirements and performance.Maintenance & Community
contact@kakaobrain.com for cooperation.Licensing & Compatibility
Limitations & Caveats
KoGPT is primarily trained on Korean text and may perform poorly on non-Korean inputs or specific Korean dialects not well-represented in the training data. Due to training on raw data, it can generate socially unacceptable or offensive text, and its output is difficult to predict.
1 year ago
Inactive
apoorvumang