Bilingual language models for Korean/English, compute-efficient vs. SOTA
Top 99.2% on sourcepulse
Kanana is a series of bilingual (Korean/English) language models designed for compute efficiency, offering strong Korean language capabilities and competitive English performance. The models are suitable for researchers and developers working with Korean NLP tasks, providing base, instruct, embedding, and RAG-optimized versions.
How It Works
Kanana models leverage several pre-training techniques for efficiency, including high-quality data filtering, staged pre-training, depth up-scaling, and pruning/distillation. Post-training involves supervised fine-tuning and preference optimization to enhance user interaction. The models are available in various sizes, from 2.1B to 32.5B parameters, with smaller versions released to foster Korean LLM research.
Quick Start & Requirements
transformers>=4.45.0
via pip.torch_dtype=torch.bfloat16
for optimal performance.Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
1 week ago
Inactive