LLM family with foundation, chat, long context, quantized, RAG, and agent models
Top 45.2% on sourcepulse
Orion-14B is a family of multilingual large language models developed by OrionStarAI, offering a 14B parameter base model and specialized variants for chat, long context, RAG, and agent functionalities. It targets developers and researchers seeking high-performing, multilingual LLMs with robust fine-tuning options and efficient inference capabilities.
How It Works
The Orion-14B models are trained from scratch on a 2.5T token multilingual corpus, emphasizing strong performance in Chinese, English, Japanese, and Korean. The architecture supports up to 320k token context length in its LongChat variant. Quantized versions (Int4) achieve a 70% size reduction and 30% speed increase with minimal performance degradation.
Quick Start & Requirements
transformers
library (Python code examples provided). Quantized versions can be used with llama.cpp
.Highlighted Details
Maintenance & Community
ai@orionstar.com
.Licensing & Compatibility
Limitations & Caveats
The project disclaims responsibility for any issues arising from model use, including data security, public opinion risks, or misuse. Users are urged not to use the models for activities that harm national or social security or violate laws, and to avoid deploying without proper security review.
1 year ago
1 day