Russian language models and instruction tuning
Top 66.0% on sourcepulse
This repository provides implementations and comparisons of language models specifically tuned for the Russian language, targeting researchers and developers working with Russian NLP. It offers pre-trained models and datasets for instruction tuning and chat-based interactions, aiming to advance Russian language AI capabilities.
How It Works
The project leverages instruction tuning and chat-based fine-tuning techniques on top of base LLaMA models. It introduces custom datasets like RuTurboAlpaca (instruction-following) and Saiga (chat-based conversations), generated using GPT models and curated for Russian. The approach focuses on adapting large language models to the nuances of the Russian language and common interaction patterns.
Quick Start & Requirements
llama_7b_ru_turbo_alpaca_lora
, saiga_7b_lora
).Highlighted Details
Maintenance & Community
The project is actively developed by Ilya Gusev. Links to demos and evaluation code are provided.
Licensing & Compatibility
Models are based on LLaMA, which has its own license. The datasets and code appear to be permissively licensed, but users should verify compatibility with LLaMA's terms for commercial use.
Limitations & Caveats
The README explicitly recommends using Saiga models over the original RuTurboAlpaca models, suggesting Saiga models are better supported and perform better on side-by-side metrics. The project is primarily focused on Russian language tasks.
11 months ago
1 week