Chinese Llama 2 model for chat, fully open-source and commercially available
Top 20.7% on sourcepulse
This repository provides an open-source, commercially viable Chinese version of the Llama 2 7B model, along with a bilingual SFT dataset. It aims to offer a readily usable and adaptable Chinese language model for developers and researchers, enabling integration into various applications and further fine-tuning.
How It Works
The project fine-tunes the Llama 2 7B base model using a 10 million-entry bilingual SFT dataset. It adheres strictly to the Llama-2-chat input format, ensuring compatibility with existing optimizations and tools designed for the original Llama 2 chat model. This approach allows for seamless integration and leverages the established ecosystem around Llama 2.
Quick Start & Requirements
transformers
library for Python.torchrun
command is provided for fine-tuning.Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
The Colab demos require high RAM and may not be usable on free tiers. The project is actively updated, implying potential for breaking changes in future releases.
1 year ago
1 day