Chinese adaptation of Meta's LLaMA2
Top 47.7% on sourcepulse
This repository provides Chinese-adapted versions of Meta's Llama 2 large language model, addressing its limited native Chinese capabilities. It targets researchers and developers seeking to deploy or fine-tune Llama 2 for Chinese language tasks, offering improved conversational abilities and specialized domain models.
How It Works
The project offers two primary approaches: supervised fine-tuning (SFT) on existing Chinese instruction datasets and continued pre-training on large Chinese corpora. For SFT, it utilizes datasets like UltraChat and Chinese Alpaca, with options for both extending the Llama 2 vocabulary and using the original. Continued pre-training aims to imbue the model with deeper Chinese knowledge. Specialized models for medical and traditional Chinese medicine domains are also under development.
Quick Start & Requirements
src/further_ft/download_checkpoints.py
.SFT-README.md
.vllm
for a ~2.7x speedup. See vllm-serving-README
.CUDA_VISIBLE_DEVICES="2,3"
).Highlighted Details
Maintenance & Community
Chinese-LlaMA2-chat-sft-v0.3
) released regularly.Licensing & Compatibility
Limitations & Caveats
2 years ago
1 day