LLM fine-tuning for Chinese language support
Top 89.6% on sourcepulse
MLE-LLaMA enhances the LLaMA model for improved Chinese language understanding and generation. It targets researchers and developers working with multilingual NLP, offering a fine-tuned LLaMA model with better Chinese fluency and instruction-following capabilities.
How It Works
The project leverages LLaMA's existing tokenizer, which naturally supports Chinese characters. Fine-tuning is achieved through two methods: a full fine-tuning script requiring significant GPU resources (80G A100) and a LoRA fine-tuning script using PEFT. The process involves a custom English-Chinese alignment dataset and instruction tuning using Chinese Alpaca and Guanaco datasets.
Quick Start & Requirements
ckpt
directory. Fine-tuning scripts (train.py
or train_lora.py
) are then executed.Highlighted Details
Maintenance & Community
No specific information on contributors, sponsorships, or community channels is provided in the README.
Licensing & Compatibility
The README does not explicitly state a license for the MLE-LLaMA project itself. It relies on LLaMA, which has its own licensing terms. Compatibility for commercial use depends on the underlying LLaMA license.
Limitations & Caveats
The full fine-tuning script has very high hardware requirements (80G A100). The project notes that LLaMA tends to generate long sentences, which may persist.
2 years ago
1 day