Chinese Llama3 fine-tunes for chat, tutorials, and deployment
Top 12.0% on sourcepulse
This repository provides fine-tuned versions of Llama 3 and Llama 3.1 models specifically for Chinese language tasks. It caters to researchers and developers looking to leverage or build upon Llama 3 for Chinese NLP applications, offering pre-trained weights, tutorials for training, inference, evaluation, and deployment.
How It Works
The project fine-tunes Llama 3 base models using large, high-quality Chinese conversational datasets. It employs various techniques including Supervised Fine-Tuning (SFT) and Direct Preference Optimization (DPO) to enhance the models' performance in Chinese dialogue, instruction following, and specific tasks. The approach prioritizes leveraging the existing strong multilingual capabilities of Llama 3 and augmenting them with targeted Chinese data, rather than expanding the vocabulary.
Quick Start & Requirements
transformers
library, or via Ollama (ollama run shareai/llama3.1-dpo-zh
).transformers
, torch
, peft
, bitsandbytes
(for quantization). GPU recommended for inference (e.g., 10GB VRAM for 4-bit quantization, 24GB for FP16).Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
2 months ago
1 day