Discover and explore top open-source AI tools and projects—updated daily.
LLM for Chinese LLaMA-2, supporting incremental pre-training
Top 70.8% on SourcePulse
Firefly-LLaMA2-Chinese offers open-source Chinese-English bilingual large language models based on LLaMA2. It addresses the need for efficient, low-resource incremental pre-training and instruction fine-tuning for various LLMs, including LLaMA2, Baichuan2, Qwen, and others. The project provides pre-trained and fine-tuned model weights, along with the full training code and datasets, enabling researchers and developers to replicate or build upon their work.
How It Works
The project employs a low-resource incremental pre-training approach, primarily using QLoRA. This involves expanding the vocabulary of base models like LLaMA2 with Chinese tokens to improve encoding efficiency, followed by incremental pre-training on a 22GB Chinese-English corpus. Subsequently, models undergo multi-turn instruction fine-tuning using a large dataset of Chinese and English conversational instructions. This methodology significantly reduces the GPU resources required compared to full fine-tuning, making advanced LLM customization more accessible.
Quick Start & Requirements
script/chat
.Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
1 year ago
Inactive