Chinese chat model fine-tuned from Llama3-8B-Instruct
Top 85.5% on sourcepulse
Llama3-Chinese-Chat is a fine-tuned version of Meta-Llama-3-8B-Instruct, specifically optimized for Chinese language interactions. It aims to address issues like "Chinese questions with English answers" and mixed language responses, offering enhanced capabilities in roleplay, function calling, and math for both Chinese and English users.
How It Works
This model is fine-tuned using ORPO (Odds Ratio Preference Optimization), a method that refines the model's preference for desired responses. It builds upon the Llama-3-8B-Instruct base model, leveraging a significantly larger dataset (up to 100K preference pairs in v2.1) to improve performance across various conversational tasks. The training framework used is LLaMA-Factory.
Quick Start & Requirements
ollama run wangshenzhi/llama3-8b-chinese-chat-ollama-q4
(for q4_0 GGUF)shenzhi-wang/Llama3-8B-Chinese-Chat-GGUF-*
repositories.transformers
library with model_id = "shenzhi-wang/Llama3-8B-Chinese-Chat"
.transformers
, torch
, llama-cpp-python
(for GGUF). GPU recommended for optimal performance.Highlighted Details
internet_search
and send_email
.Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
1 year ago
1 day