LLM for Chinese classics
Top 75.5% on sourcepulse
This project provides a suite of large language models (LLMs) specifically designed for processing and understanding classical Chinese texts. Targeting researchers, linguists, and enthusiasts of Chinese classics, these models offer advanced capabilities for information extraction, translation, and analysis of ancient literature, significantly aiding scholarly work and cultural exploration.
How It Works
The Xunzi series offers both base and chat models, built upon established open-source LLMs like Qwen, ChatGLM3, and Baichuan2. This approach leverages the robust architectures of these foundational models while specializing them for classical Chinese through targeted fine-tuning. This strategy allows for efficient development and provides users with familiar calling methods, similar to their base model counterparts.
Quick Start & Requirements
openai
Python library.
from openai import OpenAI
openai_api_key = "ANY THING"
openai_api_base = "http://xunziallm.njau.edu.cn:21180/v1"
client = OpenAI(api_key=openai_api_key, base_url=openai_api_base)
chat_response = client.chat.completions.create(
model="/home/gpu0/xunzi_web/Xunzi-Qwen1.5-7B_chat",
messages=[{"role": "system", "content": "You are a helpful assistant."}, {"role": "user", "content": '...'}]
)
print(chat_response.choices[0].message.content)
openai
library. A hosted API endpoint is provided for Xunzi-Qwen1.5-7B_chat
.Highlighted Details
Xunzi-Qwen1.5-7B_chat
is accessible via an OpenAI-compatible API.Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
The project acknowledges that models still have room for improvement and may contain unavoidable issues due to data and model complexity. The developers disclaim responsibility for any problems arising from data security, public opinion risks, or misuse of the models. Compliance with China's generative AI regulations is advised.
2 weeks ago
1 week