Chinese medical chatbot based on LLaMa, trained with RLHF
Top 77.4% on sourcepulse
Zhongjing-LLaMA is a Chinese medical large language model designed to address the limitations of general-purpose LLMs in specialized domains. It offers a comprehensive solution for medical dialogue, aiming to provide professional-level responses and improve active medical inquiry capabilities for researchers and developers in the medical AI field.
How It Works
Zhongjing-LLaMA employs a full training pipeline, including pre-training, supervised fine-tuning (SFT), and reinforcement learning with human feedback (RLHF). This approach leverages a large-scale pre-training corpus and a unique 70,000-utterance multi-turn dialogue dataset (CMtMedQA) derived from real doctor-patient interactions. The RLHF stage, incorporating expert feedback, is crucial for enhancing safety and professional accuracy, distinguishing it from models solely fine-tuned on single-turn dialogues.
Quick Start & Requirements
scripts/cli_demo.sh
with model paths and run.Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
The model is intended for research use only, and users assume all medical risks. While significant progress has been made, there is still room for improvement in safety and professionalism, with the potential for unexpected responses in certain situations.
1 year ago
1 day