Chinese medical dialogue model for proactive health applications
Top 43.1% on sourcepulse
BianQue is an open-source Chinese medical dialogue large language model developed by South China University of Technology. It aims to improve the "questioning" ability of AI in healthcare, mimicking how doctors gather information through multi-turn conversations, and also provides health suggestions. The project targets researchers and developers in the active health domain, particularly for chronic diseases and mental health.
How It Works
BianQue is built upon the ProactiveHealthGPT foundation model and fine-tuned on the "BianQue Health Big Data" corpus, a dataset of millions of Chinese health dialogues. This corpus emphasizes multi-turn questioning (Chain of Questioning - CoQ) to address the limitation of single-turn interactions in existing medical LLMs. BianQue-2.0, specifically, uses ChatGLM-6B as a base and is further fine-tuned with instruction data including drug inserts, medical encyclopedias, and ChatGPT distillation, enhancing its suggestion and knowledge retrieval capabilities.
Quick Start & Requirements
proactivehealthgpt_py38.yml
. Install dependencies with pip install cpm_kernels
and specific PyTorch versions (e.g., torch==1.13.1+cu116
).streamlit run bianque_v2_app.py --server.port 9005
.Highlighted Details
Maintenance & Community
The project is initiated by the Future Technology School and Guangdong Key Laboratory of Digital Human at South China University of Technology. Contributions are encouraged via GitHub issues and PRs. Collaboration with academic institutions, hospitals, and companies is welcomed. Contact: eeyirongchen@mail.scut.edu.cn.
Licensing & Compatibility
BianQue-2.0 uses weights from ChatGLM-6B and is restricted to non-commercial research purposes due to its MODEL_LICENSE. Users must adhere to terms prohibiting commercial, military, or illegal use, and acknowledge that the model output is not a substitute for professional medical advice.
Limitations & Caveats
BianQue-1.0, trained for one epoch, may produce errors due to noise in the training data and exhibits an "inquisitive" style that might be idiosyncratic. Its "observation," "listening," and "pulse-taking" (望闻问切) capabilities require further research. The project disclaims responsibility for model output suitability and user reliance on its advice.
1 year ago
1 week