HuatuoGPT-II  by FreedomIntelligence

Medical LLM for Chinese medical applications

created 2 years ago
390 stars

Top 74.7% on sourcepulse

GitHubView on GitHub
Project Summary

HuatuoGPT-II is an open-source Large Language Model specifically adapted for the medical domain, targeting researchers and developers in medical AI. It offers significant improvements in medical knowledge and dialogue capabilities, demonstrated by state-of-the-art performance on Chinese medical benchmarks and professional exams, even outperforming GPT-4 in expert evaluations.

How It Works

HuatuoGPT-II utilizes a novel one-stage domain adaptation method. This approach involves transforming pre-training corpora into instruction-output pairs and then applying a priority sampling algorithm for data processing before a single-stage training phase. This method is designed to enhance model adaptation across languages and domains efficiently.

Quick Start & Requirements

  • Install/Run: Inference can be performed using Hugging Face Transformers. A command-line demo is also available.
  • Prerequisites: PyTorch, Hugging Face Transformers. Models are available in 7B, 13B, and 34B parameter sizes, with quantized versions (4-bit, 8-bit) also provided.
  • Resources: Requires significant GPU resources for larger models; quantized versions mitigate this.
  • Links: Online Demo (7B), Models on Hugging Face, Paper

Highlighted Details

  • Achieves state-of-the-art performance on Chinese medical benchmarks.
  • Outperforms GPT-4 in expert evaluations for single-round medical responses (38% win rate) and multi-round dialogues (53% win rate).
  • Achieved top results on the 2023 Chinese National Pharmacist Licensure Examination.
  • Training code and a substantial portion of training data (142k fine-tuning, 5.2M pre-training) are publicly available.

Maintenance & Community

The project is associated with the School of Data Science, CUHKSZ, and the Shenzhen Research Institute of Big Data. Updates are ongoing, with recent news including paper acceptance at COLM 2024 and public release of training data.

Licensing & Compatibility

The repository does not explicitly state a license. The models are available on Hugging Face, implying a permissive license for use, but specific terms should be verified.

Limitations & Caveats

The project focuses on Chinese medical applications, and performance on other languages or medical contexts may vary. While code and data are being released, some aspects are still being organized. The evaluation benchmarks are specific to medical QA and professional exams.

Health Check
Last commit

11 months ago

Responsiveness

1+ week

Pull Requests (30d)
0
Issues (30d)
0
Star History
22 stars in the last 90 days

Explore Similar Projects

Feedback? Help us improve.