Medical LLM for instruction-following in the medical domain
Top 51.8% on sourcepulse
This repository provides the official code and models for PMC-LLaMA, a family of open-source Large Language Models specifically designed for the medical domain. It aims to improve medical question answering and instruction following by pre-training on a large medical corpus and fine-tuning with instruction datasets, offering a specialized alternative to general-purpose LLMs for medical professionals and researchers.
How It Works
PMC-LLaMA follows a two-stage approach: first, pre-training a base LLaMA model on a vast collection of medical literature (PubMed Central papers and medical books), and second, fine-tuning this pre-trained model on an instruction-following dataset. This domain-specific pre-training is crucial for imbuing the model with medical knowledge, while instruction tuning enhances its ability to understand and respond to medical queries and tasks.
Quick Start & Requirements
conda install pytorch==1.13.0 torchvision==0.14.0 torchaudio==0.13.0 pytorch-cuda=11.6 -c pytorch -c nvidia
pip install transformers==4.28.1 sentencepiece datasets
transformers
library (e.g., axiong/PMC_LLaMA_13B
). See simple_test.py
for examples.Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
1 year ago
Inactive