LLM finetuned for medical question answering
Top 60.3% on sourcepulse
MedAlpaca provides open-source large language models fine-tuned for medical question answering and dialogue. It aims to facilitate the development of medical chatbot solutions by leveraging curated medical datasets. The project is primarily for researchers and developers in the medical AI space.
How It Works
MedAlpaca builds upon the Stanford Alpaca and AlpacaLoRA methodologies, fine-tuning LLaMA and OPT models on a diverse collection of medical texts. This includes medical flashcards, wikis, and dialogue datasets, collectively termed "Medical Meadow." The fine-tuning process supports various configurations, including 8-bit training and LoRA, to optimize for memory usage and training efficiency.
Quick Start & Requirements
pip install -r requirements.txt
Highlighted Details
Maintenance & Community
The project is associated with kbressem and has a published paper: arXiv:2304.08247. Further community interaction channels are not explicitly mentioned in the README.
Licensing & Compatibility
The models are provided for research purposes only and are not intended for healthcare applications. The README does not explicitly state a license for the code or data, but the models are hosted on Hugging Face, which typically uses Apache 2.0 for code and specific licenses for models.
Limitations & Caveats
The models are experimental, have not undergone extensive validation, and their reliability cannot be guaranteed. They are not multimodal and cannot process image-based questions. The project notes that the benchmark table is subject to change as training and evaluation methods improve.
1 year ago
Inactive