Open-source medical LLMs adapted from Llama-2
Top 22.1% on sourcepulse
Meditron offers open-source medical Large Language Models (LLMs), specifically Meditron-7B and Meditron-70B, adapted from Llama-2. These models are designed for researchers and developers working with medical data, aiming to improve performance on medical reasoning tasks and provide a foundation for specialized healthcare AI applications.
How It Works
Meditron models are created by continuing the pretraining of Llama-2 on a curated medical corpus. This corpus, GAP-Replay, combines clinical guidelines, PubMed abstracts, full-text medical papers, and general domain data. This domain-adaptive pretraining approach enhances the models' understanding and generation capabilities within the medical field, outperforming general LLMs on specific medical reasoning benchmarks.
Quick Start & Requirements
transformers
.
from transformers import AutoTokenizer, AutoModelForCausalLM
tokenizer = AutoTokenizer.from_pretrained("epfl-llm/meditron-70b")
model = AutoModelForCausalLM.from_pretrained("epfl-llm/meditron-70b")
vllm >= 0.2.1
, transformers >= 4.34.0
, datasets >= 2.14.6
, torch >= 2.0.1
.Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
The models are not yet adapted for safe or appropriate delivery of medical knowledge in clinical practice. Use in medical applications is strongly discouraged without extensive alignment, testing, and potentially randomized controlled trials. The models are static and knowledge cutoff is August 2023.
1 year ago
Inactive