Codebase for research paper assessing LLMs in rare disease QA
Top 70.8% on sourcepulse
This repository provides the official codebase for the paper "Assessing and Enhancing Large Language Models in Rare Disease Question-answering." It offers a benchmark dataset, ReDis-QA, and a corpus, ReCOP, designed to evaluate and improve LLM performance on rare disease-related queries, targeting researchers and developers in medical AI.
How It Works
The project evaluates LLMs using a question-answering framework focused on rare diseases. It supports both zero-shot LLM execution and Retrieval-Augmented Generation (RAG) approaches. RAG implementations utilize metadata retrievers, BM25, and MedCPT retrievers, with options to combine the ReCOP corpus with baseline corpora like PubMed, Textbooks, and Wikipedia for enhanced retrieval accuracy.
Quick Start & Requirements
pip install -r requirements.txt
datasets
library (load_dataset
).zero-shot-bench/scripts/run_exp.sh
, meta-data-bench/scripts/run_exp.sh
, rag-bench/scripts/run_exp.sh
, combine-corpora-bench/scripts/run_exp.sh
.Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
The project is presented as the codebase for a research paper, implying it may be experimental. The absence of a specified license and community support could pose adoption challenges.
11 months ago
Inactive