Curated list of resources for multimodal large language model hallucination
Top 46.0% on sourcepulse
This repository serves as a comprehensive, curated collection of resources on hallucination in Multimodal Large Language Models (MLLMs), also known as Large Vision-Language Models (LVLMs). It targets researchers and practitioners by organizing papers, code, and datasets focused on analyzing, detecting, and mitigating visual hallucinations in MLLMs, aiming to improve the faithfulness and reliability of these models.
How It Works
The project categorizes resources into "Hallucination Survey," "Hallucination Evaluation & Analysis," and "Hallucination Mitigation." Papers are primarily listed by their contribution to new benchmarks/metrics or mitigation methods, ordered chronologically from newest to oldest within each category. This structure allows users to quickly identify the latest advancements and relevant techniques for addressing MLLM hallucinations.
Quick Start & Requirements
This is a curated list of research papers and code, not a runnable software package. No installation or execution commands are provided.
Highlighted Details
Maintenance & Community
This project is actively maintained and welcomes community contributions via pull requests for missing papers, new research, or corrections. Users can open issues or contact the maintainers directly via email.
Licensing & Compatibility
The repository itself is not software and therefore does not have a specific software license. The linked papers and code will have their own respective licenses.
Limitations & Caveats
As a curated list, this repository does not provide executable code or direct mitigation tools. Users must refer to individual linked papers for implementation details and potential usage. The rapidly evolving nature of MLLM research means the list requires continuous updates.
3 days ago
1 week