CVPR'24 research on aligning MLLMs via fine-grained human feedback
Top 92.3% on sourcepulse
RLHF-V provides a framework for aligning Multimodal Large Language Models (MLLMs) using fine-grained human feedback to reduce hallucinations. It targets researchers and developers aiming to improve MLLM trustworthiness, offering a data-efficient method to enhance model behavior.
How It Works
The framework leverages fine-grained correctional human feedback, where annotators correct hallucinated segments in MLLM responses. This approach prioritizes data efficiency, allowing for significant hallucination rate reduction with minimal training time. The core methodology involves aligning MLLM behavior through this targeted feedback, enhancing reliability.
Quick Start & Requirements
conda create -n muffin python=3.10
, conda activate muffin
). Install dependencies via pip install -e .
. Specific versions of transformers
and flash-attention
are recommended for reproducibility.flash-attention
), COCO2014 dataset annotations for Object HalBench evaluation.Highlighted Details
Maintenance & Community
The project is associated with THU NLP and has seen contributions and integrations with other models like MiniCPM-V 2.0 and OmniLMM-12B. Updates are regularly posted on Hugging Face and arXiv.
Licensing & Compatibility
Limitations & Caveats
The dataset and models are strictly for research purposes and non-commercial use. Compatibility is tied to the licenses of the base models used.
10 months ago
1 week