Science QA dataset & code for multimodal reasoning research
Top 50.7% on sourcepulse
This repository provides the dataset and code for the NeurIPS 2022 paper "Learn to Explain: Multimodal Reasoning via Thought Chains for Science Question Answering." It addresses the challenge of multimodal reasoning in science education by offering a benchmark dataset and methods for generating explanatory "thought chains" to solve complex questions. The target audience includes AI researchers and developers working on multimodal understanding, explainable AI, and educational AI applications.
How It Works
The ScienceQA dataset comprises over 21,000 multimodal multiple-choice questions spanning natural, language, and social sciences, featuring rich annotations including lectures and explanations. The core approach involves using language models to generate these "thought chains" (lectures and explanations) as a form of chain-of-thought (CoT) reasoning. This mimics human multi-hop reasoning processes, enhancing question-answering performance by providing step-by-step justifications.
Quick Start & Requirements
pip install -r requirements.txt
tools/download.sh
or download from Google Drive.cd models && python run_gpt3.py --label exp1 --test_split test --test_number -1 --shot_number 2 --prompt_format QCM-ALE --seed 3
Highlighted Details
Maintenance & Community
The project is actively maintained, with recent updates in late 2023 featuring over 100 models. Community engagement is encouraged via email, Twitter, and GitHub issues.
Licensing & Compatibility
Limitations & Caveats
The leaderboard data is manually collected and may contain errors or ambiguities. The dataset's CC BY-NC-SA 4.0 license restricts commercial use of the dataset itself.
10 months ago
1 week