Awesome-LLM-hallucination  by LuckyyySTA

Survey paper list on LLM hallucination

created 2 years ago
320 stars

Top 86.0% on sourcepulse

GitHubView on GitHub
Project Summary

This repository provides a comprehensive survey of hallucination in Large Language Models (LLMs), categorizing causes, detection methods, and mitigation strategies. It serves as a valuable resource for researchers and practitioners aiming to understand and address the phenomenon of LLM hallucination.

How It Works

The survey systematically categorizes LLM hallucinations into factuality and faithfulness types. It then breaks down the causes into data, model training, and inference stages, providing a structured overview of the problem space. The repository also curates extensive lists of papers related to each category, including surveys, detection benchmarks, and mitigation techniques.

Quick Start & Requirements

This repository is a curated list of research papers and does not require installation or execution. The primary resource is the survey paper itself, available on arXiv.

Highlighted Details

  • Comprehensive taxonomy of LLM hallucination causes (data, training, inference).
  • Extensive lists of papers covering hallucination detection and mitigation techniques.
  • Categorization of detection methods into fact-based, classifier-based, QA-based, and uncertainty estimation.
  • Detailed review of hallucination evaluation and detection benchmarks.

Maintenance & Community

The repository is maintained by LuckyyySTA and associated authors from Harbin Institute of Technology and Huawei Inc. The first version of the paper was released on arXiv in November 2023.

Licensing & Compatibility

The repository itself does not specify a license. The survey paper is available under an arXiv license.

Limitations & Caveats

As a survey, this repository is a snapshot of the field as of its publication date. The rapidly evolving nature of LLM research means new papers and techniques are constantly emerging.

Health Check
Last commit

1 year ago

Responsiveness

Inactive

Pull Requests (30d)
0
Issues (30d)
0
Star History
6 stars in the last 90 days

Explore Similar Projects

Starred by Chip Huyen Chip Huyen(Author of AI Engineering, Designing Machine Learning Systems) and Travis Fischer Travis Fischer(Founder of Agentic).

HaluEval by RUCAIBox

0%
497
Benchmark dataset for LLM hallucination evaluation
created 2 years ago
updated 1 year ago
Starred by Jerry Liu Jerry Liu(Cofounder of LlamaIndex), Chip Huyen Chip Huyen(Author of AI Engineering, Designing Machine Learning Systems), and
1 more.

hallucination-leaderboard by vectara

0.9%
3k
LLM leaderboard for hallucination detection in summarization
created 1 year ago
updated 1 day ago
Feedback? Help us improve.