llm-hallucination-survey  by HillZhang1999

Survey of hallucination in LLMs

Created 2 years ago
1,051 stars

Top 35.8% on SourcePulse

GitHubView on GitHub
Project Summary

This repository serves as a comprehensive reading list and survey of research papers focused on hallucinations in Large Language Models (LLMs). It aims to provide researchers and practitioners with a structured overview of the problem, its various types, evaluation methods, sources, and mitigation strategies.

How It Works

The project categorizes LLM hallucinations into three main types: input-conflicting, context-conflicting, and fact-conflicting. It then meticulously lists and links to relevant research papers for each category, covering evaluation benchmarks, potential sources of hallucination, and diverse mitigation techniques applied during pretraining, fine-tuning, RLHF, and inference.

Quick Start & Requirements

This repository is a curated list of research papers and does not involve code execution or installation. All requirements are met by having internet access to view the linked papers.

Highlighted Details

  • Features a survey paper: "Siren’s Song in the AI Ocean: A Survey on Hallucination in Large Language Models."
  • Categorizes hallucinations into input-conflicting, context-conflicting, and fact-conflicting types.
  • Provides extensive lists of papers for evaluation, source analysis, and mitigation techniques.
  • Covers a wide range of mitigation strategies, including data curation, fine-tuning, RLHF, inference-time decoding, and external knowledge integration.

Maintenance & Community

The project is maintained by HillZhang1999. Contact is available via email for suggestions or contributions.

Licensing & Compatibility

The repository itself does not specify a license, but it links to numerous research papers, each with its own licensing and usage terms.

Limitations & Caveats

This is a curated list of research papers and does not provide code or tools for direct experimentation. The rapidly evolving nature of LLM research means new papers and findings may not be immediately reflected.

Health Check
Last Commit

2 weeks ago

Responsiveness

Inactive

Pull Requests (30d)
0
Issues (30d)
0
Star History
10 stars in the last 30 days

Explore Similar Projects

Starred by Chip Huyen Chip Huyen(Author of "AI Engineering", "Designing Machine Learning Systems"), Travis Fischer Travis Fischer(Founder of Agentic), and
1 more.

HaluEval by RUCAIBox

0.4%
516
Benchmark dataset for LLM hallucination evaluation
Created 2 years ago
Updated 1 year ago
Feedback? Help us improve.