Curated list of LLM safety resources for researchers/practitioners
Top 27.9% on sourcepulse
This repository curates a comprehensive collection of resources on Large Language Model (LLM) safety, targeting researchers and practitioners. It provides a structured overview of papers, tutorials, articles, and datasets covering security, privacy, truthfulness, jailbreaking, defenses, and benchmarks, aiming to keep users updated on the latest advancements and challenges in LLM safety.
How It Works
The repository organizes information into distinct categories such as Security & Discussion, Privacy, Truthfulness & Misinformation, Jailbreak & Attacks, Defenses & Mitigation, and Datasets & Benchmarks. Within each category, resources are further categorized into papers and tutorials/articles/presentations, with links and brief descriptions provided. The project is actively updated with new findings and aims to be a central hub for LLM safety research.
Quick Start & Requirements
Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
The repository is a curated list and does not host any code or models itself. The quality and relevance of the information are dependent on the contributors and the ongoing maintenance of the links.
4 days ago
1 day