LLM security/privacy resources: papers, tools, datasets, blogs
Top 61.4% on sourcepulse
This repository compiles papers and resources on the security and privacy of Large Language Models (LLMs). It serves as a curated reference for researchers and practitioners entering or working within this nascent field, offering quick access to key studies on vulnerabilities, attacks, and defenses.
How It Works
The collection is organized by primary contribution, covering areas like prompt injection, jailbreaking, adversarial attacks, privacy concerns (data extraction, membership inference), watermarking, and LLM security in general. Papers are tagged with symbols indicating their focus (e.g., ⭐ for personally recommended, 💽 for datasets, 👤 for PII focus). The curator emphasizes that ⭐ is a subjective indicator of personal understanding and enjoyment, not a measure of paper quality.
Quick Start & Requirements
Highlighted Details
Maintenance & Community
The repository is maintained by the author, with contributions welcomed via GitHub issues or pull requests. The author also manually transfers updates to a Notion page.
Licensing & Compatibility
The repository itself is likely under a permissive license (e.g., MIT, Apache 2.0) given its nature as a curated list of research papers. The licensing of the individual papers referenced would vary by their original publication.
Limitations & Caveats
The paper selection is noted to be biased towards the curator's research interests, potentially leading to an incomplete overview. Distinctions between prompt injection, jailbreaking, and adversarial attacks can be fluid, and some papers may fit into multiple categories.
1 month ago
1 week