Discover and explore top open-source AI tools and projects—updated daily.
yunwei37LLM prompt hacking and defense resource
Top 92.4% on SourcePulse
Summary
This repository provides a curated collection of prompt adversarial techniques, focusing on prompt injection attacks, defenses, and reverse engineering examples. It serves as a valuable resource for researchers, students, and security professionals seeking to understand and practice LLM security vulnerabilities and mitigation strategies.
How It Works
The project organizes various prompt types, including jailbreaks, reverse engineering, attacks, and defenses, primarily in YAML format for easy parsing and application. It details concepts and provides concrete examples, such as the "DAN 11.0" jailbreak prompt, illustrating methods to bypass LLM restrictions and explore their security boundaries.
Quick Start & Requirements
This repository primarily serves as a reference collection rather than a runnable application. Installation instructions or specific software/hardware prerequisites are not detailed. Users are expected to leverage the provided prompt examples directly within their LLM interaction environments. Relevant external resources on LLM safety are linked.
Highlighted Details
Maintenance & Community
The project encourages community contributions through issues and pull requests, fostering collaborative development. Specific details regarding active maintainers, community channels (like Discord/Slack),
10 months ago
Inactive
protectai
protectai
meta-llama