Discover and explore top open-source AI tools and projects—updated daily.
LLMSecurityAutomated prompt injection for LLM applications
Top 99.3% on SourcePulse
<2-3 sentences summarising what the project addresses and solves, the target audience, and the benefit.> HouYi is an automated framework for prompt injection attacks against LLM-integrated applications. It enables researchers and security professionals to systematically test and identify vulnerabilities by automatically crafting and injecting malicious prompts, thereby facilitating the discovery of prompt injection flaws.
How It Works
The framework defines "harnesses" to simulate or interact with target LLM applications and "attack intentions" for desired malicious outcomes. HouYi uses GPT models (requiring an OpenAI API key) to iteratively generate and refine injection prompts. It then employs the harness to deliver these prompts, evaluating success based on the application's response to achieve the defined intention.
Quick Start & Requirements
pip install -r requirements.txt.example.config.json to config.json and insert your OpenAI API key.python main.py for a translator attack demonstration.Highlighted Details
Maintenance & Community
Primary contributors are Yi Liu and Gelei Deng, with contact emails provided. No community channels (Discord, Slack), public roadmaps, or active development signals beyond the initial release are indicated.
Licensing & Compatibility
The repository's README does not specify a software license. This omission may restrict usage, modification, and distribution, particularly for commercial applications. Users should seek clarification on licensing terms.
Limitations & Caveats
HouYi depends on the OpenAI API, incurring potential costs and external service dependency. As a research artifact, its production readiness may require significant adaptation. Effectiveness is tied to the underlying LLM and the target application's harness implementation.
1 year ago
Inactive
protectai
greshake
protectai