SDK for LLM prompt injection detection
Top 31.0% on sourcepulse
Rebuff is an open-source prompt injection detector designed to protect AI applications from malicious inputs. It targets developers and security professionals seeking to safeguard LLM-based systems. Rebuff offers a multi-layered defense strategy to identify and mitigate prompt injection attacks.
How It Works
Rebuff employs a four-layer defense mechanism: heuristic filtering of suspicious inputs, LLM-based analysis of prompts, a vector database to store and recognize embeddings of known attacks, and canary tokens to detect data leakage. This layered approach aims to provide robust protection by combining signature-based detection with behavioral analysis and proactive monitoring.
Quick Start & Requirements
pip install rebuff
gpt-3.5-turbo
).Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
Rebuff is a prototype and does not guarantee 100% protection against prompt injection attacks. A Python SDK is still under development, and features like local-only mode and user-defined detection strategies are planned.
1 year ago
1 day