LLM fuzzer for automated jailbreak detection
Top 51.6% on sourcepulse
FuzzyAI is an automated LLM fuzzing tool designed for developers and security researchers to identify and mitigate jailbreaks and vulnerabilities in LLM APIs. It supports a wide range of LLM providers and offers various attack techniques for comprehensive security testing.
How It Works
FuzzyAI employs a multi-pronged approach to LLM fuzzing, incorporating mutation-based, generation-based, and intelligent fuzzing techniques. It supports custom input generation and offers an extensible architecture for tailored testing. The tool can target various LLM providers, including OpenAI, Anthropic, Gemini, and Ollama, as well as custom REST APIs.
Quick Start & Requirements
poetry install
and poetry shell
.ollama pull llama3.1
).python run.py
.streamlit run webui.py
.Highlighted Details
Maintenance & Community
The project is maintained by CyberArk. Contributions are welcome via CONTRIBUTING.md
. Contact is available at fzai@cyberark.com
.
Licensing & Compatibility
Released under the Apache License 2.0, permitting commercial use and integration with closed-source projects.
Limitations & Caveats
Some classifiers are not compatible with all attack methods due to design differences (e.g., single-output vs. multi-output classifiers). When using Ollama, ensure all Ollama models are added before other model types. The port for Ollama defaults to 11434 and can be specified with -e port=...
. The SI-Attack implementation is text-based only.
2 weeks ago
1 day