agentic_security  by msoedov

Open-source vulnerability scanner for LLMs and agent workflows

created 1 year ago
1,572 stars

Top 27.1% on sourcepulse

GitHubView on GitHub
Project Summary

Agentic Security is an open-source toolkit for AI red teaming and vulnerability scanning, designed to protect Large Language Models (LLMs) from various attacks. It targets developers, researchers, and security teams seeking to proactively identify and mitigate risks in AI systems, offering robust defenses against jailbreaks, fuzzing, and multimodal threats.

How It Works

The tool employs an agentic approach, simulating sophisticated attack sequences and stress-testing LLMs with diverse inputs. It supports multimodal attacks (text, image, audio), multi-step jailbreaks, comprehensive fuzzing, API integration, and RL-based adaptive attacks. This methodology allows for proactive identification and mitigation of AI system vulnerabilities.

Quick Start & Requirements

  • Install via pip: pip install agentic_security
  • Run using the CLI: agentic_security or python -m agentic_security
  • Supports custom datasets via CSV files.
  • For detailed usage and advanced features, refer to the official documentation.

Highlighted Details

  • Supports multimodal attacks across text, image, and audio inputs.
  • Enables multi-step jailbreak simulations and comprehensive fuzzing.
  • Integrates with LLM APIs for stress testing and features RL-based attack capabilities.
  • Offers a CLI for initialization, listing datasets, and running CI checks.

Maintenance & Community

The project is actively developed, with a roadmap including RL-powered attacks, massive dataset expansion, daily attack updates, and community modules. Contributions are welcome via pull requests on GitHub.

Licensing & Compatibility

Released under the Apache License v2.0, allowing for commercial use and integration with closed-source projects.

Limitations & Caveats

The project is described as "just getting started," with some features like RL-powered attacks and extensive dataset expansion still under development. Custom integration examples for multimodal probes are provided, but specific requirements for these may vary.

Health Check
Last commit

5 days ago

Responsiveness

Inactive

Pull Requests (30d)
18
Issues (30d)
0
Star History
248 stars in the last 90 days

Explore Similar Projects

Starred by Chip Huyen Chip Huyen(Author of AI Engineering, Designing Machine Learning Systems), Michele Castata Michele Castata(President of Replit), and
2 more.

rebuff by protectai

0.4%
1k
SDK for LLM prompt injection detection
created 2 years ago
updated 1 year ago
Starred by Chip Huyen Chip Huyen(Author of AI Engineering, Designing Machine Learning Systems), Carol Willing Carol Willing(Core Contributor to CPython, Jupyter), and
2 more.

llm-security by greshake

0.2%
2k
Research paper on indirect prompt injection attacks targeting app-integrated LLMs
created 2 years ago
updated 2 weeks ago
Feedback? Help us improve.