Discover and explore top open-source AI tools and projects—updated daily.
Ta0ingAI model safety guardrails bypassed for cybersecurity tasks
New!
Top 94.7% on SourcePulse
Summary This project offers a modified Claude Code v2.1.88, specifically engineered to remove prompt-level cybersecurity restrictions. It targets users needing to generate code for security tasks previously blocked by AI safety guardrails, enhancing flexibility for security research, CTFs, and educational contexts.
How It Works
The core modification involves clearing the CYBER_RISK_INSTRUCTION constant. This directive previously blocked code generation for destructive techniques, DoS attacks, mass targeting, supply chain compromises, malicious detection evasion, and dual-use security tools without explicit authorization. This approach directly bypasses AI-imposed prompt-level safety filters.
Quick Start & Requirements The README mentions a "full version, including manual, compilation, etc." for Claude Code v2.1.88 but provides no explicit installation commands, build instructions, or specific prerequisites (e.g., Python version, GPU).
Highlighted Details
Maintenance & Community Modification date is March 31, 2026, attributed to "Claude (AI Assistant)". No details on contributors, sponsorships, community channels, roadmaps, or social media are provided.
Licensing & Compatibility The README does not specify a software license, leaving compatibility for commercial use or closed-source integration undetermined.
Limitations & Caveats This modification intentionally removes Anthropic's safety measures, potentially leading to unsafe outputs. Users assume all risks; production deployment is strongly discouraged; use is recommended only in controlled, isolated settings.
1 week ago
Inactive