Discover and explore top open-source AI tools and projects—updated daily.
Safety guardrails for LLM interactions
New!
Top 98.4% on SourcePulse
Qwen3Guard is a series of multilingual safety moderation models designed to protect against harmful content in LLM applications. Targeting developers and researchers, it offers robust prompt and response analysis, with specialized variants for real-time, token-level monitoring during text generation, providing comprehensive, multi-language safety solutions.
How It Works
Qwen3Guard is built upon Qwen3 and trained on a large safety-labeled dataset. It comprises two main types: Qwen3Guard-Gen
for static classification of prompts and responses, and Qwen3Guard-Stream
for real-time, token-level safety assessment during incremental generation. The models support 119 languages and classify content into three severity levels (safe, controversial, unsafe) across nine defined safety categories, enabling adaptable risk management.
Quick Start & Requirements
transformers>=4.51.0
. Deployment options include sglang>=0.4.6.post1
or vllm>=0.9.0
for OpenAI-compatible API endpoints.trust_remote_code=True
is required for Qwen3Guard-Stream
models.Highlighted Details
-Gen
) and real-time token-level streaming moderation (-Stream
).Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
Qwen3Guard-Stream
model requires using the same tokenizer as Qwen3 for optimal performance; integration with different tokenizers necessitates re-tokenization.Qwen3Guard-Stream
in vLLM and SGLang is listed as "coming soon."2 weeks ago
Inactive