Discover and explore top open-source AI tools and projects—updated daily.
QwenLMSafety guardrails for LLM interactions
Top 76.6% on SourcePulse
Qwen3Guard is a series of multilingual safety moderation models designed to protect against harmful content in LLM applications. Targeting developers and researchers, it offers robust prompt and response analysis, with specialized variants for real-time, token-level monitoring during text generation, providing comprehensive, multi-language safety solutions.
How It Works
Qwen3Guard is built upon Qwen3 and trained on a large safety-labeled dataset. It comprises two main types: Qwen3Guard-Gen for static classification of prompts and responses, and Qwen3Guard-Stream for real-time, token-level safety assessment during incremental generation. The models support 119 languages and classify content into three severity levels (safe, controversial, unsafe) across nine defined safety categories, enabling adaptable risk management.
Quick Start & Requirements
transformers>=4.51.0. Deployment options include sglang>=0.4.6.post1 or vllm>=0.9.0 for OpenAI-compatible API endpoints.trust_remote_code=True is required for Qwen3Guard-Stream models.Highlighted Details
-Gen) and real-time token-level streaming moderation (-Stream).Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
Qwen3Guard-Stream model requires using the same tokenizer as Qwen3 for optimal performance; integration with different tokenizers necessitates re-tokenization.Qwen3Guard-Stream in vLLM and SGLang is listed as "coming soon."1 month ago
Inactive
protectai
NVIDIA-NeMo
guidance-ai