Trained models for toxic comment classification
Top 35.7% on sourcepulse
Detoxify provides pre-trained models and code for classifying toxic comments across multiple datasets and languages. It is designed for researchers and developers working on content moderation, bias detection, and natural language understanding, offering a user-friendly interface to identify various forms of toxicity in text.
How It Works
The library leverages state-of-the-art transformer models (BERT, RoBERTa, XLM-RoBERTa) fine-tuned on Jigsaw's toxic comment datasets. It employs PyTorch Lightning for efficient training and Hugging Face Transformers for model architecture and tokenization. This approach allows for high performance and broad language support, with specific models optimized for general toxicity, unintended bias, and multilingual classification.
Quick Start & Requirements
pip install detoxify
Highlighted Details
original-small
, unbiased-small
) for reduced resource usage.Maintenance & Community
Licensing & Compatibility
LICENSE
file (not detailed here) is typical for open-source projects. Users should verify the license for commercial use or closed-source integration.Limitations & Caveats
4 days ago
1 week