detoxify and toxic

These are competitors: both independently solve the same Jigsaw Toxic Comment Classification task, with Unitary AI's detoxify being the more mature and widely-adopted solution (production API, active maintenance, higher adoption metrics) compared to Ostyakov's implementation.

detoxify
83
Verified
toxic
49
Emerging
Maintenance 13/25
Adoption 24/25
Maturity 25/25
Community 21/25
Maintenance 0/25
Adoption 10/25
Maturity 16/25
Community 23/25
Stars: 1,202
Forks: 141
Downloads: 94,691
Commits (30d): 2
Language: Python
License: Apache-2.0
Stars: 266
Forks: 73
Downloads:
Commits (30d): 0
Language: Python
License: MIT
No risk flags
Stale 6m No Package No Dependents

About detoxify

unitaryai/detoxify

Trained models & code to predict toxic comments on all 3 Jigsaw Toxic Comment Challenges. Built using ⚡ Pytorch Lightning and 🤗 Transformers. For access to our API, please email us at contact@unitary.ai.

Provides three distinct model variants—`original`, `unbiased`, and `multilingual`—each optimized for different toxicity detection scenarios, with lightweight ALBERT-based alternatives for resource-constrained deployments. Leverages transformer-based architectures with bias-aware training on aggregated annotator judgments, supporting multi-label classification across toxicity subtypes (obscenity, threats, identity attacks, etc.) and identity mentions. Exposes predictions via a simple Python API returning per-category confidence scores and supports inference across seven languages with per-language performance metrics.

About toxic

PavelOstyakov/toxic

Toxic Comment Classification Challenge

Implements multi-label toxic comment classification using Keras with fastText embeddings (300d), processing six toxicity categories simultaneously. The pipeline combines NLTK preprocessing with scikit-learn utilities to train deep learning models that achieve competitive leaderboard performance on the Kaggle Jigsaw competition dataset.

Scores updated daily from GitHub, PyPI, and npm data. How scores work