yaekobB/Toxic-Comment-Classification
Multi-label toxic comment classification using DistilBERT with explainable AI via Captum Integrated Gradients (IG). Trained on the Jigsaw dataset, the model predicts six toxicity categories : toxic, severe toxic, obscene, threat, insult, and identity hate , while highlighting key words driving each prediction.
Stars
1
Forks
—
Language
Jupyter Notebook
License
—
Category
Last pushed
Oct 16, 2025
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/yaekobB/Toxic-Comment-Classification"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
StyrbjornKall/TRIDENT
A collection of transformer-based models and developmental scripts presented in the publication...
Nithin-Holla/meme_challenge
Repository containing code from team Kingsterdam for the Hateful Memes Challenge
viddexa/moderators
One package to moderate them all
jaygala24/fed-hate-speech
The official code repository for the paper titled "A Federated Approach for Hate Speech...
richouzo/hate-speech-detection-survey
Trained Neural Networks (LSTM, HybridCNN/LSTM, PyramidCNN, Transformers, etc.) & comparison for...