Skip to content

Classifying various toxic comments using DistilBert

Notifications You must be signed in to change notification settings

rimo02/jigsaw-toxicity

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

5 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Toxic Comment Classification using DistilBERT

Overview

This project aims to build a multi-headed model for detecting various types of toxicity in online comments, including:

  • Toxic
  • Severe Toxic
  • Obscene
  • Threat
  • Insult
  • Identity Hate

Utilizing the DistilBERT model, this approach fine-tunes the model on a dataset of comments from Wikipedia's talk page edits. The model is then deployed using FastAPI, allowing for easy interaction and testing of its capabilities.

Sample

image


Note

Run the Jupyter notebook to train the model or contact me to get the pretrained model weights

About

Classifying various toxic comments using DistilBert

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published