bert-hateXplain / README.md
Kanit's picture
Update README.md
297b087
|
raw
history blame
912 Bytes
metadata
datasets:
  - hatexplain
language:
  - en
pipeline_tag: text-classification
metrics:
  - accuracy
  - f1
  - precision
  - recall

BERT for hate speech classification

The model is based on BERT and used for classifying a text as toxic and non-toxic. It achieved an F1 score of 0.81 and an Accuracy of 0.77.

The model was fine-tuned on the HateXplain dataset found here: https://huggingface.co/datasets/hatexplain

How to use

from transformers import AutoModelForSequenceClassification, AutoTokenizer, pipeline

# Load model and tokenizer
tokenizer = AutoTokenizer.from_pretrained('tum-nlp/bert-hateXplain')
model = AutoModelForSequenceClassification.from_pretrained('tum-nlp/bert-hateXplain')

# Create the pipeline for classification
hate_classifier = pipeline("text-classification", model=model, tokenizer=tokenizer)

# Predict
hate_classifier("I like you. I love you")