moderation-abuse-korean

A text classification/moderation model that classifies Korean text into four concepts: hate speech, offensive language, gender bias, or other bias.

1

여자는 의사가 될 수 없습니다

Clarifai

Sign up or Log in to view model predictions.

Notes

moderation-abuse-korean

Purpose

This model detects hate speech in Korean text. Outputs four concepts:

  • "3" : Hate
  • "2" : Offensive
  • "1" : Gender bias
  • "0" : Other bias

Architecture

This is a KcELECTRA model that is finetuned for hate speech detection.

The original KcELECTRA-base model is a pretrained ELECTRA model that is trained from scratch using comments from Naver News.

The dataset used for finetuning:

Intended Use

Multi-class classification of Korean text

Resources

  • ID
  • Model Type ID
    Text Classifier
  • Input Type
    text
  • Output Type
    concepts
  • Description
    A text classification/moderation model that classifies Korean text into four concepts: hate speech, offensive language, gender bias, or other bias.
  • Last Updated
    Jan 22, 2023
  • Privacy
    PUBLIC
  • Use Case
  • Toolkit
  • License
  • Share
  • Badge
    moderation-abuse-korean