• Community
  • Model
  • moderation-abuse-korean

moderation-abuse-korean

A text classification/moderation model that classifies Korean text into four concepts: hate speech, offensive language, gender bias, or other bias.

Notes

moderation-abuse-korean

Purpose

This model detects hate speech in Korean text. Outputs four concepts:

  • "3" : Hate
  • "2" : Offensive
  • "1" : Gender bias
  • "0" : Other bias

Architecture

This is a KcELECTRA model that is finetuned for hate speech detection.

The original KcELECTRA-base model is a pretrained ELECTRA model that is trained from scratch using comments from Naver News.

The dataset used for finetuning:

Intended Use

Multi-class classification of Korean text

Resources

  • ID
  • Name
    moderation-abuse-korean
  • Model Type ID
    Text Classifier
  • Description
    A text classification/moderation model that classifies Korean text into four concepts: hate speech, offensive language, gender bias, or other bias.
  • Last Updated
    Jan 22, 2023
  • Privacy
    PUBLIC
  • Use Case
  • Toolkit
  • License
  • Share
    • Badge
      moderation-abuse-korean