Workflow uses Gemma-7b-it Model with specified prompt template for Hate speech moderation that identify and filter out based on race, ethnicity, gender,
Use Workflow
Delete Workflow
Notes
Hate Speech Moderation
Workflow uses Gemma-7b-it Model with specified prompt template for Hate speech moderation that identify and filter out based on race, ethnicity, gender, sexual orientation, religion, or disability.
Text Moderation
Large Language Models (LLMs), like GPT (Generative Pretrained Transformer) variants, since they are pre-trained on diverse internet text, enabling them to understand context, nuance, and the subtleties of human language, to identify and filter out inappropriate or harmful content from digital platforms.
How to use the Toxicity Moderation workflow?
Using Clarifai SDK
Export your PAT as an environment variable. Then, import and initialize the API Client.
from clarifai.client.workflow import Workflow
workflow_url ='https://clarifai.com/clarifai/text-moderation/workflows/text-moderation-hate-speech-gemma-7b'text ='I love this movie and i would watch it again and again!'prediction = Workflow(workflow_url).predict_by_bytes(text.encode(), input_type="text")# Get workflow resultsprint(prediction.results[0].outputs[-1].data)
Using Workflow
To utilize the Text Moderation workflow, you can input text through the Blue Plus Try your own Input button and it will filter out 'Hate speech' sentiment based on race, ethnicity, gender, sexual orientation, religion, or disability. Indicate 'Hate Speech Detected' if hate speech is found, otherwise 'Not Detected'.
Workflow ID
text-moderation-hate-speech-gemma-7b
Description
Workflow uses Gemma-7b-it Model with specified prompt template for Hate speech moderation that identify and filter out based on race, ethnicity, gender, sexual orientation, religion, or disability.
Last Updated
Apr 09, 2024
Privacy
PUBLIC
Share
Copy URL
Twitter
Facebook
Reddit
LinkedIn
Email
Help
text-moderation-hate-speech-gemma-7b workflow by clarifai | Clarifai - The World's AI