Skip to content

minhvyha/TikGuard

Repository files navigation

TikGuard

Links

About

TikGuard is a frontend interface developed using Next.js and deployed via Vercel. It demonstrates a potential machine learning model for detecting hate speech and offensive content in TikTok posts, captions, and comments.

TikGuard features a chatbot-like interface where users can upload text or media. The app then uses the selected API to identify and detect hateful, offensive, and inappropriate content in the uploaded attachments.

Demo

To access TikGuard, simply launch your preferred browser and enter the following URL:

https://tikguard.vercel.app/

You will see a landing page similar to this (Dark Mode):

TikGuardLanding

Supported Analysis

TikGuard supports the following types of analysis:

TikGuard uses a machine learning model built with Scikit-learn to determine the level of hate speech in a given text. This model leverages natural language processing techniques to analyze and classify text inputs, providing an effective tool for detecting and mitigating harmful content on social media platforms. You can find the code for the ML model here.

Text Analysis

As mentioned earlier, we use 4 different APIs to analyze the severity of text in different ways:

Expert.ai measures the severity of text based on the frequency count of inappropriate words used in the context, these counts are also categorized as follows:

  • Personal Insults
  • Racism
  • Sexism
  • Ableism
  • Religous Hatred
  • Homophobia
  • Classism
  • Body Shaming
  • Threat and Violence

Azure AI content safety works similarly to Expert.ai however with a different categorization:

  • Hate
  • Self-harm
  • Sexual
  • Violence

HateHoundApi is different from the other APIs and measures the severity based on the probability of hateful and offensive content. Content is categorized as follows:

  • Toxic
  • Severe toxicity
  • Obscene words
  • Threats
  • Insults
  • Discrimination

Sightengine also provides a probability however it is a much more robust API than hatehound. Content in the Sightengine API is categorized as follows:

  • Sexual based words
  • Discrimination
  • Self-harm statements
  • Insulting words
  • Violent words
  • Toxic words

Image Analysis

We use two kinds of APIs for image analysis:

Sightengine provides a probability between 0-1 for images. Content in the Sightengine API is categorized as follows:

  • Nudity
  • Recreational drug usage
  • Offensive
  • Tobacco
  • Self-harm
  • Gambling
  • Weaponry
  • Alcohol
  • Gore
  • Violence
  • Money

Azure AI content moderator works with a percentage on two measures:

  • Racism
  • Sexual Content

Video Analysis

We only provide one kind of API for video analysis:

Sightengine provides a probability between 0-1 for videos. This also turns the audio in videos into text to be analysed. Content in the Sightengine API is categorized as follows:

  • Nudity
  • Recreational drug usage
  • Offensive
  • Tobacco
  • Self-harm
  • Gambling
  • Weaponry
  • Alcohol
  • Gore
  • Violence
  • Money

About

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors