content-checker
TypeScript icon, indicating that this package has built-in type declarations

1.0.9 • Public • Published

Content-Checker (by OpenModerator)

content-checker is designed to be a modern, open-source library for programmatic and AI content moderation. Currently content-checker supports image and text moderation. Thanks to Google's Perspective API, in addition to detecting specific profane words, we can detect malicious intent in text. So, a user who tries to circumvent the AI profanity filter by using a variation of a profane word, or even just a malicious phrase without a specific word in the profanity list, will still be flagged. Image moderation is also supported, using the Inception V3 model of the NSFW JS library.

Future features will include moderation tools (auto-ban, bots), more powerful models, and multimedia support for video and audio moderation.

How It Works

content-checker builds on the popular bad-words package, but updated to use TypeScript and ES6, and merging in badwords-list, which it used as a dependency.

The AI moderation is powered by a database of profane words, Google's Perspective API (for text analysis) and the NSFW JS library (for image analysis). The models will likely be changed in the future as more powerful models become available. The importance of AI moderation for text is we need to be able to detect malicious intent, not just specific words.

The API returns responses in the following format for text moderation (the possible profanity types for now include "TOXICITY", "SEVERE_TOXICITY", "IDENTITY_ATTACK", "INSULT", "PROFANITY", "THREAT", and "SEXUALLY_EXPLICIT"). Text that has a 60% or greater probability of being in one of these categories will be flagged:

Unsafe text:

{
  "profane": true,
  "type": ["TOXICITY", "SEVERE_TOXICITY"]
}

Safe text:

{
  "profane": false,
  "type": []
}

The API returns responses in the following format for image moderation (the possible image types for now include "Porn" and "Hentai"). If the highest probability category is a NSFW category like the ones mentioned then the image will be flagged as NSFW:

Unsafe image:

{
  "nsfw": true,
  "type": ["Porn"]
}

Safe image:

{
  "nsfw": false,
  "type": []
}

Note that type is an array, so it can contain multiple types of profanity or none at all (if it hits the fine-tuned model). The API is rate limited to 10 requests from the same IP address per 10 seconds. If you need more, please contact me.

Installation

Use npm to install content-checker.

npm install content-checker

Table of Contents

  1. Standard Text Moderation
  2. AI Text Moderation
  3. AI Image Moderation

Usage

Standard Text Moderation

Initialize a filter

import { Filter } from "content-checker";
const filter = new Filter();

console.log(filter.clean("Don't be an ash0le")); //Don't be an ******

Placeholder Overrides

import { Filter } from "content-checker";
const customFilter = new Filter({ placeHolder: "x" });

customFilter.clean("Don't be an ash0le");

Regex Overrides

const filter = new Filter({ regex: /\*|\.|$/gi });

const filter = new Filter({ replaceRegex: /[A-Za-z0-9가-힣_]/g });

Add words to the blacklist

const filter = new Filter();

filter.addWords("some", "bad", "word");

filter.clean("some bad word!");

// or use an array using the spread operator

const newBadWords = ["some", "bad", "word"];

filter.addWords(...newBadWords);

filter.clean("some bad word!"); //**** *** ****!

// or

const filter = new Filter({ list: ["some", "bad", "word"] });

filter.clean("some bad word!"); // **** *** ****!

Instantiate with an empty list

const filter = new Filter({ emptyList: true });

filter.clean("hell this wont clean anything"); // hell this wont clean anything

Remove words from the blacklist

const filter = new Filter();

filter.removeWords("hells", "sadist");

filter.clean("some hells word!"); // some hells word!

// or use an array using the spread operator

let removeWords = ["hells", "sadist"];

filter.removeWords(...removeWords);

filter.clean("some sadist hells word!");

AI Text Moderation

To use AI text moderation, ensure you have the OPEN_MODERATOR_API_KEY set in your environment variables (one can be generated for free at www.openmoderator.com) or passed as a parameter during the initialization of the Filter class.

Initialize a filter

import { Filter } from "content-checker";

const filter = new Filter({ openModeratorAPIKey: "your_api_key_here" });

Check a string for profanity

The isProfaneAI method checks if a string contains profane language using AI. It returns a promise that resolves to an object containing a profane boolean and a type array which lists the categories of profanity detected (e.g., "harassment").

filter.isProfaneAI("your string here").then((response) => {
  if (response.profane) {
    console.log("Profanity found. Types: ", response.type.join(", "));
  } else {
    console.log("No profanity found");
  }
});

AI Image Moderation

To use AI image moderation, ensure you have the OPEN_MODERATOR_API_KEY set in your environment variables (one can be generated for free at www.openmoderator.com) or passed as a parameter during the initialization of the Filter class. For now the NSFW JS library is used for image moderation, but this will be replaced with a more powerful model in the future. Ensure you're uploading either a PNG or JPEG image.

Check an image for NSFW content

Raw JS example:

const imageInput = document.getElementById("imageInput");

imageInput.addEventListener("change", async (event) => {
  const file = event.target.files[0];
  if (file) {
    try {
      const response = await filter.isImageNSFW(file);
      if (response.nsfw) {
        console.log("NSFW content detected. Types:", response.type.join(", "));
      } else {
        console.log("Image is safe.");
      }
    } catch (error) {
      console.error("Error checking image:", error);
    }
  }
});

React example:

import React, { useState } from "react";
import { Filter } from "content-checker";

const ImageModeration = () => {
  const [image, setImage] = useState(null);
  const [moderationResult, setModerationResult] = useState("");

  const filter = new Filter({ openModeratorAPIKey: "your_api_key_here" });

  const handleImageChange = (event) => {
    setImage(event.target.files[0]);
  };

  const checkImage = async () => {
    if (image) {
      try {
        const response = await filter.isImageNSFW(image);
        if (response.nsfw) {
          setModerationResult(
            `NSFW content detected. Types: ${response.type.join(", ")}`,
          );
        } else {
          setModerationResult("Image is safe.");
        }
      } catch (error) {
        console.error("Error checking image:", error);
        setModerationResult("Error occurred while checking the image.");
      }
    }
  };

  return (
    <div>
      <input
        type="file"
        onChange={handleImageChange}
        accept="image/png, image/jpeg"
      />
      <button onClick={checkImage}>Check Image</button>
      {moderationResult && <p>{moderationResult}</p>}
    </div>
  );
};

export default ImageModeration;

Contributing

Pull requests are welcome. For major changes, please open an issue first to discuss what you would like to change.

Please make sure to update tests as appropriate.

Authors

  • Jacob Habib (@jahabeebs), founder of OpenModerator

Package Sidebar

Install

npm i content-checker

Weekly Downloads

9

Version

1.0.9

License

Apache-2.0

Unpacked Size

58.3 kB

Total Files

30

Last publish

Collaborators

  • jahabeebs