New image toxicity model

New image toxicity model

The new image toxicity model adds a single but robust label for detecting and preventing harmful images.

Where the image NSFW model can distinguish between multiple types of unwanted content, it can fail to generalise to toxic content outside of the provided labels.

The toxicity model on the other hand is better at understanding a wider range of toxicity, but provides just a single label.

Feel free to use the two models alongside each other. For example use the NSFW model if you want to give users a reason for the flag, but include the toxicity model to make sure you catch all unwanted content.

How to use the model

You can find the toxic image model in your project configuration now. Add it under "Pre-built models".

Support and feedback

We are committed to providing you with the best tools for content moderation. If you have any questions, encounter issues, or have suggestions for further improvements, please don't hesitate to contact our support team.

We look forward to seeing how this new model enhance your automated moderation and contribute to maintaining a safe and respectful online environment.

Happy Moderating!

Read more