Get the latest tech news

Mistral launches a moderation API


AI startup Mistral has launched an API to moderate possibly toxic — or otherwise problematic — text in a range of languages.

It’s powered by a fine-tuned model ( Ministral 8B) trained to classify text in a range of languages, including English, French, and German, into one of nine categories: sexual, hate and discrimination, violence and threats, dangerous and criminal content, self-harm, health, financial, law, and personally identifiable information. “Our content moderation classifier leverages the most relevant policy categories for effective guardrails and introduces a pragmatic approach to model safety by addressing model-generated harms such as unqualified advice and PII.” “We’re working with our customers to build and share scalable, lightweight, and customizable moderation tooling,” the company said, “and will continue to engage with the research community to contribute safety advancements to the broader field.”

Get the Android app

Or read this on TechCrunch

Read more on:

Photo of API

API

Photo of Mistral

Mistral

Photo of moderation API

moderation API

Related news:

News photo

Apple Preparing for Upcoming Siri Onscreen Awareness Feature With New iOS 18.2 API for Developers

News photo

xAI woos developers with $25/month worth of API credits, support for OpenAI, Anthropic SDKs

News photo

Model Distillation in the API