Get the latest tech news
Mistral launches a moderation API
AI startup Mistral has launched an API to moderate possibly toxic — or otherwise problematic — text in a range of languages.
It’s powered by a fine-tuned model ( Ministral 8B) trained to classify text in a range of languages, including English, French, and German, into one of nine categories: sexual, hate and discrimination, violence and threats, dangerous and criminal content, self-harm, health, financial, law, and personally identifiable information. “Our content moderation classifier leverages the most relevant policy categories for effective guardrails and introduces a pragmatic approach to model safety by addressing model-generated harms such as unqualified advice and PII.” “We’re working with our customers to build and share scalable, lightweight, and customizable moderation tooling,” the company said, “and will continue to engage with the research community to contribute safety advancements to the broader field.”
Or read this on TechCrunch