Get the latest tech news
Composo helps enterprises monitor how well AI apps work
AI and the large language models (LLMs) that power them have a ton of useful applications, but for all their promise, they're not very reliable. No one
London-based startup Composo feels it has a headstart in trying to solve that problem, thanks to its custom models that can help enterprises evaluate the accuracy and quality of apps that are powered by LLMs. The company’s similar to Agenta, Freeplay, Humanloop and LangSmith, which all claim to offer a more solid, LLM-based alternative to human testing, checklists and existing observability tools. That’s notable because this widens the scope of its potential market — you don’t have to be a developer to use it, and domain experts and executives can evaluate AI apps for inconsistencies, quality and accuracy themselves.
Or read this on TechCrunch