Get the latest tech news

AWS brings prompt routing and caching to its Bedrock LLM service


As businesses move from trying out generative AI in limited prototypes to putting them into production, they are becoming increasingly price conscious.

So basically, you want to create this notion of ‘Hey, at run time, based on the incoming prompt, send the right query to the right model,’” Deo explained. Startups like Martian and a number of open source projects also tackle this, but AWS would likely argue that what differentiates its offering is that the router can intelligently direct queries without a lot of human input. Since those customers are asking the company to support these, AWS is launching a marketplace for these models, where the only major difference is that users will have to provision and manage the capacity of their infrastructure themselves — something that Bedrock typically handles automatically.

Get the Android app

Or read this on TechCrunch

Read more on:

Photo of AWS

AWS

Photo of Bedrock LLM

Bedrock LLM

Photo of Bedrock LLM service

Bedrock LLM service

Related news:

News photo

AWS brings third-party apps to its SageMaker AI platform

News photo

AWS makes its SageMaker HyperPod AI platform more efficient for training LLMs

News photo

AWS says AI could disrupt everything – and hopes it will do just that to Windows