Get the latest tech news

Lambda launches ‘inference-as-a-service’ API claiming lowest costs in AI industry


This GPU advantage enables the platform to support scaling to trillions of tokens monthly, providing flexibility for devs and enterprises.

“Our platform is fully verticalized, meaning we can pass dramatic cost savings to end users compared to other providers like OpenAI,” said Robert Brooks, Lambda’s Vice President of Revenue, in a video call interview with VentureBeat. Lambda’s Inference API supports leading-edge models such as Meta’s Llama 3.1, Nous’s Hermes-3, and Alibaba’s Qwen 2.5, making it one of the most accessible options for the machine learning community. Brooks noted, “Lambda has deployed tens of thousands of GPUs over the past decade, allowing us to offer cost-effective solutions and maximum utility for both older and newer AI chips.”

Get the Android app

Or read this on Venture Beat

Read more on:

Photo of API

API

Photo of service

service

Photo of inference

inference

Related news:

News photo

Upvest, a stock trading API used by N26, Revolut and others, raises $105 million

News photo

The UC Berkeley Project That Is the AI Industry's Obsession

News photo

Training myself to run farther with Strava's API and an IoT dog feeder of M&Ms