Get the latest tech news
Lambda launches ‘inference-as-a-service’ API claiming lowest costs in AI industry
This GPU advantage enables the platform to support scaling to trillions of tokens monthly, providing flexibility for devs and enterprises.
“Our platform is fully verticalized, meaning we can pass dramatic cost savings to end users compared to other providers like OpenAI,” said Robert Brooks, Lambda’s Vice President of Revenue, in a video call interview with VentureBeat. Lambda’s Inference API supports leading-edge models such as Meta’s Llama 3.1, Nous’s Hermes-3, and Alibaba’s Qwen 2.5, making it one of the most accessible options for the machine learning community. Brooks noted, “Lambda has deployed tens of thousands of GPUs over the past decade, allowing us to offer cost-effective solutions and maximum utility for both older and newer AI chips.”
Or read this on Venture Beat