Get the latest tech news

SambaNova and Gradio are making high-speed AI accessible to everyone—here’s how it works


SambaNova and Gradio partner to simplify AI development, offering faster inference and improved energy efficiency, challenging Nvidia's dominance in the evolving AI chip market.

SambaNova’s platform can run Meta’s Llama 3.1 405B model at 132 tokens per second at full precision, a speed that is particularly crucial for enterprises looking to deploy AI at scale. As enterprises integrate AI into their operations, they will need to balance speed with sustainability, considering the total cost of ownership, including energy consumption and cooling requirements. Although SambaNova and others offer powerful hardware, Nvidia’s CUDA ecosystem maintains an edge with its wide range of optimized libraries and tools that many AI developers already know well.

Get the Android app

Or read this on Venture Beat

Read more on:

Photo of SambaNova

SambaNova

Photo of Gradio

Gradio

Photo of speed AI

speed AI

Related news:

News photo

Gradio 5 is here: Hugging Face’s newest tool simplifies building AI-powered web apps

News photo

SambaNova challenges OpenAI’s o1 model with Llama 3.1-powered demo on HuggingFace

News photo

SambaNova makes Llama gallop in inference cloud debut