Get the latest tech news

Hugging Face offers inference as a service powered by Nvidia NIM


Hugging Face is offering developers an inference-as-a-service powered by Nvidia NIM microservices.

The new service will bring up to five times better token efficiency with popular AI models to millions ofdevelopers and enables immediate access to NIM microservices running on Nvidia DGX Cloud. Announced today at the Siggraph conference, the service will help developers quickly prototype with open-source AI models hosted on the Hugging Face Hub and deploy them in production. The Nvidia DGX Cloud platform is purpose-built for generative AI, offering developers easy access to reliable accelerated computing infrastructure that can help them bring production-ready applications to market faster.

Get the Android app

Or read this on Venture Beat

Read more on:

Photo of service

service

Photo of inference

inference

Photo of Hugging Face

Hugging Face

Related news:

News photo

From surveillance to denial-of-service: How the DHS's new NEO robot dog will change raids | The robot takes out IoT devices

News photo

Sony's next live-service sci-fi shooter Concord won't have a battle pass

News photo

OpenAI is set to lose $5B this year