Get the latest tech news
Hugging Face offers inference as a service powered by Nvidia NIM
Hugging Face is offering developers an inference-as-a-service powered by Nvidia NIM microservices.
The new service will bring up to five times better token efficiency with popular AI models to millions ofdevelopers and enables immediate access to NIM microservices running on Nvidia DGX Cloud. Announced today at the Siggraph conference, the service will help developers quickly prototype with open-source AI models hosted on the Hugging Face Hub and deploy them in production. The Nvidia DGX Cloud platform is purpose-built for generative AI, offering developers easy access to reliable accelerated computing infrastructure that can help them bring production-ready applications to market faster.
Or read this on Venture Beat