Get the latest tech news

Ironwood: The first Google TPU for the age of inference


We’re introducing Ironwood, our seventh-generation Tensor Processing Unit (TPU) designed to power the age of generative AI inference.

Ironwood delivers the massive parallel processing power necessary for the most demanding AI workloads, such as super large size dense LLM or MoE models with thinking capabilities for training and inference. Google Cloud is the only hyperscaler with more than a decade of experience in delivering AI compute to support cutting edge research, seamlessly integrated into planetary-scale services for billions of users every day with Gmail, Search and more. Leading thinking models like Gemini 2.5 and the Nobel Prize winning AlphaFold all run on TPUs today, and with Ironwood we can’t wait to see what AI breakthroughs are sparked by our own developers and Google Cloud customers when it becomes available later this year.

Get the Android app

Or read this on Hacker News

Read more on:

Photo of Age

Age

Photo of inference

inference

Photo of Ironwood

Ironwood

Related news:

News photo

Google’s new Ironwood chip is 24x more powerful than the world’s fastest supercomputer

News photo

Ironwood is Google’s newest AI accelerator chip

News photo

IBM releases a new mainframe built for the age of AI