Start your day with intelligence. Get The OODA Daily Pulse.
Google Cloud unveiled its seventh-generation Tensor Processing Unit (TPU), Ironwood, on Wednesday. This custom AI accelerator, the company claims, delivers more than 24 times the computing power of the world’s fastest supercomputer when deployed at scale. The new chip, announced at Google Cloud Next ’25, represents a significant pivot in Google’s decade-long AI chip development strategy. While previous generations of TPUs were designed primarily for both training and inference workloads, Ironwood is the first purpose-built specifically for inference — the process of deploying trained AI models to make predictions or generate responses. “Ironwood is built to support this next phase of generative AI and its tremendous computational and communication requirements,” said Amin Vahdat, Google’s Vice President and General Manager of ML, Systems, and Cloud AI, in a virtual press conference ahead of the event. “This is what we call the ‘age of inference’ where AI agents will proactively retrieve and generate data to collaboratively deliver insights and answers, not just data.”
Full report : Google launches new Ironwood chip to speed AI applications which it says is fastest in the world.