Google’s new Trillium AI chip offers four times the speed and supports Gemini 2.0

Google’s new Trillium AI chip offers four times the speed and supports Gemini 2.0


Subscribe to our daily and weekly newsletters to receive the latest updates and exclusive content on industry-leading AI reporting. Learn more


Google just unveiled Trillium, its sixth-generation artificial intelligence accelerator chip, promising performance improvements that could fundamentally change the economics of AI development while pushing the boundaries of what’s possible in machine learning.

The custom processor powering the training of Google’s newly announced Gemini 2.0 AI model delivers four times the training performance of its predecessor while using significantly less energy. This breakthrough comes at a crucial time as technology companies race to develop increasingly sophisticated AI systems that require massive computing resources.

“TPUs support 100% of Gemini 2.0 training and inference,” Google CEO Sundar Pichai said in an announcement post highlighting the chip’s central role in the company’s AI strategy. The scale of the deployment is unprecedented: Google connected more than 100,000 Trillium chips in a single network structure, creating one of the most powerful AI supercomputers in the world.

How Trillium’s 4x performance improvement is transforming AI development

Trillium’s specifications represent significant advances in several dimensions. The chip delivers a 4.7x increase in peak processing power per chip compared to its predecessor, while doubling both high-bandwidth storage capacity and inter-chip interconnect bandwidth. Perhaps most importantly, energy efficiency is increased by 67% – a crucial metric as data centers struggle with the enormous energy demands of AI training.

“When training the Llama-2-70B model, our testing shows that Trillium achieves near-linear scaling from a 4-slice Trillium 256 chip pod to a 36-slice Trillium 256 chip pod with scaling efficiency of 99%,” said Mark Lohmeyer, vice president of computing and AI infrastructure at Google Cloud. This level of scaling efficiency is particularly notable given the challenges typically associated with distributed computing at this scale.

The Economics of Innovation: Why Trillium is changing the game for AI startups

Trillium’s business impact goes beyond just performance metrics. Google claims the chip offers up to 2.5x improvement in training performance per dollar compared to the previous generation, potentially reshaping the economics of AI development.

This cost-effectiveness could prove particularly important for companies and startups developing large language models. AI21 Labs, an early Trillium customer, has already reported significant improvements. “The advances in scale, speed and cost-effectiveness are significant,” noted Barak Lenz, CTO of AI21 Labs, in the announcement.

Reaching new heights: Google’s 100,000-chip AI supernetwork

Google’s use of Trillium within its AI hypercomputer architecture demonstrates the company’s integrated approach to AI infrastructure. The system combines over 100,000 Trillium chips with a Jupiter network structure that enables a bisection bandwidth of 13 petabits per second – allowing a single distributed training job to scale across hundreds of thousands of accelerators.

“The growth in Flash usage has exceeded 900%, which is incredible to see,” noted Logan Kilpatrick, product manager on Google’s AI Studio team, during the developer conference, citing the rapidly increasing demand for AI computing Resources.

Beyond Nvidia: Google’s bold move in the AI ​​chip war

Trillium’s release intensifies competition in AI hardware, where Nvidia has dominated with its GPU-based solutions. While Nvidia’s chips remain the industry standard for many AI applications, Google’s tailored silicon approach could offer advantages for certain workloads, particularly when training very large models.

Industry analysts suggest that Google’s massive investments in developing customized chips reflect a strategic bet on the growing importance of AI infrastructure. The company’s decision to make Trillium available to cloud customers shows its desire to compete more aggressively in the cloud AI market, where it faces stiff competition from Microsoft Azure and Amazon Web Services.

Driving the future: What Trillium means for the AI ​​of tomorrow

The impact of Trillium’s abilities goes beyond immediate performance increases. The chip’s ability to efficiently handle mixed workloads – from training large models to running inference for production applications – suggests a future in which AI computing becomes more accessible and cost-effective.

For the broader tech industry, Trillium’s release signals that the race for dominance in AI hardware is entering a new phase. As companies push the boundaries of what is possible with artificial intelligence, the ability to design and deploy specialized hardware at scale could become an increasingly crucial competitive advantage.

“We are still in the early stages of what is possible with AI,” wrote Demis Hassabis, CEO of Google DeepMind, in the company’s blog post. “The right infrastructure – both hardware and software – will be critical as we continue to push the boundaries of what AI can do.”

As the industry moves toward more sophisticated AI models that can act autonomously and reason across multiple information modes, the demands on the underlying hardware will only increase. With Trillium, Google has shown that it intends to remain at the forefront of this evolution and invest in the infrastructure that will power the next generation of AI advancements.

Leave a Reply

Your email address will not be published. Required fields are marked *