Google puts Nvidia on high alert as it showcases Trillium, its rival AI chip, while promising to bring H200 Tensor Core GPUs within days

Trending 2 hours ago
Trillium TPU
(Image credit: Google)

  • Trillium offers 4x training boost, 3x conclusion betterment complete TPU v5e
  • Enhanced HBM and ICI bandwidth for LLM support
  • Scales up to 256 chips per pod, perfect for extended AI tasks

Google Cloud has unleashed its latest TPU, Trillium, nan sixth-generation exemplary successful its civilization AI spot lineup, designed to powerfulness precocious AI workloads.

First announced backmost successful May 2024, Trillium is engineered to grip large-scale training, tuning, and inferencing pinch improved capacity and costs efficiency.

The merchandise forms portion of Google Cloud’s AI Hypercomputer infrastructure, which integrates TPUs, GPUs, and CPUs alongside unfastened package to meet nan expanding demands of generative AI.

A3 Ultra VMs arriving soon

Trillium promises important improvements complete its predecessor, TPU v5e, pinch complete a 4x boost successful training capacity and up to a 3x summation successful conclusion throughput. Trillium delivers doubly nan HBM capacity and doubled Interchip Interconnect (ICI) bandwidth, making it peculiarly suited to ample connection models for illustration Gemma 2 and Llama, arsenic good arsenic compute-heavy conclusion applications, including diffusion models specified arsenic Stable Diffusion XL.

Google is keen to accent Trillium’s attraction connected power ratio arsenic well, pinch a claimed 67% summation compared to erstwhile generations.

Google says its caller TPU has demonstrated substantially improved capacity successful benchmark testing, delivering a 4x summation successful training speeds for models specified arsenic Gemma 2-27b and Llama2-70B. For conclusion tasks, Trillium achieved 3x greater throughput than TPU v5e, peculiarly excelling successful models that request extended computational resources.

Scaling is different spot of Trillium, according to Google. The TPU tin nexus up to 256 chips successful a single, high-bandwidth pod, expandable to thousands of chips wrong Google’s Jupiter information halfway network, providing near-linear scaling for extended AI training tasks. With Multislice software, Trillium maintains accordant capacity crossed hundreds of pods.

Sign up to nan TechRadar Pro newsletter to get each nan apical news, opinion, features and guidance your business needs to succeed!

Tied successful pinch nan presence of Trillium, Google besides announced nan A3 Ultra VMs featuring Nvidia H200 Tensor Core GPUs. Scheduled for preview this period they will connection Google Cloud customers a high-performance GPU action wrong nan tech giant’s AI infrastructure.

Trillium TPU, built to powerfulness nan early of AI - YouTube Trillium TPU, built to powerfulness nan early of AI - YouTube

Watch On

You mightiness besides like

  • Google Cloud: No-one tin present business AI worth for illustration us
  • Google's TPU v5p spot is faster and has much representation and bandwidth
  • Intel and Google Cloud squad up to motorboat super-secure VMs

Wayne Williams is simply a freelancer penning news for TechRadar Pro. He has been penning astir computers, technology, and nan web for 30 years. In that clip he wrote for astir of nan UK’s PC magazines, and launched, edited and published a number of them too.

More
Source Technology
Technology