21.5 C
New York
Tuesday, May 14, 2024

Sundar Pichai on Gemini, AI progress and extra


Infrastructure for the AI period: Introducing Trillium

Coaching state-of-the-art fashions requires a number of computing energy. Business demand for ML compute has grown by an element of 1 million within the final six years. And yearly, it will increase tenfold.

Google was constructed for this. For 25 years, we’ve invested in world-class technical infrastructure. From the cutting-edge {hardware} that powers Search, to our customized tensor processing models that energy our AI advances.

Gemini was skilled and served fully on our fourth and fifth era TPUs. And different main AI firms, together with Anthropic, have skilled their fashions on TPUs as properly.

At present, we’re excited to announce our sixth era of TPUs, known as Trillium. Trillium is our most performant and most effective TPU so far, delivering a 4.7x enchancment in compute efficiency per chip over the earlier era, TPU v5e.

We’ll make Trillium obtainable to our Cloud clients in late 2024.

Alongside our TPUs, we’re proud to supply CPUs and GPUs to assist any workload. That features the brand new Axion processors we introduced final month, our first customized Arm-based CPU that delivers industry-leading efficiency and vitality effectivity.

We’re additionally proud to be one of many first Cloud suppliers to supply Nvidia’s cutting-edge Blackwell GPUs, obtainable in early 2025. We’re lucky to have a longstanding partnership with NVIDIA, and are excited to convey Blackwell’s breakthrough capabilities to our clients.

Chips are a foundational a part of our built-in end-to-end system. From performance-optimized {hardware} and open software program to versatile consumption fashions. This all comes collectively in our AI Hypercomputer, a groundbreaking supercomputer structure.

Companies and builders are utilizing it to sort out extra advanced challenges, with greater than twice the effectivity relative to only shopping for the uncooked {hardware} and chips. Our AI Hypercomputer developments are made potential partially due to our method to liquid cooling in our knowledge facilities.

We’ve been doing this for practically a decade, lengthy earlier than it grew to become state-of-the-art for the {industry}. And at the moment our complete deployed fleet capability for liquid cooling programs is almost 1 gigawatt and rising — that’s near 70 instances the capability of another fleet.

Underlying that is the sheer scale of our community, which connects our infrastructure globally. Our community spans greater than 2 million miles of terrestrial and subsea fiber: over 10 instances (!) the attain of the subsequent main cloud supplier.

We are going to maintain making the investments essential to advance AI innovation and ship state-of-the-art capabilities.



Supply hyperlink

Related Articles

LEAVE A REPLY

Please enter your comment!
Please enter your name here

Latest Articles