COMPUTERS

The AMD Advancing AI & Instinct MI300 Launch Live Blog (Starts at 10am PT/18:00 UTC)

AnandTech Live Blog: The newest updates are at the top. This page will auto-update, there’s no need to manually refresh your browser.

01:16PM EST – And MS will be offering MI300X Azure instances

01:15PM EST – Microsoft and AMD have been building the foundation for several years here

01:14PM EST – Lisa is asking Kevin for his thoughts on where the industry is on this AI journey

01:13PM EST – MS CTO, Kevin Scott

01:12PM EST – And now AMD’s first guest of many, Microsoft

01:12PM EST – Bloom 176B (throughput) and Llama 2 70B (latency) inference performance.

01:11PM EST – Comparing a single 8 accelerator server

01:11PM EST – And how does MI300X scale?

01:10PM EST – AMD finds they have the performance advantage in FlashAttention-2 and Llama 2 70B. At the kernel level in TFLOPS

01:10PM EST – AMD has the advantage in memory capacity and bandwidth due to having more HBM stacks. And they think that’s going to help carry them to victory over H100

01:09PM EST – And immediately jumping to the H100 comparisons

01:09PM EST – And 8 stacks of HBM3 attached to the IODs, for 192GB of memory, 5.3 TB/second of bandwidth

01:08PM EST – Wired to the IODs via TSVs

01:08PM EST – 304 CDNA 3 compute units

01:08PM EST – 8 XCD compute dies stacked on top

01:08PM EST – 256MB AMD Infinity Cache, Infinity Fabric Support, etc

01:08PM EST – 4 I/O Dies in the base layer

01:08PM EST – A dozen 5nm/6nm chiplets

01:07PM EST – 153B transistors for MI300X

01:07PM EST – 3.4x more perf for BF16, 6.8x INT8 perf, 1.6x memory bandwidth

01:07PM EST – CDNA 3 comes wiht a new compute engine, sparsity support, industry-leading memory bandwidth and capacity, etc

01:06PM EST – “Highest performance accelerator in the world for generative AI”

01:06PM EST – Launching today: AMD Instinct MI300X accelerator

01:06PM EST – The more compute, the better the model, the faster the answers

01:06PM EST – Generative AI requires tens of thousands of accelerators at the high-end

01:05PM EST – Now to products, starting with the cloud

01:05PM EST – (AMD has historically struggled with software in particular)

01:05PM EST – A broad hardware portfolio, an open and proven software ecosystem, and partnerships to co-innovate with

01:04PM EST – AMD’s AI strategy is centered around 3 big strategic priorities

01:04PM EST – A greater than 70% compound annual growth rate

01:04PM EST – Now they think it’s going to be $400B+ by 2027

01:03PM EST – In 2023 AMD projected the CAGR for the AI market would be $350B by 2027

01:03PM EST – (Which NVIDIA has captured the lion’s share of thus far)

01:03PM EST – And the key to it? Generative AI. Which requires significant investments in infrastructure

01:02PM EST – Lisa’s listing off some of the use cases for AI

01:02PM EST – And with a rather quick adoption rate, despite being at the very beginning of the AI era

01:02PM EST – AMD views AI as the single most transformative technology in the last 50 years

01:02PM EST – It’s only been just a bit over a year since ChatGPT was launched. And it’s turned the computing industry on its head rather quickly

01:01PM EST – And Lisa is diving right in

01:01PM EST – Today “is all about AI”

01:01PM EST – And here’s AMD’s CEO, Dr. Lisa Su

01:00PM EST – Advancing AI… together

01:00PM EST – (And joining me on this morning’s live blog is the always-awesome Gavin Bonshor)

01:00PM EST – Starting with an opening trailer

12:59PM EST – And hey, here we go. Right on time

12:59PM EST – We’re supposed to start at 10am local time here – so in another minute or so

12:59PM EST – I’d say the event is being held in AMD’s backyard, but since AMD sold their campus here in the bay area several years ago, this is more like NVIDIA’s backyard. Which is fitting, given that AMD is looking to capture a piece of the highly profitable Generative AI market from NVIDIA

12:58PM EST – Including MI300A, their first chiplet-based server APU, and MI300X, their stab at the most powerful GPU/accelerator possible for the AI market

12:57PM EST – Today AMD is making the eagerly anticipated launch of their next-generation MI300 series of accelerators

12:56PM EST – We’re here in San Jose for AMD’s final and most important launch event of the year: Advancing AI

This morning is an important one for AMD – perhaps the most important of the year. After almost a year and a half of build-up, and even longer for actual development, AMD is launching their next generation GPU/APU/AI accelerator family, the Instinct MI300 series. Based on AMD’s new CDNA 3 architecture, and combining it with AMD’s proven Zen 4 cores, AMD will be making a full-court press for the high-end GPU and accelerator market with their new product, aiming to lead in both big-metal HPC as well as the burgeoning market for generative AI training and inference.

Taking the stage for AMD’s launch event will be AMD CEO Dr. LIsa Su, as well as a numerous AMD executives and ecosystem partners, to detail, at last, AMD’s latest generation GPU architecture, and the many forms it will come in. With both the MI300X accelerator and MI300A APU, AMD is aiming to cover most of the accelerator market, whether clients just need a powerful GPU or a tightly-coupled GPU/CPU pairing.

The stakes for today’s announcement are significant. The market for generative AI is all but hardware constrained at the moment, much to the benefit of (and profits for) AMD’s rival NVIDIA. So AMD is hoping to capitalize on this moment to cut off a piece – perhaps a very big piece – of the market for generative AI accelerators. AMD has made breaking into the server space their highest priority over the last half-decade, and now, they believe, is their time to take a big piece of the server GPU market.

Join us here at 10am Pacific/1pm Eastern/18:00 UTC for our live blog coverage of AMD’s event.


Source link

Related Articles

Back to top button