AMD's MI300: The Future of Semiconductors!

Deep-dive into the technology of AMD's MI300

Estimated read time: 1:20

    AI is evolving every day. Don't fall behind.

    Join 50,000+ readers learning how to use AI in just 5 minutes daily.

    Completely free, unsubscribe at any time.

    Summary

    AMD's MI300 is a groundbreaking leap in the world of semiconductors, incorporating five key technologies crucial for competitive chip design. The MI300 melds chiplets, advanced packaging, SoC designs, unified memory, and AI acceleration into one powerful product. It features an astounding 147 billion transistors, integrating 10,000 CDNA3 GPU cores with 24 Zen 4 CPU cores. This innovative chip is aimed at advancing efficiency and performance, particularly in achieving future zettascale computing capabilities. With the MI300, AMD sets the stage for future developments and challenges the dominance of competitors like Nvidia.

      Highlights

      • AMD MI300 integrates revolutionary technologies for the next-gen semiconductor. 🤖
      • The extensive use of chiplets allows flexible and efficient design beyond traditional monolithic chips. 🌌
      • Advanced packaging with true 3D stacking makes efficient power and data management a reality. 🏗️
      • Unified memory architecture simplifies programming and enhances system-wide efficiency. 🧠
      • MI300's breakthrough in AI acceleration positions AMD as a strong competitor to Nvidia. 🏎️

      Key Takeaways

      • AMD MI300 combines five crucial technologies: chiplets, advanced packaging, SoC design, unified memory, and AI acceleration. 🚀
      • It's a powerhouse with 147 billion transistors and integrates 10,000 CDNA3 GPU cores and 24 Zen 4 CPU cores. 💪
      • The MI300 is a fully functioning APU, breaking new ground in high-performance computing efficiency. 🌟
      • Aims at challenging Nvidia and driving zettascale computing advancement through enhanced efficiency. ⚡
      • It represents AMD's first step towards achieving integrated systems aimed at extreme performance and efficiency. 🌐

      Overview

      AMD's MI300 is nothing short of a marvel in the tech world, marrying five game-changing technologies into a single semiconductor chip. Each one of these technologies—chiplets, advanced packaging, SoC (System on Chip) design, unified memory, and AI acceleration—significantly enhances the performance and efficiency of the MI300, setting a new standard for high-performance computing and AI applications.

        This technological symphony results in a chip boasting an overwhelming 147 billion transistors, equipped with over 10,000 CDNA3 GPU cores and 24 Zen 4 CPU cores. It's designed not just to perform but to redefine performance benchmarks, posing a serious challenge to industry giants like Nvidia. AMD's MI300 isn't just about raw power; it's about delivering power smartly, optimizing efficiency to work towards the colossal aim of zettascale computing.

          What makes MI300 truly revolutionary is its design approach. By embracing a fully integrated APU that merges CPUs and GPUs into one harmonious unit, it breaks away from traditional concepts and opens new frontiers in efficiency and performance. This chip is AMD’s bold step towards a future where zettascale isn't a dream, it's an upcoming reality, driving everything from supercomputers to comprehensive AI solutions.

            Deep-dive into the technology of AMD's MI300 Transcription

            • 00:00 - 00:30 The semiconductor landscape is rapidly  changing. Over the past couple of years, five   new technologies have surfaced, which I think will  become fundamental requirements for anyone trying   to design and produce competitive silicon chips in  the future. Each single of these technologies on   its own is already very impactful, but combining  all five of them will be the ultimate challenge.   And that's exactly what AMD is trying to do with  its next-gen HPC and AI accelerator MI300. Let's   take a closer look at what these five important  technologies are, how AMD is using them for
            • 00:30 - 01:00 MI300 and what all of this has to do with the  race to the first zettascale supercomputer.   Until now the design of silicon chips  primarily revolved around two key aspects:   architecture and process node. However, modern  chips have evolved to become much more than just   the sum of these two elements. Architecture and  process node have expanded to encompass broader,   overarching concepts. This shift has given rise  to next generation technologies. Number one won't
            • 01:00 - 01:30 be a surprise for you, if you have watched  this channel before, it's chiplets. Instead   of a single monolithic chip, almost all future  semiconductors will be chiplet based. The benefits   of chiplet designs are plentiful and range from  simple cost savings, due to increased yield,   the ability to use the perfect process node for  each individual chiplet, all the way to scaling   up an entire product stack based on only a few  tapouts. There's no future without them. Number   two is advanced packaging and 3D stacking. The  shift towards chiplets has made it essential to
            • 01:30 - 02:00 closely integrate each individual chiplet into the  final product. While the process node was once the   primary focus, packaging is increasingly gaining  prominence. In the future I expect that packaging   will surpass the importance of the process node,  reflecting a fundamental change in chip design   priorities. Number three is the transition  towards semiconductors that are primarily   SoCs or APUs. These chips consolidate functions  that were previously separated into CPU, GPU and
            • 02:00 - 02:30 I/O components. Apple has been a pioneer in this  area and many current CPUs essentially function   as SoCs. This development also relies heavily  on the chiplet approach, with each building   block of the SoC potentially serving as its own  distinct chiplet. Number four is the adoption   of a unified memory architecture. As systems  become more tightly integrated, the importance   of the memory architecture growths. Unified memory  enables all components of the SoC to access the
            • 02:30 - 03:00 same memory pool, resulting in more efficient  memory allocation and simplified programming,   since every part of the system can access the  same data. We have to move away from the idea that   each component needs its own dedicated memory.  And last, but certainly not least, we have AI.   While you could argue it's already part of a SoC  design, for me the integration of AI and machine   learning acceleration is such an important aspect  of any future semiconductor, that it has its own   dedicated point. And as we will learn later,  AI also pays a very important role when trying
            • 03:00 - 03:30 to achieve zettascale. By the way, all the logos  I'm using here are AI generated. In a nutshell,   future semiconductors will be chiplet based,  rely on advanced packaging, function as an SoC   with unified memory and accelerate AI and machine  learning code. AMD's MI300 is such an interesting   product because it's the first of its kind to  incorporate all five technologies. Before we dive   into each of these technologies and examine how  the MI300 implements them, let's briefly review
            • 03:30 - 04:00 the specifications. AMD's Instinct accelerators  are designed for supercomputers and as such   MI300 is targeting the HPC and AI market. Its  predecessor, MI250xX, powers Frontier, the world's   first exascale supercomputer. I've talked about  very large chips before, but compared to MI300 and   it's 147 billion transistors all other chips seem  tiny. It's truly a monster of a chip and easily   outclasses Intels 100 billion transistor Ponte  Vecchio or Nvidia's 80 billion Hopper GPU. These
            • 04:00 - 04:30 numbers are a clear indication of AMD's mmbitions  for MI300 and while we don't have a official die   size yet MI300 will use well over 1000 millimeters  squared of active silicon. In total MI300 combines   more than 10,000 CDNA3 GPU cores with 24 Zen 4  CPU cores, lots of I/O, lots of cache and 128   gigabytes of HBM3 memory. The whole chip is  manufactured in a combination of TSMC N6 and
            • 04:30 - 05:00 N5 process nodes, which brings us right to the  first of the five technologies: chiplets. Just   by looking at pictures of MI300 the complexity  of its chiplet design isn't really apparent.   It doesn't look that different to its predecessor  MI200 with what seems like four medium-sized dies   in the middle, surrounded by a couple of smaller  chiplets, most likely the HBM3 memory. But this   appearance is very deceiving, because what we are  looking at is the absolutely crazy combination of   a large interposer with four six nanometer  base layer dies on top, on top of which AMD
            • 05:00 - 05:30 stacked another nine, yes nine, five nanometer  chiplets. To round it off the interposer also   connects eight 16 gigabyte stacked HBM3 memory  modules and another 8 smaller unidentified chips,   most likely spacers used for physical integrity.  Io understand the complexity of this layout let's   take a look at this mock-up AMD provided. The  foundation of MI300 is the interposer, it connects   the HBM3 memory and the base chiplets. The marked  area in yellow might not be the exact size,
            • 05:30 - 06:00 but it's good enough to understand the concept.  On top of the interposer we have 8 stacked HBM3   memory chips, each adding 16 gigabytes for a  total of 128 gigabytes of memory. On pictures   of the physical chip we can see smaller chips  sitting in between the HBM stacks that aren't   present in this illustration, which means they  are most likely inert silicon used as spacers   and don't have any active functionality. Next are  the four six nanometer base chips marked in pink.   They sit on top of the interposer and underneath  the five nanometer compute chiplets,.sSnce the
            • 06:00 - 06:30 base chiplets are active silicon that most  likely houses I/O functionality and cache,   AMD is using true 3D stacking. More on packaging  and 3D stacking in just a bit. And finally,   on top of the four base chiplets, we have six  five nanometer CDNA3 MCDs (*GCDs) containing   the GPU cores and three five nanometers Zen 4  CCDs with the CPU cores. Each CCD holds 8 cores   for a total of 24 Zen 4 cores on MI300. I think  this overview is the best way to understand how
            • 06:30 - 07:00 AMD counts four 6 nanometer and nine 5 nanometer  chiplets, and the interesting part is that while   it's a complicated layout from a packet point of  view, it's a very simple design from a chiplet   perspective. All AMD has to do is to tape out  one six nanometer cache and I/O base chiplet,   one five nanometer CDNA3 MCD (*GCD) chiplet and  use the already existing 5 nanometer 8 cores Zen   4 chiplet. It's literally just three different  chiplets and only two new individual tape outs
            • 07:00 - 07:30 for MI300. All these chiplets are rather small  and not too complex in design. We know how tiny a   singles Zen 4 CCD is and while a CDNA3 MCD (*GCD)  will be slightly bigger, at this size and we can   still expect close to perfect yields. It's a mix  and match system that allows AMD to effortlessly   scale their HPC and AI products. You could also  match nine Zen 4 CCDs with two CDNA3 MCDs (*GCDs)   or basically any other configuration.  Depending on what the customer wants,
            • 07:30 - 08:00 AMD always delivers the perfect product. And if  you are building a supercomputer ,semi-custom   silicon is not uncommon. We have seen chiplet  based architectures since the release of Zen 2   and AMDs's X3D CPUs up a notch in verticality.  But nothing comes close to MI300! This chip is   what AMD's R&D has been working towards for the  past decade, it's basically science fiction,   the end goal: a truly modular chiplet design.  And it's not some far away product either,   mMI300 already exists in silicon and will launch  in the second half of this year. I think it's
            • 08:00 - 08:30 impossible to overstate the importance of a design  like this! What if I told you that I didn't study   anything related to semiconductors or computer  science and everything I know is based on many   years of just following my passion and trying to  absorb as much knowledge as I can? That's why I'm   really excited that brilliant.org is sponsoring  this video! The cool thing about Brilliant is   that it doesn't feel like working, more like  exploring new topics with its intuitive and
            • 08:30 - 09:00 interactive concepts. Brilliant offers thousands  of lessons related to math and computer science,   from beginners to advanced levels. Just recently  they added a whole new introductory course on   artificial intelligence, which really helps you  understand how neural networks function. AI stops   being just "magic" and turns into this tangible  thing that you now understand. The best way to   reach your goals and increase your knowledge about  the things you are interested in is to never stop   learning. Visit brilliant.org/HighYield for a free  30-day trial plus the first 200 of you also get 20
            • 09:00 - 09:30 off Brilliants annual premium subscription. If you  are curious and enjoy my videos I think you will   love Brilliant! With that, let's get back to MI300  and its advanced packaging technology. If chiplets   are the brain of MI300, the advanced packaging has  has to be the heart. It distributes all the energy   and data and is a equally important part of the  chip. Without proper packaging MI300 would be just   a collection of useless chiplets. For MI300 AMD  uses a mix of 2.5D and 3D stacking. The easiest
            • 09:30 - 10:00 way to explain the difference between 2.5 and 3D  stacking is that while in both cases a silicon   chip sits on top of another silicon chip, only  with 3D stacking both of these ships are actually   active. The 5 nanometer MCD and CCD chiplets on  top of a 6 nanometer I/O and cache based chiplets   are true 3D stacking, while the HBM memory on  top of the interposer is 2.5D. It's vertical
            • 10:00 - 10:30 but the interposer is only used for data and power  routing. A really interesting aspect are the five   nanometer compute chiplets which sit on top of the  infinity cache, located in the base layer, while   AMD's 3D V-Cache has it the other way around. On  a 7950X3D the cache chiplet sits on top of the   CPU chiplet, creating rather enter interesting  engineering problems, something I extensively   talked about in my previous video. MI300 proves  that AMD has the technology to implement a reverse   stacking method. Currently we can only speculate  on the specific 3D stacking technology AMD and
            • 10:30 - 11:00 TSMC are using to enable MI300. The first thing  that comes to mind is the same TSMC SoIC and TSV   based technology as used in Zen 4 X3D, where the  base chipletwould use TSVs to directly wire into   the compute chiplet above. The only question is if  this method provides enough bandwidth and if it's   the most efficient and cost effective solution.  Another way would be using the large interposer   to route the data and power connections. The idea  here is that the six nanometer base chiplets might
            • 11:00 - 11:30 actually be a bit smaller than the 5 nanometer  compute chiplets stacked on top. This approach   would allow for direct data and power routing,  without having to use TSVs to go through the   chiplet itself. Twitter user AMDFanUwe created  some very interesting drawings which illustrate   this concept very well. Such a approach would be  less complex and most likely cheaper to implement.   The challenge is to create a stacking method that  fulfills the bandwidth needs, is low latency,   doesn't complicate the power routing and at the  same time isn't too expensive. As you can see,
            • 11:30 - 12:00 a super easy task. I'm hoping AMD will share more  information on the exact technology used sooner   than later. Let me know in the comments down below  what technology you think AMD is using for MI300   and how you would approach the 3D stacking of  MI300. Next let's talk about another first in   the HPC space. MI300 isn't a simple GPU like  its predecessors, but it's a fully featured   APU combining CPU and CPU cores into one single  package, and with the included I/O functionality
            • 12:00 - 12:30 of the base chiplets is actually a true SoC. We  have seen the rise of SoCs in mobile and desktop,   but high performance computing is a entirely new  frontier. If we talk about power draw in modern   computers, most of us probably think of power  hungry GPUs and high-speed CPUs, but data connects   are becoming another huge factor for power  consumption. As chips become faster, interconnects   need to keep up. Transporting large amounts of  data at a high speeds between physically separated
            • 12:30 - 13:00 chips like CPU and GPU consumes a lot of power. By  reducing the physical distance you not only reduce   latency, but it actually virtually eliminates  the energy requirements for data transfers.   When I first heard about APU and SoC concepts, I  primarily saw it as space optimization, but now I   understand that the main benefit is the increase  in efficiency. Especially supercomputers use a   lot of their power budget for data connection,  the less data you have to move off silicon the   better. With MI300s APU approach not only will  the interconnect power usage decrease drastically,
            • 13:00 - 13:30 but motherboards also don't need to be as complex.  Two huge advantages in the server space! A single   MI300 is already a fully functioning chip in of  itself. Somewhat connected to the SoC design and   another very impactful technology is is the use  of unified memory. Apple's M-chips are already   a great example how unified memory benefits the  entire system. There are two reasons why unified   memory is beneficial: first is the software.  If all parts of your system have access to the
            • 13:30 - 14:00 same data, because they use the same memory pool,  programming becomes easier. The other reason is   hardware related: having unified memory eliminates  redundant memory copies, which also means you   need less total physical memory. And as you can  guess, on-package memory is more efficient too.   In a nutshell, with unified memory you need less  memory capacity, it's on package and thus power   efficient, it reduces mainboard complexity and  simplifies programming. A well-implemented unified
            • 14:00 - 14:30 memory architecture has many advantage over the  legacy approach, but of course you need to be able   to build such a system with the proper GPU and  CPU IP. And currently only AMD is capable of that,   although Nvidia's Grace Superchip isn't far  behind, something we will look at in a future   video. On the topic of Nvidia, let's talk about AI  acceleration. Artificial intelligence and machine   learning are currently the number one topic  in tech and there is only one company that's   clearly leading a pack: Nvidia. With MI300 AMD  is trying its best to catch up. CDNA3 focuses on
            • 14:30 - 15:00 machine learning acceleration, AMD claims up to 8  times the AI training performance and 5 times the   AI power efficiency. MI300 will also support new  math formats, most likely low precision INT and   floating point operations to further accelerate  deep neural networks. This won't be enough to   surpass Nvidia in my opinion, especially since  Nvidia's advantage is deeply rooted in its CUDA   software package, something AMD has to invest a  lot of time and money into if they really want to   challenge Nvidia. MI300 isn't only for AI, there  are plenty of other applications within the HPC
            • 15:00 - 15:30 space it will excel at, but the overall direction  of the design is clear. AMD's goal is to challenge   Nvidia and if MI300 won't bring victory, MI400  might. Now that we have talked about all five   technologies AMD implemented with MI300, let me  tell you why AMD is so focused on integrating them   into a single product. It was only a few weeks  ago, while watching an ISSCC presentation from   AMD CEO Lisa Su, that I finally understood AMD's  goal. There's actually a bigger picture behind
            • 15:30 - 16:00 all of this: achieving Zettascale performance.  Not with MI300, but with one of its hopefully   many successors. I highly recommend you watch  the keynote yourself, but here's a gist of it:   going from exascale to zettascale is not a  performance problem, eventually we will get there,   it's a efficiency problem. Currently supercomputer  efficiency roughly doubles every 2.2 years,   which doesn't sound too bad, does it? But if  we continue this trend and we finally are able
            • 16:00 - 16:30 to build a zettascale supercomputer in the  mid-2030s, even if we achieve the same 2x   efficiency increase every two years until then,  a single zettascale supercomputer would consume   about 500 megawatts. That's a lot, that's nuclear  power plant levels of energy draw! Cooling such   a system alone would be impossible! That's  why, if we want to enter the zettascale age,   we need to find a way to drastically increase  efficiency, well above the current levels. And   Mi 300 is exactly that, it's AMD's first  step towards a fully integrated system,
            • 16:30 - 17:00 with a focus on not only increasing performance  but specifically using every technology available   to be as efficient as possible, The chiplets, the  advanced packaging, the SoC design, the unified   memory and the focus on AI all target efficiency.  You could get a similar performance with less   engineering expense, with less R&D and with a less  complex layout, but the goal is not to create the   cheapest or most convenient product, the goal  is a first step towards continuous efficiency
            • 17:00 - 17:30 scaling. Reaching zettascale will require a  lot more R&D over the years, it will require   technology that's not available yet. Off-chip  communication will be optical in the future,   huge AI networks will exponentially increase  efficiency for specific calculations and 3D   stacking will get even more crazy. But eventually  we will get there. Now it's your turn! Let me   know your thoughts about the design of MI300 and  to be honest, I'm even more interested how you   think we can reach zettascale in the future. What  technologies will stay and which ones will fail?
            • 17:30 - 18:00 Leave a comment down below and let's see if we can  correctly predict the future of semiconductors.   You know what to do if you found this video  interesting and see you in the next one :)