sekar nallalu AMD,AMD:CA,Cryptocurrency,SA Transcripts Advanced Micro Devices, Inc. (AMD) Deutsche Bank Technology Conference (Transcript)

Advanced Micro Devices, Inc. (AMD) Deutsche Bank Technology Conference (Transcript)

0 Comments

Advanced Micro Devices, Inc. (NASDAQ:AMD) Deutsche Bank Technology Conference August 28, 2024 4:15 PM ET Company Participants Mark Papermaster – Executive Vice President & Chief Technology Officer Conference Call Participants Ross Seymore – Deutsche Bank Ross Seymore I’m Ross Seymore, cover semiconductors here at Deutsche Bank. We’re very pleased to have Mark Papermaster, the CTO of AMD on stage with us today. Mark, thank you very much for joining here at the DB Tech Conference. So I think it was just what last week you guys announced an AI related acquisition. AI is obviously a topic we’ll talk about in many different ways, but the ZT Systems acquisition, talk a little bit about the logic of that deal and seems like there’s some nuance of what you were really going after in that with the engineers. So just talk a little bit about that. Mark Papermaster Absolutely. Well, first, Ross, thanks for having me here at your conference. Excited to be here, and an incredibly exciting time at AMD. When you think about the announcement that we made last week in acquiring ZT Systems, it’s the next stage in really our strategic growth of ensuring that we have the full complement of skills that we need to have not only the best AI hardware, not only competitive and leadership AI software, but the ability to integrate it and optimize it at the system level. And so, ZT Systems represents exactly that. They’re 15 years of experience with building some of the most complex heterogeneous system designs, integrating CPUs, GPUs, networking, advanced thermal management and cooling capabilities, as well as the kind of control software that you need to efficiently run these complex rack designs. So, a very, very strategic addition for us and not a typical acquisition, because the goal for AMD is indeed to leverage the 1,000 plus skills that have that design expertise, and what we did, Ross, was stated right up front that, we will — for the manufacturing side, Frank, the CEO of ZT Systems will continue to run that 1,500 person manufacturing side, but that we’ll be looking for strategic partners over time, of course, once we close. So the strategic impact for AMD is really one of accelerating our time to market, and optimizing that time — that solution which is so critical going forward to have the hardware, software, and the system design co-optimize together. Question-and-Answer Session Q – Ross Seymore And is this something that will have an immediate benefit, because the system design side of things can pull in the existing AMD silicon, mainly on the MI side of things, I would assume. Or is it something that you’ll have to wait a little bit longer for the benefits to accrue because it will be dependent upon kind of integrating the roadmap from the silicon perspective as well. Mark Papermaster The full benefit, of course, will be upon close. We anticipate close by mid-2025. This requires U.S. and EU approval, because that’s where ZT does business. And when you think about between now and then, we’ve already been working with ZT Systems and other ODM and OEM in terms of that system design. So we already have, I know that we have a very good work relationship. Of course, it needs to be still as separate companies, but we expect to continue that deep dialogue and partnership as we have. And then upon close, you’ll see a much deeper integration, you’ll see a true co-design from silicon to systems. Ross Seymore And talk a little bit about the rising importance of systems. How did that come to be, and how does AMD expect to differentiate in its approach? Mark Papermaster When you look at AI applications, they’re incredibly demanding of high-performance compute and so they demand the absolute best of silicon capability. Well, in AMD, we’re well armed there, are the leaders in chiplet technology, we’re the leaders in bringing together high performance CPUs, GPUs, accelerators. And we have a strong network portfolio based on our acquisitions with Xilinx and Pensando. We’ve been building our software skills. And so that’s critically important to actually bring that computation to bear. But again, when I go back to the demands of the most complex AI models, the foundational models, which really require massive clusters of heterogeneous systems, the optimization doesn’t stop at that hardware and software of the compute engines. You need to optimize through the networking, through the connectivity, through the rack. And anything can become a bottleneck. If you have issues from wiring capabilities, if you have issues from a cooling standpoint, if you didn’t fully integrate the networking in the most optimum way, then you are not delivering the most optimized system. So we, of course, partner with our system providers today. It’s a serial process. We’re designing that compute complex, and then we work with our system providers and continue that optimization so that part doesn’t change. But with the ZT design capabilities coming in-house from the earliest conception of our next-generation AI-based processing complexes, we will take into account the system aspects. We’ll be actually optimizing for all those facets that I just described and that is a game changer. Earlier time to market of an optimized system all the way through the rack implementation. Ross Seymore Does it change at all? Your relationship with other partners from a system perspective? Mark Papermaster That’s what’s critical about our strategy and why, as I described at the outset, that we are partitioning off the manufacturing capability. It is not our intent to compete with any of our partners. What we want to do though, is offer a significantly strengthened starting point, a reference design, where we have taken into account facets of the design, as I said, all the way through that system optimization. The 1,000 skills that come in from ZT that have been doing this for many, many years, will allow us to actually partner better with OEMs, with ODMs, with our end customers. Because as we understand those requirements, now, the starting point will be optimized from the get go. Ross Seymore Is this something that is more cloud attuned than kind of enterprise adopters of AI from your perspective? And I know cloud’s the most aggressive adopter of it, does it really matter the type of customer? Mark Papermaster You said it right, the cloud, which are driving the foundational model training and inference, are the most demanding of that system and rack level optimization, and so they’re the first customers that will benefit from this marriage of technologies that we will achieve upon close with the addition of ZT systems. But what happens — and this — I will give you an analogy to server systems today. So often the biggest server clusters are at the hyperscalers, but that learning and then optimization comes right down into the on-prem and enterprise installations, which are deployed across the Fortune 500. You’ll see a very similar trend here, the utmost of scale up, meaning the largest of the GPU compute nodes of the scale out, meaning the building of the largest clusters will be hyperscale. But all of that technology honing can be applied to ensure that enterprise class implementations that may not be running those massive foundational models, they’re going to benefit equally from scale back systems in tune with their workload demands. Ross Seymore And with the benefits on this, I assume everybody’s going to think the MI family would benefit from this, but would it also be beneficiary or would it be a positive also for the EPYC side of things? Mark Papermaster There is no doubt about the benefit for both our EPYC and our instinct lines. First of all, these AI clusters are CPU and GPU based. They are indeed heterogeneous. So that integration of CPU and GPU into that rack level design will be an immediate beneficiary of the ZT System skill inclusion into the AMD family, but beyond that this is 1,000 engineers, extremely experienced in system optimization. So that will be applied to standalone GPU optimization and X86 EPYC server systems. Ross Seymore So that is enough on, at least, unless you want to talk about it more on the ZT side of things. Let’s talk about the instinct side. The MI300, you guys have ramped that amazingly fast. The people debating what the number ends up being this year, but going from basically zero to somewhere around $5 billion over the course of a year or so is pretty impressive. Talk about what the biggest challenge is in the next-generations of product. I know the race never ends, but as you go to the next gen of that, the 325, et cetera, et cetera, where does AMD start to differentiate or continue to differentiate in that roadmap? Mark Papermaster Well, let’s start with the differentiation that we have right now. How did we get that differentiation? We listen to our customers. That’s a hallmark of AMD. We’re able to collaborate extremely well, listen to our customers, when they saw what we had done with the MI250, which powers the largest supercomputer in the world today with the Department of Energy Frontier System at Oak Ridge, and LUMI and other installations. We earned the trust and understood their requirement, and were able to show our agility and pivot the next-generation design that we have for HPC for supercomputers, called MI300A over to MI300X AI optimized for the most demanding inference and training applications. And in fact for inferencing, we’ve been differentiated because we have up to 192 megabytes of HBM3 memory on that starting point. And so, in fact, by the way, I’ll mention that we did publish MLPerf results. And you’ll see today this came out and you’ll see the competitive stance of MI300 on that standard benchmark. So where do we go in our roadmap going forward? What’s our strategy? One, we’ve embraced an annual cadence. If you think about Data Centers, typically they’ve been on an 18 to 24-month type of cadence, but AI systems are on an annual cadence. And the reason is that the innovation on the models and the model size, the number of parameter of the models and what the capability of what those foundational AI models can do is dependent on the growth of the compute engines. To support that expanded parameter growth, it has to be more efficient or energy alone. Just the power would gate the advancement of AI and the industry. And so it is a necessity to have a much more rapid cadence and we embrace that. And so what you should expect from AMD is we’re leveraging all of the know-how that we’ve had for years, supplying the industry leading edge, high performance computing. Being the leader in chiplet, we are increasingly going to leverage that to hit that annual cycle. We have the MI325 enhancing to HBM3E, which we have come into production next quarter. We’re on track with the MI350. And both the MI325 and the MI350 family will stay in that same socket that we have that same baseboard, the UBB. So it’s an easy adoption cycle for our customers. We’re not demanding a new system infrastructure, but we’re adding a doubling of memory capacity with 325. So that differentiation that we have on inferencing, we expect to persist with the MI325. And then with the MI350 again dropping right into that same system infrastructure with ease of adoption, we’re going to have the next generation of our GPU compute engine, and that’s going to have new math formats. That’s going to support FP4 and FP6. It’s going to yet leverage not only the same socket, but the same HBM3E memory. So it’s a plan of how do you get customers a significant increase in capability every generation, but work closely with them and make it easy for them to consume these new capabilities. And then beyond that, with the MI400 family, which will be for the year following 2026. Really, really exciting advancements across the board. Again, great partnership with our customers, and even more focus on the networking capabilities and system optimization, because by the time we come out with the MI400 family, assuming successful close of ZT System acquisitions, we’ll have thumbprints all over that in terms of even further system optimization. Ross Seymore So you and your primary competitor in this space have talked about that annual cadence and demand is off the charts. That’s great. But can the customers truly implement these solutions at the speed with which you are introducing them. Do they have a diversity of workloads? They’re not going to go whole hog with the 300 and no 325 and skip to the 350. How do we match the sheer dollars of CapEx and the time it takes to install these systems relative to the speed that both you and NVIDIA are trying to move? Mark Papermaster What’s key in terms of adoption is making sure that there’s a consistent software and software application story as you move generation to generation. When you look at MI300X, we launched in December of last year, and that was with our software base of ROCm 6.0. So then we brought rock to production level of 6.0, and we’re in production. So look at Azure as our announcements that GPT-4, GPT-4.0, we look at Llama 3.1 coming out on day one on AMD, Meta you saw on stage with us, as we announced 300 and they’ve been such a great partner. It’s the learning from now real world production applications, running on MI300 that we fold into ROCm 6.1, 6.2. We’ve tuned the performance, we’ve increased the ease of adoption. We’ve taken important features like the transformer optimizations with Flash Attention two built in now to ROCm. And so, it becomes rather than just hardware generation, it becomes a fluid flow Ross, in terms of making sure that the software enhancement learning that we do is a consistent and seamless flow generation to generation. And then with the hardware capabilities, what we do is increase the performance, and most importantly, the performance for watt of energy, again, this is an energy gated sector. And so, we really are very keen on the productivity that our roadmap will bring in every generation, yet on the base of a consistent software platform. Ross Seymore So today, I believe your solutions are largely have a leadership position on the imprint side, a little less so on the training side. Talk about what gives that advantage or disadvantage and, and perhaps more importantly, looking forward, what side are you actually focusing more on between those two, if it might not be an either or. Mark Papermaster Well, it’s certainly not an either or. We’ve targeted the instinct family squarely at the most demanding applications of both inferencing and training in the AI space. We did lean in, of course, upfront on inferencing with MI300, and the reason is we were the new entrant to actually bring competition to this market. And it was important that we establish ourselves where we had a clear differentiation. Our chiplet capabilities and prowess, and the ability to have a leadership memory capacity, as well as bandwidth was where we wanted to lead into market. And it’s proven out to be very beneficial. It’s providing immediate total cost of ownership advantages to our customers, and it’s a commitment that we’ll maintain throughout our roadmap. So as I said a moment ago, you’ll see us very focused on I’ll say staying on the torrid pace that our customers demand. But training where we are certainly competitive from the outset with MI300. Training, no doubt is a harder task, because you need to take additional time, additional time to hone the networking capability, the scale out capability as you build clusters. You need to spend the time on the communications library, the software which is managing how the GPUs are working close knit across the build out of that cluster. And that’s exactly what we’ve done. So we are working with initial implementations on training, we’re tuning the training, so we’re not waiting for the future roadmap of instinct. We are starting now to bring training to bear, as well as inference, which of course is the production instances that we have out there today. Ross Seymore The last hardware related question on this AI infrastructure for you, ASIC versus the traditional — well, GPU traditional otherwise. One, do you see ASIC as a threat or is that an opportunity? You do ASIC in another part of your business, is there any reason why you couldn’t at some point choose to do it within your Data Center infrastructure business as well? Mark Papermaster Well, if it’s a silicon-based solution that’s in need of high performance computing, we’re going to view all of that as an opportunity, because across our general offerings of CPU, GPU accelerated computing, we have in addition to that, of course, a Semi-Custom division. And we’ve been driving our Semi-Custom division to be more and more efficient to get our cost structure very close to ASIC like. And so we do view that as an opportunity. But moreover, I think the broader point is the size of the market. When you look at a TAM that we project at being $400 billion by 2027, what you’re going to see is there’s a need for a range of solutions. It is not one type of computing solution. Of course, the CPU and GPU combination is the most flexible. You have programming environments out there that are very, very well established. As the models change, they can adapt immediately to new algorithms. And so that will persist. But you will see whether with startups or with hyperscalers are investing in their own custom silicon teams, where you have an algorithm that’s stabilized and where you can come at a unique angle or tailor on a piece of market, that’s very much going to be a part of that. Overall, $400 billion TAM, and they’re going to coexist together. It’s going to be a key part of driving more efficiency. Efficiency of compute is the name of the game in AI, and we welcome that as both competitive pressure to make sure our general offerings stay in the utmost of their competitiveness, as well as to participate in to our Semi-Custom division. Ross Seymore So actually, lied one more question on the hardware side. Do FPGAs play any role in this? On the Data Center infrastructure side? Mark Papermaster They do. Our FPGAs or adaptive compute as we call it, are actually one of the fastest ways to tailor and to customize based on a known workload that our customers have. So we already have and are working with major customers in terms of hyperscaler deployments which can allow tailoring in terms of workload flow, network management and actually storage optimizations. Ross Seymore So why don’t we switch gears away from Instinct and over to EPYC for a bit? AMD has done an amazing job of targeting that market that forever was dominated by one of your competitors and taking significant share, at least by my math, potentially a majority share in some of the markets. Talk a little bit about the — or GPU versus CPU crowding out dynamic. I know it’s more the business side than the technology side, but do you see that as something that we’re kind of towards the end of it and both sides are going to start to be appreciated a little bit more going forward or do you think more or less people are going to keep focusing on your Instinct side and not your EPYC side for some time to come? Mark Papermaster Well, it’s a great question. I think the question is borne out of some of the dynamics that we saw over, I’ll say, the backward looking 12 months. And there was a huge rush to strengthen the GPU infrastructure. Clearly, hyperscalers are making a huge investment, given the jump in model size. But you also saw a number of enterprise wanting to make sure they didn’t get caught flat-footed. It didn’t have AI compute capability. So there was some impact in terms of a GPU demand taking, I’ll say, a bigger share of the CapEx allocation on a year-to-year basis. But I think that’s normalized now. You look at — going forward, guess what? The applications that always ran on a general purpose CPU still want to run on a general-purpose CPU. You need to close the books every quarter. You need to have your ledger analysis. You have your customer relationship management. You may have an AI augment on top of that, but you’re still running your CIM as a general-purpose x86 compute. And likewise, on many other workloads that don’t go away, but add an AI augment. And so what that means is where our customers that have a legacy x86 installed base, they’re finding that they’re at the end of their depreciation cycle. And now when we show them what we can do with our fourth gen EPYC, our gen [AI] that we’ve been shipping over last year, our fifth gen EPYC that is already in production sampling and goes into full announced and release mode in Q4. What we’re showing is that we can save dramatic 50%, 60% time of TCO cost savings. You can take that legacy data center and shrink the footprint, shrink the power, shrink the floor footprint and make room for your next-generation workloads, whether they be CPU or CPU and GPU based. So I think I do see more of the last 12 months being more of that debate of CPU versus GPU. But I think we’re normalizing. And frankly, the growth of AI and accelerated compute with the GPU is also going to drive yet more demand on those general-purpose x86 workloads. We’re very excited about the advantages we offer our customers with our EPYC product line. Ross Seymore So on the — sticking within the EPYC side of things, talk about the architectural choices that your customers want you to make as far as the core count side. So kind of Bergamo versus Genoa. How do you see the market moving? And from a business perspective, we’re moving towards pricing per core count versus pricing per chip per se. I know they end up at the same place. But just talk about the reasons economically why that could be a good thing for AMD? Mark Papermaster Well, there’s been a very significant shift in the driver of how you configure your x86 server systems. VMware did shift their pricing schemes. And it turned out that having a more performance CPU and the efficiency, energy efficiency that you get with more performance CPU. And now the licensing costs that you get, lower licensing cost if you have a more performance CPU are, in fact, really driving a major driver of our Genoa sales for those that are leveraging that type of virtualization capability. So it moreover, underscores how demand is driven by workloads. Just like data centers today are heterogeneous across CPU and GPU. When you look at CPU alone, it’s not one type of server that makes sense for most customers. Again, for those virtualized environments you want, Genoa’s perfect with that high performance 96 Core CPU that gives you that TCO advantage. But you mentioned Bergamo. It’s really our native Cloud optimized workload. So where you have Cloud native and don’t need a necessarily the highest performance of CPU, but you want the most efficient. And so we go from 96 to 128 Core, very, very efficient. And so throughput systems, things like recommendation engines, those are workloads that really excel. Beyond that, we’ve further diversified. We have stacked V-Cache for where you might have high performance or database workloads, where we really increase the Cache size of the CPU and we have telco optimized. So again, the diversification of our EPYC CPU product line is really a result of listening to our customers and making sure that we have the diversification to match where their workloads are. Ross Seymore Have you seen any change in the competitive intensity? Your roadmap seems to be very aggressive and you’ve executed superbly on it, whether it be from the x86 side. Is that gap closing at all? Or more recently with Arm-based solutions. Do you see that as a meaningful threat or I know the answer is it’s always, the world is always competitive, we always have to run fast, et cetera? But any meaningful change you’ve seen? Mark Papermaster Well, your last comment is true. It is always competitive. And our view is actually very simple. Like the best defense is a super strong offense. So I mean, our strategy is don’t slow down, don’t slow down on the innovation, don’t slow down on our pace, don’t slow down on aggressive use of semiconductor leading edge technology, but doing it in a chiplet-based approach where we bring the leading edge where you need it and not where you don’t. And so that’s what drives us, as well as the diversification to the customer workloads, making sure that we truly deliver TCO advantage based on the workload. That is what positions us best to rise in competition. Of course, there are new entrants. You see Arm-based designs that are tailored for specific workloads. Had we not diversified our portfolio, we would be at competitive disadvantages in that stance. So bottom line, we expect nothing but strong competition. This is what our DNA at AMD is all about. We thrive on that competitive, competitive spirit. And it will drive us to get the absolute best systems out for our customers. Ross Seymore So just to front run the AIPC topic and get the competition side out of the equation in both ways, is there something inherent in the Arm architecture, you believe, versus the x86 architecture that is an advantage, whether it’s on power consumption, performance, whatnot, and changes the game, versus an x86 to x86 competitive environment that we’ve seen more predominantly over the last, say, 20 years? Mark Papermaster We don’t see that the instruction set architecture is a differentiator. At the end of the day. I mean, we love x86 because there’s a massive install base. It’s a CPU architecture that we’ve been able to hone microarchitectural implementations over decades to every bit of performance out of it and the software tool chain that is out there and the familiarity of applications development make it a superb play. Yet, we offer ARM across our portfolio. It’s in our adaptive embedded compute. It’s in various controllers we have. That — end of the day is we’re about providing the best solutions for our customers. You look at Zen 5 that we’re rolling out now. We just brought it out in our PC market with a strict point. What do we do with Zen 5? We widened the execution pipeline? We added AVX-512, full physical implementation. And you might think, well, that’s for more like server, and that’s for those vector calculations and floating point calculations. But what we have is the capability to also excel with AI because the — the core of the AI is, in fact, a multiply accumulate, which we performed very, very well and we added FP16 in Zen 5. So we’ll continue to evolve the x86 architecture to excel at the workloads needed across PC, embedded in our data center customers. And we also were an ARM ecosystem, not the ARM ISA, there are differences there, but more it’s the ARM ecosystem. Where does it play and where that’s an advantage. Like our embedded road map will persist and continue to partner closely with ARM in those markets. In the PC, we’ll see. Now it’s a dual ISA play. So we’re going to leverage x86 where that’s best. But again, we are not married to ISO. We’re married to the best experience we can provide to our customers. Ross Seymore How does AI change the road map of your PC CPUs? Mark Papermaster AI changes dramatically in terms of the horsepower needed to operate AI workloads in an incredibly restricted power environment. So if you think of what does an AIPC do, it’s allowing you to have, at your fingertips, on the device you’re using for the bulk of your content creation, the ability that — to run incredibly efficiently AI applications. Yet you won’t accept if your battery life degrade significantly. So the demand upon us and our competitors, frankly, for the AIPC was how to provide AI acceleration without impacting that experience. And this will all come to light as we close 2024, but as we ramp into 2025 and you start seeing the amazing applications that will be available to you in the PC space. So Copilot+, which we now support and run on our AMD systems, will bring a plethora of new capabilities. In 2 or 3 years, you won’t remember what it was like when you didn’t have those AI tools to assist you and all of your day-to-day work tasks and your personal life. And so it is — it’s added a whole new tier because we now have high MPCs, which have significant investment in the AI capability in AMD and our new strict point, we added 50 tops of AI acceleration on top of the Zen 5 CPU and on top of Radeon 3.5 GPU that was optimized based on the partnership we have with Samsung on that GPU technology, the [indiscernible] in Galaxy. So an incredibly hone combination of IPs to deliver — hone combination of IP to deliver a superior AIPC experience. Ross Seymore Do you think there’s any, from a CTO perspective, any kind of cost tradeoffs that you have to hit? You just said you’re going to add on top of everything you’re going to add all this functionality. Is the price of the end device just going to go up and it’s going to be a higher end — establish a new higher end PC or when you’re thinking about the roadmap, does cost come into the equation as well? Mark Papermaster Well, the outset it will be a higher price tier for the highest capability of AIPC. So where our six point with 50 top capability is positioned in the higher tier from an ASP strata. But it won’t take long before the AI enablement is really end-to-end across the PC portfolio, across the industry. That’s how quickly I believe the adoption will be of these AI enabled PC-based applications. Ross Seymore Well, Mark, we are officially out of time. Thank you so much for all your insights and joining us here at the conference. I appreciate it. Mark Papermaster Ross, thanks for having me.

Buy cryptocurrency



Source link

Refer And Earn Demat Account – Get ₹300 | Referral Program

Open Demat Account In Angel One For FREE

Leave a Reply

Your email address will not be published. Required fields are marked *