Transcript for:
NVIDIA's Role in AI Innovations

What they achieved is singular, never been done before. Just to put in perspective, 100,000 GPUs, that's easily the fastest supercomputer on the planet. That's one cluster.

A supercomputer that you would build would take normally three years to plan. Right. And then they deliver the equipment, and it takes one year to get it all working. Yes. Jensen's nice glasses.

Hey, yeah. It's great to be with you. Yeah, I got my ugly glasses on just like come on Those aren't ugly. It's pretty good.

Do you like the red ones better? There's something only your family could love Well, it's Friday October 4th. We're at the NVIDIA headquarters just down the street from altimeter Thank you.

Thank you. And we have our investor meeting our annual investor meeting on Monday Where we're going to debate all the consequences of AI, how fast we're scaling intelligence, and I couldn't think of anybody better really to kick it off with than you. Appreciate that. As both a shareholder, as a thought partner, kicking ideas back and forth, you really make us smarter.

And we're just grateful for the friendship. So thanks for being here. Happy to be here. You know, this year, the theme is scaling intelligence to AGI. And it's pretty mind boggling that when we did this two years ago, we did it on the age of AI.

And that was two months before ChatGPT. And to think about all this change. So I thought we would kick it off with a thought experiment and maybe a prediction.

If I colloquially think of AGI as that personal assistant in my pocket. If I think of AGI as that colloquial assistant in my pocket. I was getting used to it. Exactly. You know, that knows everything about me.

That has perfect memory of me. That can communicate with me. They can book a hotel for me or maybe book a doctor's appointment for me.

When you look at the rate of change in the world today, when do you think we're going to have that personal assistant in our pocket? Soon, in some form. Yeah.

Yeah, soon in some form. And that assistant will get better over time. That's the beauty of technology as we know it. So I think in the beginning, it'll be...

quite useful, but not perfect. And then it gets more and more perfect over time, like all technology. When we look at the rate of change, I think Elon has said the only thing that really matters is rate of change.

It sure feels to us like the rate of change has accelerated dramatically, is the fastest rate of change we've ever seen on these questions, because we've been around the rim like you on AI for a decade now, you even longer. Is this the fastest rate of change you've seen in your career? It is because we've reinvented computing. You know, a lot of this is happening because we drove the marginal cost of computing down by 100,000x over the course of 10 years.

Moore's Law would have been about 100x. And we did it in several ways. We did it by, one, introducing accelerated computing, taking what is work that is not very effective on CPUs and put it on top of GPUs.

We did it by... Inventing new numerical precisions, we did it by new architectures, inventing the Tensor Core, the way systems are formulated, MVLink, added insanely fast memories, HBM, and scaling things up with MVLink and InfiniBand, and working across the entire stack. Basically, everything that I described about how NVIDIA does things that led to A super Moore's Law rate of innovation.

Now the thing that's really amazing is that as a result of that, we went from human programming to machine learning. And the amazing thing about machine learning is that machine learning can learn pretty fast, as it turns out. And so as we reformulated the way we distribute computing, we did a lot of... the parallelism of all kinds, right?

Tensor parallelism, pipeline parallelism, parallelism of all kinds. And we became good at inventing new algorithms on top of that and new training methods. And all of this invention is compounding on top of each other as a result, right?

And back in the old days, if you look at the way Moore's Law was working, the software was static. It was pre-compiled, it was shrink-wrapped, put into a store. It was static. And the hardware underneath was growing at Moore's Law rate.

Now we've got the whole stack growing, innovating across the whole stack. And so I think that that's the... Now all of a sudden we're seeing scaling. That is extraordinary, of course. But we used to talk about pre-trained models and scaling at that level.

and how we're doubling the model size and doubling, therefore, appropriately doubling the data size. And as a result, the computing capacity necessary is increasing by a factor of four every year. Right.

That was a big deal. Right. But now we're seeing scaling with post-training and we're seeing scaling at inference. Isn't that right? Right.

And so people used to think that pre-training was hard and inference was easy. Now everything is hard. Right. Which is kind of sensible. You know, the idea that...

that all of human thinking is one shot, is kind of ridiculous. And so there must be a concept of fast thinking and slow thinking and reasoning and reflection and iteration and simulation and all that. And that now it's coming in. I think to that point, one of the most misunderstood things about NVIDIA is how deep the true NVIDIA moat is. I think there's a notion out there that as soon as someone invents a new chip, a bomb.

a better chip that they've won. But the truth is you've been spending the past decade building the full stack from the GPU to the CPU to the networking and especially the software and libraries that enable applications to run on NVIDIA. So I think you spoke to that, but when you think about NVIDIA's moat today, do you think NVIDIA's moat today is greater or smaller than it was three to four years ago?

Well, I appreciate you recognizing how computing has changed. In fact, the reason why people thought, and many still do, that you designed a better chip, it has more flops, has more flips and flops and bits and bytes, you know what I'm saying? Yeah.

And you see their keynote slides, and it's got all these flips and flops and bar charts and things like that. And that's all good. I mean, look, horsepower does matter. Yes.

So these things fundamentally do matter. However, unfortunately, that's old thinking. It is old thinking in the sense that the software was some application running on Windows.

And the software is static. Right. Which means that the best way for you to improve the system is just making faster and faster ships. But we realized that machine learning is not human programming.

Machine learning is not about just the software. It's about the entire data pipeline. It's about, in fact, the flywheel.

of machine learning is the most important thing. So how do you think about enabling this flywheel on the one hand and enabling data scientists and researchers to be productive in this flywheel? And that flywheel, is starts at the very, very beginning. A lot of people don't even realize that it takes AI to curate data to teach an AI.

And that AI alone is pretty complicated. Yeah. And as that AI itself is improving, is it also accelerating, you know, again, when we think about the competitive advantage, right? It's combinatorial of all these systems. Exactly, exactly.

And that was exactly going to lead to that because of... Smarter AIs to curate the data. We now even have synthetic data generation and all kinds of different ways of curating data, presenting data. So before you even get the training, you've got massive amounts of data processing involved. So people think about PyTorch, that's the beginning and end of the world, and it was very important.

But don't forget, before PyTorch, there's an amount of work. After PyTorch, there's an amount of work. And the thing about the flywheel is really the way you ought to think. How do I think about this entire flywheel? And how do I design a computing system, a computing architecture, that helps you take this flywheel and be as effective as possible?

It's not one slice of an application, training. Does that make sense? That's just one step. Every step along that flywheel is hard.

And so the first thing that you should do, instead of thinking about, how do I make Excel faster? How do I make, you know, Doom faster? That was kind of the old days, isn't that right?

Now you have to think about, how do I make this flywheel faster? And this flywheel has a whole bunch of different steps. And there's nothing easy about machine learning, as you guys know. There's nothing easy about what OpenAI does or X does or Gemini and the team at DeepMind does.

I mean, there's nothing easy about what they do. And so we decided. Look, this is really what you ought to be thinking about.

This is the entire process. You want to accelerate every part of that. You want to respect Amdahl's law.

Amdahl's law would suggest, well, if this is 30% of the time, and I accelerated that by a factor of three, I didn't really accelerate the entire process by that much. Does that make sense? And you really want to create a system that accelerates every single step of that because only in doing the whole thing can you really materially improve. That's cycle time. And that flywheel, that rate of learning is really, in the end, what causes the exponential rise.

And so what I'm trying to say is that our perspective about, you know, a company's perspective about what you're really doing manifests itself into the product. Right. And notice, I've been talking about this flywheel.

The entire cycle, yeah. That's right. Yeah.

And we accelerate everything. Right. Right now.

Right now. The main focus is video. A lot of people are focused on physical AI and video processing. Just imagine that front end.

The terabytes per second of data that are coming into the system. Give me an example of a pipeline that is going to ingest all of that data, prepare it for training in the first place. So that entire thing is CUDA accelerated. And people are only thinking about...

text models today. Yeah. But the future is, you know, this video models, as well as, you know, using, you know, some of these text models, like O1, to really process a lot of that data before we even get there. Yeah.

Right? Yeah, yeah. So language models are going to be involved in every single... Yeah.

What it took us, took the industry enormous technology and effort. to train a language model, to train these large language models. Now we're using a large language model in every single step of the way.

It's pretty phenomenal. I don't mean to be overly simplistic about this, but again, we hear it all the time from investors, right? Yes, but what about custom ASICs? Yes, but their competitive mode is going to be pierced by this. What I hear you saying is that in a combinatorial system, the advantage grows over time.

So I heard you say that our advantage is greater today than it was three to four years ago because we're improving every component and that's combinatorial. Is that, you know, when you think about, for example, as a business case study, Intel, right, who had a dominant mode, a dominant position in the stack relative to where you are today, perhaps just, you know, again, boil it down a little bit, you know, compare, contrast your competitive advantage. to maybe the competitive advantage they had at the peak of their cycle? Well, Intel is extraordinary. Intel is extraordinary because they were probably the first company that was incredibly good at manufacturing, process engineering, manufacturing.

And that one click above manufacturing, which is building the chip. Right. And designing the chip and architecting the chip in the x86 architecture and building building faster and faster x86 chips. That was their brilliance, and they fused that with manufacturing.

Our company is a little different in the sense that, and we recognize this, that in fact, parallel processing... doesn't require every transistor to be excellent. Serial processing requires every transistor to be excellent. Parallel processing requires lots and lots of transistors to be more cost-effective. I'd rather have 10 times more transistors, 20% slower, than 10 times less transistors, 20% faster.

Does that make sense? They would like the opposite. And so single-threaded performance, single-threaded processing, and parallel processing was very different.

And so we... We observed that, in fact, our world is not about being better going down. We want to be very good, as good as we can be.

But our world is really about much better going up. Parallel computing, parallel processing is hard because... Every single algorithm requires a different way of refactoring and re-architecting the algorithm for the architecture.

What people don't realize is that you can have three different ISAs, CPU ISAs. They all have their own C compilers. You could take software and compile down to that ISA. That's not possible in accelerated computing. That's not possible in parallel computing.

The company who comes up with the architecture has to come up with their own OpenGL. So we revolutionized deep learning because of our domain-specific library called KU-DNN. Without KU-DNN, nobody talks about KU-DNN because it's one layer underneath PyTorch and TensorFlow and back in the old days, Cafe and Theano and now Triton. And there's a whole bunch of different frameworks.

And so that domain-specific library, KU-DNN, a domain-specific library called Optics, we have a domain-specific library called KU-Quantum. Yeah. um rapids uh the list of you know ariel for for uh industry specific algorithms that sit below you know that pie torch layer that everybody's focused on like i've heard oftentimes well you know if llms if i didn't if we didn't invent that uh no application on top could work right you guys understand what i'm saying so the mathematics is really what nvidia is really good at is algorithm right that in the fusion between the the science above the architecture on the bottom That's what we're really good at. There's all this attention now on inference, finally.

But I remember two years ago, Brad and I had dinner with you, and we asked you the question, do you think your moat will be as strong in inference as it is in training? Yeah. And I'm sure I said it would be greater.

Yeah, yeah. And you touched upon a lot of these elements just now, just the composability between, or... We don't know the total mix at one point. And to a customer, it's very important to be able to be flexible in between. That's right.

But can you just touch upon now that we're in this era of inference? It was inference. Training is inferencing at scale.

I mean, you're right. And so if you train well, it is very likely you'll inference well. If you built it on this architecture, without any consideration, it will run on this architecture.

You could still go and optimize it for other architectures, but at the very minimum, since it's already been built on NVIDIA, it will run on NVIDIA. Now, the other aspect, of course, is just kind of... You know, capital investment aspect, which is when you're training new models, you want your best new gear to be used for training, which leaves behind gear that you used yesterday. gear is perfect for inference.

And so there's a trail of free gear. There's a trail of free infrastructure behind the new infrastructure that's CUDA compatible. And so we're very disciplined about making sure that we're compatible throughout so that everything that we leave behind will continue to be excellent.

Now, we also put a lot of energy into continuously reinventing new algorithms so that... When the time comes, the Hopper architecture is two, three, four times better than when they bought it. So that infrastructure continues to be really effective. And so all of the work that we do improving new algorithms, new frameworks.

Notice, it helps every single install base that we have. Hopper is better for it. Ampere is better for it.

Even Volta is better for it. And I think Sam was just telling me that they had just decommissioned the Volta infrastructure that they have at OpenAI recently. And so I think we leave behind this trail of install base. Just like all computing, install base matters.

And NVIDIA is in every single cloud. We're on-prem and all the way out to the edge. And so the...

The Vela vision language model that's been created in the cloud works perfectly at the edge on a robot without modification. It's all CUDA compatible. And so I think this idea of architecture compatibility was important for large it's no different for iPhones, no different for anything else.

I think the install base is really important for inference. But the thing that I really, really We really benefit from is because we're working on training these large language models and the new architectures of it, we're able to think about how do we create architectures that's excellent at inference someday when the time comes. And so we've been thinking about... about iterative models for reasoning models and how do we create very interactive inference experiences for this personal agent of yours.

You don't want to say something and have to go off and think about it for a while. You want it to interact with you quite quickly. So how do we create such a thing?

And what came out of it was MVLink. MVLink so that we could take these systems that are excellent for training, but when you're done with it, the inference performance is exceptional. And so you want to optimize for this time to first token. Right.

And time to first token is insanely hard to do, actually, because time to first token requires a lot of bandwidth. But if your context is also rich. then you need a lot of flops.

And so you need an infinite amount of bandwidth, infinite amount of flops at the same time in order to achieve just a few millisecond response time. And so that architecture is really hard to do. And we invented Grace Blackwell and V-Link for that. Right.

In the spirit of time, I have more questions about that, but... Don't worry about the time. Hey, guys.

Hey, hey, hey, listen. Janine? Yeah. Look.

Let's do it until right. Let's do it until right. There you go.

I love it. I love it. So, you know... I was at dinner with Andy Jassy earlier this week, and Andy said, you know, we've got Tranium coming and Inferencia coming, and I think most people, again, view these as a problem for NVIDIA. But in the very next breath, he said, NVIDIA is a huge and important partner to us and will remain a huge and important partner for us as far as I can see into the future.

The world runs on NVIDIA, right? So when you think about the custom ASICs that are being built, that are going to go after targeted application, maybe the inference accelerator at Meta, maybe, you know, Tranium at Amazon, you know, or Google's TPUs, and then you think about the supply shortage that you have today, do any of those things change that dynamic, right? Or are they complements to the systems that they're all buying from you?

We're just doing different things. Yes. We're trying to accomplish different things. What NVIDIA is trying to do is build a computing platform for this new world, this machine learning world, this generative AI world, this agentic AI world. We're trying to create, as you know, what's just so deeply profound is after 60 years of computing, we reinvented the entire computing stack.

The way you write software from programming to machine learning, the way that you process software from CPUs to GPU, the way that the applications from software to artificial intelligence, right? And so software tools to artificial intelligence. So every aspect of the computing stack and the technology stack has been changed.

What we would like to do is to create a computing platform that's available everywhere. And this is really the complexity of what we do. The complexity of what we do is, if you think about what we do, we're building an entire AI infrastructure, and we think of it as one computer.

I've said before, the data center is now the unit of computing. To me, when I think about a computer, I'm not thinking about that chip. I'm thinking about this thing.

That's my mental model. And all the software and all the orchestration, all the machinery that's inside, that's my computer. And we're trying to build a new one every year.

Yes. That's insane. Nobody has ever done that before. We're trying to build a brand new one every single year.

And every single year, we deliver two or three times more performance. As a result, every single year, we reduce the cost by two or three times. Every single year, we improve the energy efficiency by two or three times. Incredible.

Right? And so we ask our customers, don't buy everything at one time. Buy a little every year.

Right. Okay? And the reason for that, we want them to cost average into the future. All of it's architecturally compatible.

Okay. Now, so that building that alone at the pace that we're doing is incredibly hard. Now, the double part, the double hard part is then we take that all of that. And instead of selling it as a infrastructure. We're selling it as a service.

We disaggregate all of it and we integrate it into GCP. We integrate it into AWS. We integrate it into Azure. We integrate it into X.

Does that make sense? Yes. And so everybody's integration is different. We have to get all of our architectural libraries and all of our algorithms and all of our frameworks and integrate it into theirs. We get our security system integrated into theirs.

We get our networking integrated into theirs. Isn't that right? Right.

Then we do. Basically 10 integrations. And we do this every single year. Right.

Now, that is the miracle. That is the miracle. Why were you? I mean, it's madness.

It's madness that you're trying to do this every year. I'm going insane thinking about it. So what drove you to do it every year? And then related to that, Clark's just back from Taipei and Korea and Japan when meeting with all your supply partners who you have decade-long relationships with.

How important are those relationships to, again, the combinatorial math that builds that competitive moat? Yeah, that's when you break it down systematically, the more you guys break it down, the more everybody breaks it down, the more amazed that they are. Yes. And how is it possible that the entire ecosystem of electronics today is dedicated in working with us?

to build ultimately this cube of a computer integrated into all of these different ecosystems. And the coordination is so seamless. So there's obviously APIs and methodologies and business processes and design rules that we've propagated backwards and methodologies and architectures and APIs that we've propagated forward.

That have been hardened for decades. Hardened for decades, yeah. And also evolving as we go. Right.

But these APIs have to come together. Right. When the time comes, all these things in Taiwan, all over the world being manufactured, they're going to land somewhere in Azure's data center. They're going to come together, click, click, click, click.

Someone just calls it OpenAI API and it just works. That's right. Yeah, exactly. It's craziness, right?

There's a whole chain. So that's what we invented. That's what we invented, this massive infrastructure of computing.

The whole planet is working with us on it. It's integrated into everywhere. It's, you could sell it through Dell, you could sell it through HPE. It's hosted in the cloud. It's in, it's all the way out at the edge.

People use it in robotic systems now, and, you know, human robots. They're in self-driving cars. They're all architecturally compatible. Pretty kind of craziness.

It's craziness. Clark, I don't want to, I don't want you to leave the impression I didn't answer the question. In fact, I did.

What I meant by that, when relating to your ASIC, is... Is the way to think about we're just doing something different. Yes.

As a company, as a company, we want to be situationally aware, and I'm very situationally aware of everything around our company and our ecosystem. Right. I'm aware of all the people doing alternative things and what they're doing and sometimes it's adversarial to us, sometimes it's not. I'm super aware of it. But that doesn't change what the purpose of the company is.

The singular purpose of the company is to build an architecture, a platform that could be everywhere. That is our goal. We're not trying to take any share from anybody.

NVIDIA is a market maker, not share taker. If you look at our company slides, we don't show... Not one day does this company talk about market share. Not inside. All we're talking about is how do we create the next thing?

What's the next problem we can solve? In that flywheel, how can we do a better job for people? How do we take that flywheel that used to take about a year, how do we crank it down to about a month?

Yes. You know? What's the speed of light of that? Isn't that right? And so we're thinking about all these different things, but the one thing we're not, we're situationally aware of everything, but we're certain that what our mission is, is very singular.

The only question is whether that mission is necessary. Does that make sense? And all companies, all great companies, ought to have that at its core.

It's about what are you doing? For sure. The only question, is it necessary? Is it valuable?

Right. Is it impactful? Does it help people? And I am certain that you're a developer, you're a generative AI startup, and you're about to decide how to become a company, the one choice that you don't have to make is which one of the A6 do I support? If you just support a CUDA, you know you could go everywhere.

You could always change your mind later. But we're the on-ramp. to the world of AI, isn't that right? Once you decide to come onto our platform, the other decisions you could defer. You could always build your own ASIC later.

We're not against that, we're not offended by any of that. When we work with all the GCPs, the GCPs Azure, we present our roadmap to them years in advance. They don't present their ASIC roadmap to us.

And it doesn't ever offend us. Does that make sense? We create where in it, if you have a sole purpose and your purpose is meaningful and your mission is dear to you and is dear to everybody else, then you could be transparent.

Notice my roadmap is transparent at GTC. My roadmap goes way deeper to our friends at Azure and AWS and others. We have no trouble doing any of that, even as they're building their own ASIC.

I think, you know, when... when people observe the business, you said recently that the demand for Blackwell is insane. You said one of the hardest parts of your job is the emotional toll of saying no to people in a world that has a shortage of the compute that you can produce and have on offer.

But critics say this is just a moment in time, right? They say, this is just like Cisco in 2000. We're overbuilding fiber. It's going to be boom and bust.

You know, I think about the start of 23 when we were having dinner. The forecast for NVIDIA at that dinner in January of 23 was that you would do $26 billion of revenue for the year 2023. You did $60 billion. The 25 people...

Let the truth be known, that is the single greatest failure of forecasting the world has ever seen. Right, right, right. Can we all at least admit that?

To me... That was my takeaway. And that was... We got so excited in November 22 because we had folks like Mustafa from Inflection and Noam from Character coming in our office talking about investing in their companies. And they said, well, if you can't...

Pence allowed investing in our companies, then buy Nvidia because everybody in the world is trying to get Nvidia chips to build these applications that are going to change the world. And of course, the Cambrian moment occurred with ChatGPT. And notwithstanding that fact, these 25 analysts were so focused on the crypto winner that they couldn't get their head around an imagination of what was happening in the world.

OK, so it ended up being way bigger. You say in very plain English. The demand is insane for Blackwell, that it's going to be that way for as far as you can see.

Of course, the future is unknown and unknowable. But why are the critics so wrong that this isn't going to be the Cisco-like situation of overbuilding in 2000? Yeah. The best way to think about the future is reason about it from first principles. Correct.

Okay, so... the question is what are the first principles of what we're doing? Number one, what are we doing?

What are we doing? The first thing that we are doing is we are reinventing computing. Do we not?

We just said that. The way that computing will be done in the future will be highly machine learned. Yes.

Highly machine learned. Okay. Almost everything that we do, almost every single application, Word, Excel, PowerPoint, Photoshop, Premiere, you know, AutoCAD. You give me your favorite application that was all hand engineered, I promise you it will be highly machine learned in the future.

Isn't that right? And so all these tools will be... And on top of that, you're going to have machines, agents, that help you use them.

Right. Okay. And so we know this for a fact at this point, right? Isn't that right? We've reinvented computing.

We're not going back. The entire computing technology stack has been reinvented. Okay.

So now that we've done that, we said that software is going to be different. What software can write is going to be different. How we use software will be different.

So let's now acknowledge that. Those are my ground truth now. Yes. Now the question therefore is what happens?

And so let's go back and let's just take a look at how's computing done in the past. So we have a trillion dollars worth of computers in the past. We look at it, just open the door, look at the data center, and you look at it and say, are those the computers you want doing that, doing that future? And the answer is no. Right.

You got all these CPUs back there. We know what it can do and what it can't do. And we just know that we have a trillion dollars worth of data centers that we have to modernize.

And so right now as we speak, if we were to have a trajectory over the next four or five years to modernize that old stuff, that's not unreasonable. Right. Sensible. So we have a trillion.

And you're having those conversations with the people who have to modernize it. Yeah. And they're modernizing it on GPU.

That's right. Well, let's make another test. You have $50 billion of CapEx you'd like to spend.

Option A, option B, build CapEx for the future or build CapEx like the past. Now, you already have the capex of the past. Right. It's sitting right there. It's not getting much better anyways.

Moore's Law has largely ended. And so why rebuild that? Let's just take $50 billion, put it into generative AI.

Isn't that right? And so. So now your company just got better.

Now, how much of that $50 billion would you put in? Well, I would put in 100% of the $50 billion because I've already got four years of infrastructure behind me. That's of the past. And so now I just reasoned about it from the perspective of somebody thinking about it from first principles.

And that's what they're doing. Smart people are doing smart things. Now the second part is this.

So now we have a trillion dollars worth of capacity to go build, right? Trillion dollars worth of infrastructure. What about, you know, call it $150 billion into it.

Okay, so we have a trillion dollars in infrastructure to go build over the next four or five years. Well, the second thing that we observe is that the way that software is written is different, but how software is going to be used is different. In the future, we're going to have agents.

Isn't that right? We're going to have digital employees in our company. In your inbox, you have all these little dots and these little faces.

In the future, there's going to be little icons of AIs. Isn't that right? I'm going to be sending them. I'm going to be, I'm no longer going to program computers with C++, I'm going to program AIs.

with prompting, isn't that right? Now, this is no different than me talking to my, you know, this morning, I wrote a bunch of emails before I came here. I was prompting my teams, right? And I would describe the context.

I would describe the fundamental constraints that I know of. And I would describe the mission for them. I would leave it sufficiently, I would be sufficiently directional so that they understand what I need. And I want to be clear about what the outcome should be, as clear as I can be.

But I leave enough ambiguous space, you know, a creativity space so they can surprise me. Isn't that right? Absolutely.

It's no different than how I prompt an AI today. Yeah. It's exactly how I prompt an AI.

And so what's going to happen is on top of this infrastructure of IT that we're going to modernize, there's going to be a new infrastructure. This new infrastructure are going to be AI factories that operate these digital humans. And they're going to be running all the time, 24-7. Right. We're going to have them.

For all of our companies all over the world, we're going to have them in factories, we're going to have them in autonomous systems, isn't that right? So there's a whole layer of computing fabric, a whole layer of what I call AI factories that the world has to make that doesn't exist today at all. So the question is, how big is that?

Unknowable at the moment, probably a few trillion dollars. Unknowable at the moment, but as we're sitting here building into the beautiful thing is the architecture for this modernizing this new data center. And the architecture for the AI factory is the same.

That's the nice thing. And you made this clear. You've got a trillion of old stuff you've got to modernize.

You at least have a trillion of new AI workloads coming on. Give or take, you'll do $125 billion in revenue this year. You know, there was at one point somebody told you the company would never be worth more than a billion. As you sit here today, is there any reason, right, if you're only $125 billion out of a multi-trillion, Tam, That you're not going to have 2x the revenue, 3x the revenue in the future that you have today? Is there any reason your revenue doesn't?

No. Yeah. Yeah. As you know, it's not about, it's not about, everything is, you know, companies, companies are only limited by the size of the.

The fish pond, you know? Yes, yes. A gold fishing can only be so big. And so the question is, what is our fish pond?

What is our pond? And that requires a little imagination. And this is the reason why market makers think about that future, creating that new fish pond. It's hard to figure this out looking backwards and try to take share. You know, share takers can only be so big.

For sure. Market makers can be quite large. For sure. Yeah.

For sure. So, you know, I think the good fortune that our company has is that since the very beginning of our company, we had to invent the market for us to go swim in. And people don't realize this back then anymore, but, you know, we were at ground zero of creating the 3D gaming PC market.

Right. We largely invented this market and all the ecosystem and all the graphics card ecosystem. We invented all that. And so the need to invent a new market to go serve it later is something that's very comfortable for us. Exactly.

And speaking to somebody who's invented a new market, let's shift gears a little bit to models and open AI. Open AI raised, as you know, $6.5 billion this week. At like a hundred fifty billion dollar evaluation.

We both participated Yeah, we're really happy for them really really happy that came together, right? Yeah, they did a great Sam and the team did a great job Yeah reports are that they'll do five billion ish of revenue or run rate revenue this year Maybe going to ten billion next year if you look at the business today It's about twice the revenue as Google was at the time of its IPO They have 250 million weekly average users, which we estimate is twice the amount Google had at the time of its IPO. And if you look at the multiple of the business, if you believe 10 billion next year, it's about 15 times the forward revenue, which is about the multiple of Google and Meta at the time of their IPO.

When you think about a company that had zero revenue, zero weekly average users, 22 months ago. Brad has an incredible command of history. When you think about that, talk to us about the importance of OpenAI as a partner to you and OpenAI as a force in kind of driving forward, you know, kind of public awareness and usage around AI. Well, this is one of the most consequential companies of our time. A pure play AI company.

pursuing the vision of AGI and whatever its definition. I almost don't think it matters fully what the definition is, nor do I really believe that the timing matters. The one thing that I know is that AI is going to have a roadmap of capabilities over time, and that roadmap of capabilities over time is going to be quite spectacular.

Along the way, long before it even gets to anybody's definition of AGI, we're going to put it to great use. Right. All you have to do is, right now as we speak, go talk to digital biologists, climate tech researchers, material researchers, physical sciences, astrophysicists, quantum chemists.

You go ask video game designers. manufacturing engineers, roboticists, pick your favorite, whatever industry you want to go pick. And you go deep in there and you talk to the people that matter and you ask them, has AI revolutionized the way you work.

Right. And you take those data points and you come back and you then get to ask yourself, how skeptical do you want to be? Right, right.

Because they're not talking about AI as a concept. conceptual benefit someday. They're talking about using AI right now. Right now.

Ag tech, material tech, climate tech, you pick your tech. You pick your field of science. They are...

advancing, AI is helping them advancing their work right now as we speak. Every single industry, every single company, every university, unbelievable. Isn't that right? Right. It is absolutely going to somehow transform business.

We know that. Right. I mean, it's so tangible you could-It's happening today. It's happening today. It's happening today.

Yeah. And so I think- I think that the awakening of AI, chat GPT triggered, is completely incredible. And I love their velocity and their singular purpose of advancing this field.

And so really, really consequential. And they build an economic engine that can finance the next frontier of models, right? And I think there's a growing consensus in Silicon Valley. that the whole model layer is commoditizing.

Lama is making it very cheap for lots of people to build models. And so early on here, we had a lot of model companies, you know, Character and Inflection and Cohere and Mistral and go through the list. And a lot of people question whether or not those companies can build the escape velocity on the economic engine that can continue funding those next generation.

My own sense, is that there's going to be, that's why you're seeing the consolidation, right? Open AI clearly has hit that escape velocity. They can fund their own future. It's not clear to me that many of these other companies can. Is that a fair kind of review of the state of things in the model layer that we're going to have this consolidation, like we have in lots of other markets, to market leaders who can afford, who have an economic engine, an application, that allows them to continue to invest?

There's a, first of all, there's a different fundamental difference between a model and artificial intelligence. Yes. Right?

Yeah. A model is an essential ingredient for artificial intelligence. It's necessary but not sufficient.

Correct. And so, and artificial intelligence is a capability, but for what? Right. Then what's the application? Right.

The artificial intelligence for self-driving cars is related to the artificial intelligence for human or robots, but it's not the same. which is related to the artificial intelligence for a chatbot, but not the same. Correct.

And so you have to understand the taxonomy of the stack. And at every layer of the stack, there will be opportunities, but not infinite opportunities for everybody at every single layer of the stack. Now, I just said something. All you have to do is replace the word model with GPU. In fact, this was the great observation of our company 32 years ago.

That there's a fundamental difference between GPU, graphics chip or GPU, versus accelerated computing. And accelerated computing is a different thing than... the work that we do with AI infrastructure. It's related, but it's not exactly the same.

It's built on top of each other. It's not exactly the same. And each one of these layers of abstraction requires fundamental different skills. Somebody who's really, really good at building GPUs have no clue how to be an accelerated computing company. I can, there are a whole lot of people who build GPUs.

And I don't know which one came, you know, we invented the GPU, but you know that we're not, we're... we're not the only company that makes GPUs today. Correct. And so there are GPUs everywhere. But they're not accelerated computing companies.

And there are a lot of people who, they're accelerators, accelerators that does application acceleration. But that's different than an accelerated computing company. And so, for example, a very specialized AI application could be a very successful thing.

Correct. Meta's MTIA. That's right.

But it might not be. the type of company that had broad reach and broad capabilities. And so you've got to decide where you want to be. There's opportunities probably in all these different areas, but like building companies, you have to be mindful of the shifting of the ecosystem and what gets commoditized over time, recognizing what's a feature versus a product versus a company. For sure.

Okay. I just went through. Okay.

And there's a lot of different ways you can think about this. Of course, there's one new entrant. that has the money, the smarts, the ambition, that's x.ai.

Yeah. Right? And, well, there are reports out there that you and Larry and Elon had dinner.

They talked to you out of 100,000 H100s. They went to Memphis and built a large, coherent super cluster in a matter of months. So, first, three points don't make a line, okay?

Yes, I had dinner with them. Causality is it. What do you think about their ability to stand up that super cluster?

And there's talk out there that they want another 100,000 H200s, right, to expand the size of that super cluster. You know, first talk to us a little bit about X and their ambitions and what they've achieved. But also, are we already at the age of clusters of 200,000 and 300,000 GPUs?

The answer is yes. And then the... First of all, acknowledgement of achievement where it's deserved. From the moment of concept to a data center that's ready for NVIDIA to have our gear there, to the moment that we powered it on, had it all hooked up, and it did its first training.

Yeah. Okay? Correct. So that first part, just building a...

Massive factory, liquid-cooled, energized, permitted in the short time that was done. I mean, that is like superhuman. And as far as I know, there's only one person in the world who could do that.

Elon is singular in this understanding of engineering and construction and large systems. and marshalling resources. Incredible.

Yeah, it's unbelievable. And of course, then his engineering team is extraordinary. I mean, the software team is great, the networking team.

team is great. The infrastructure team is great. You know, Elon understands this deeply. And from the moment that we decided to get to go, the planning with our engineering team, our networking team, our infrastructure computing team, the software team, all of the preparation advance, then all of the infrastructure, all of the logistics and the amount of technology and equipment that came in.

On that day, NVIDIA's infrastructure and computing infrastructure and all that technology, to training, 19 days. You know what? Did anybody sleep 24-7?

No question that nobody slept. But first of all, 19 days is incredible. But it's also kind of nice to just take a step back and just, do you know how nice... How many days in 19 days is? It's just a couple of weeks.

And the mountain of technology, if you were to see it, is unbelievable. All of the wiring and the networking. Networking NVIDIA gear is very different than networking hyperscale data centers. The number of wires that goes in one node, the back of a computer is all wires.

Just getting this mountain of technology integrated and all the software, incredible. So I think what Elon and the X team did. And I'm really appreciative that he acknowledges the engineering work that we did with him and the planning work and all that stuff. But what they achieved is singular. Never been done before.

Just to put in perspective, 100,000 GPUs, that's easily the fastest supercomputer on the planet as one cluster. A supercomputer that you would build would take normally three years to plan. Right.

and then they deliver the equipment and it takes one year to get it all working yes we're talking about 19 days wow what's the credit of the nvidia platform right that it's the whole processes are hardened that's right yeah yeah everything's already working and yeah and of course there's a whole bunch of you know x algorithms and x framework and x stack and things like that And we had a ton of integration we have to do. But the planning of it was extraordinary. Just pre-planning of it. N of 1 is right.

Elon is an N of 1. But you answered that question by starting off saying, yes, 200,000 to 300,000 GPU clusters are here. Yeah. Right?

Does that scale to 500,000? Does it scale to a million? And does the demand for your products... depend on it scaling to millions? That part, the last part is no.

My sense is that distributed training will have to work. Right. My sense is that distributed computing will be invented. Right. Some form of federated learning and distributed asynchronous distributed computing is going to be discovered.

And I'm very enthusiastic and very optimistic about that. Of course, the thing to realize is that the scaling law used to be about pre-training. Now we've gone to multimodality.

We've gone to synthetic data generation. Post-training has now scaled up incredibly. Synthetic data generation, reward systems, reinforcement learning-based.

And then now... inference scaling has gone through the roof. The idea that a model, before it answers your answer, had already done internal inference 10,000 times is probably not unreasonable. And it's probably done tree search.

It's probably done reinforcement learning on that. It's probably done some simulations. It's surely done a lot of reflection.

It probably looked up some data. It looked at some information. Isn't that right? And so its context is probably fairly large. I mean, this type of...

intelligence is, well, that's what we do. That's what we do. Isn't that right?

And so the ability, this scaling, if you did that math and you compound that with 4x per year, on model size and computing size. And then on the other hand, demand continues to grow in usage. Do we think that we need millions of GPUs? No doubt. Yeah.

Yeah, that is a for certainty now. Yeah. And so the question is, how do we architect it from a data center perspective? And that has a lot to do with, you know, are there data centers that are gigawatts at a time, or are they 250 megawatts at a time?

And my sense is that, you know, you're going to get both. I think analysts always focus on the current architectural bet. But I think one of the biggest takeaways from this conversation is that you're thinking about the entire ecosystem and many years out. So the idea that because NVIDIA is just scaling up or scaling out, it's to meet the future.

It's not such that you're only dependent on a world where there's a 500,000 or a million GPU cluster. It's, you know, by the time there's distributed training, you'll have written, you know, the software to enable that. That's right. Remember, without Megatron that we developed some seven years ago now, the scaling of these large training jobs wouldn't have happened. Right.

And so we invented Megatron, we invented Nickel, GPU Direct, right, all of the work that we did with RDMA. That made it possible for easily to do pipeline parallels, you know. right and so you know all the all the model parallelism that's being done you know all the breaking of the distributed training and all the batching and all that all of that stuff is is uh because we did the early work and now we're doing the early work for the future future generation so so let's talk about strawberry and no one yeah i want to be respectful of your time so we got all the time in the world actually well you're you're very generous yeah we've got all the world but First, I think it's cool that they named 01 after the 01 visa, right?

Which is about recruiting the world's best and brightest, you know, and bringing them to the United States. It's something I know we're both deeply passionate about. So I love the idea that building a model that thinks and that takes us to the next level of scaling intelligence, right, is an homage to the fact that it's these people who come to the United States by way of immigration that have made it. made us what we are, bring their collective intelligence to the United States. Surely an alien intelligence.

Certainly. It was spearheaded by our friend, Noam Brown, of course. He worked out in Pluribus and Cicero when he was at Meta.

How big a deal is inference time reasoning as a totally new vector of scaling intelligence, separate and distinct from just building larger models? It's a huge deal. It's a huge deal.

I think the... A lot of intelligence can't be done a priori. Right.

You know? And a lot of computing, even a lot of computing can't be reordered. I mean, just, you know, out-of-order execution can't be done a priori. You know?

And so a lot of things can only be done in runtime. Right. And. And so whether you think about it from a computer science perspective or you think about it from an intelligence perspective, too much of it requires context, the circumstance, the type of...

answer you're looking for, sometimes just a quick answer is good enough. Depends on the consequential impact of the answer. Depend on the nature of the usage of that answer. So some answers, please take a night. Some answers, take a week.

Yes. Is that right? So I could totally imagine me sending off a prompt to my AI and telling it, you know, think about it for a night. Right.

Think about it overnight. Don't tell me right away. Right. I want you to think about it all night.

And then come back and tell me tomorrow what's your best answer and reason about it for me. And so I think the quality, the segmentation of intelligence. now from a product perspective, there's going to be one-shot versions of it.

Right. For sure. Yeah.

And then there'll be some that take five minutes. And the intelligence layer that roots those questions to the right model for the right use case. I mean, we were using advanced voice mode and O1 preview last night. I was coaching my son for his AP history test. And it was like having the world's best AP history teacher sitting right next to you thinking, I mean...

about these questions. It was truly extraordinary. Again, they're-My tutor is an AI today.

Right. Of course, they're here today, which comes back to this, over 40% of your revenue today is inference. But inference is about ready because of chain of reasoning. Yeah. Right?

It's about ready-It's about to go up by a billion times. Right. By a million X, by a billion X.

That's right. That's the part that most people haven't completely internalized. This is that industry we were talking about, but this is the industrial revolution.

That's the production of intelligence. That's right. Right?

It's going to go up a billion times. Right. And so, you know, everybody's so hyper-focused on NVIDIA as kind of like doing training on bigger models.

Yeah. Right? Isn't it the case that your revenue, if it's 50-50 today, you're going to do way more inference in the future?

Yeah. Right. Then, I mean, training will always be important, but just the growth of inference is going to be way larger than the growth in training. We hope. It's almost impossible to conceive otherwise.

Yeah, we hope. That's right. That's right.

Right. I mean, it's good to go to school. Yes.

But the goal is so that you can be productive in society later. And so it's good that we train these models, but the goal is to inference them, you know. Are you already using chain of reasoning and, you know, tools like O1 in your own business? to improve your own business?

Yeah, our cybersecurity system today can't run without our own agents. We have agents helping to design chips. Hopper wouldn't be possible. Blackwell would be possible. Ruben, don't even think about it.

We have digital. We have AI chip designers, AI software engineers, AI verification engineers. And we build them all inside because we have the ability and we rather use the opportunity to explore the technology ourselves. When I walked into the building today, somebody came up to me and said, ask Jensen about the culture.

It's all about the culture. I look at the business. We talk a lot about fitness and efficiency, flat organizations that can execute quickly, smaller teams. NVIDIA is in a league of its own, really, at about $4 million of revenue per employee, about $2 million of profits or free cash flow per employee. You've built a culture of efficiency that really has unleashed creativity and innovation and ownership and responsibility.

You've broken the mold on kind of functional management. Everybody likes to talk about all of your direct reports. Is the leveraging of AI the thing that's going to continue to allow you to be hyper-creative while at the same time being efficient? No question.

I'm hoping that someday, NVIDIA has 32,000 employees today. Right. And we have 4,000 families in Israel. I hope they're well.

I'm thinking of you guys. Yes. And I'm hoping that NVIDIA someday will be a 50,000-employee company.

With 100 million AI assistants. Wow. and they're in every single group. We will have a whole directory of AIs that are just generally good at doing things.

We'll also have our inbox is going to full of directories of AIs that we work with that we know are really good specialized at our skill. AIs will recruit other AIs to solve problems. AIs will be in Slack channels with each other.

And with humans. Right, and with humans. And so we'll just be one large.

you know, employee base, if you will. Some of them are digital and AI, some of them are biological. And I'm hoping some of them are even in megatronics.

I think from a business perspective, it's something that's greatly misunderstood. You just described a company that's producing the output of a company with 150,000 people, but you're doing it with 50,000 people. That's right.

Now, you didn't say, I was going to get rid of all my employees. You're still growing the number of employees in the organization. But the output of that organization is going to be dramatically more. This is often misunderstood.

AI will change every job. AI will have a seismic impact on how people think about work. Let's acknowledge that. AI has the potential to do incredible good.

It has the potential to do harm. We have to build safe AI. Let's just make that foundational.

The part that is overlooked is when companies become more productive using artificial intelligence, it is likely that it manifests itself into either better earnings or better growth or both. Right. And when that happens, the next email from the CEO is likely not a layoff announcement. Right. Of course.

Because you're growing. Yeah. And the reason for that is because we have more ideas than we can explore. And we need people to help us think through it before we automate it. And so the automation part of it, AI can help us do.

Obviously, it's going to help us think through it as well. But it's still going to require us to go figure out what problems do I want to solve? There are a trillion things we can go solve. What problems does this company have to go solve? And select those ideas and figure out a way to automate and scale.

And so as a result, we're going to hire more people as we become more productive. People forget that, you know? And if you go back in time, obviously we have more ideas today than 200 years ago.

That's the reason why GDPs are larger and more people are employed. And even though we're automating like crazy underneath. It's such an important point of this period that we're entering.

One, almost all human productivity, almost all human prosperity is the byproduct. of the automation and the technology of the last 200 years. I mean, you can look at from Adam Smith and Schumpeter's creative destruction, you can look at chart of GDP growth per person over the course of the last 200 years, and it's just accelerated.

Which leads me to this question. If you look at the 90s, our productivity growth in the United States was about 2.5% to 3% a year. And then in the 2000s, it slowed down to about 1.8%.

So, And then the last 10 years has been the slowest productivity growth. So that's the amount of labor and capital, or the amount of output we have for a fixed amount of labor and capital. The slowest we've had on record, actually. And a lot of people have debated the reasoning for this.

But if the world is as you just described, and we're going to leverage and manufacture intelligence, then isn't it the case that we're on the verge of a dramatic expansion in terms of human productivity? That's our hope. Right. That's our hope. And of course...

You know, we live in this world, so we have direct evidence of it. Right. We have direct evidence of it, either as isolated of a case as an individual researcher.

For sure. Who is able to, with AI, now explore science at such an extraordinary scale that is unimaginable. That's productivity.

Right, 100%. Measure of productivity. Or that we're designing chips that are so incredible.

at such a high pace. And the chip complexities and the computer complexities we're building are going up exponentially while the company's employee base is not a measure of productivity. Correct. The software that we're developing better and better and better because we're using AI and supercomputers to help us.

The number of employees is growing barely linearly. Okay. Okay. Okay.

Another demonstration of productivity. So whether it's... I can spot check it in a whole bunch of different industries.

Yes. I could gut check it myself. Yes.

You're in business. That's right. And so I can, you know, and of course, you can't, we could be overfit. But the artistry, of course, is to generalize what is it that we're observing and whether this could manifest in other industries. And there's no question that intelligence is the single most valuable commodity the world's ever known.

And now we're going to manufacture it at scale. And we, all of us, have to get good at, you know, what would happen if you're surrounded by these AIs and they're doing things so incredibly well? and so much better than you.

Right. And when I reflect on that, that's my life. Right.

I have 60 direct reports. Right. The reason why they're on, the reason why they're on eStaff is because they're world-class at what they do and they do it better than I do. Right.

much better than I do. I have no trouble interacting with them. And I have no trouble prompt engineering them.

I have no trouble programming them. And so I think that that's the thing that that people are going to learn is that they're all going to be CEOs. Right.

They're all going to be CEOs of AI agents. Right. And their ability to have the creativity, the will, and some knowledge on how to reason, break problems down so that you can program these AIs to help you achieve something like I do. Right.

That's called running companies. Right. Now, you mentioned something, this alignment, this safe AI.

You mentioned the tragedy going on in the Middle East. We have a lot of autonomy and a lot of AI that's being used in different parts of the world. So let's talk for a second about bad actors, about safe AI, about coordination with Washington.

How do you feel today? Are we on the right path? Do we have a sufficient level of coordination? You know, I think Mark Zuckerberg has said the way we beat the bad AIs is we make the good AIs better.

How would you characterize your view of how we make sure that this is a positive benefit for humanity as opposed to, you know, leaving us in this dystopian world without purpose? The conversation about safety is really important and good. Yes.

The. The abstracted view, this conceptual view of AI being a large, giant neural network, not so good. Right, right. Okay.

And the reason for that is because, as we know, artificial intelligence and large language models are related, not the same. There are many things that are being done that I think are excellent. One, open sourcing models so that...

the entire community of researchers and every single industry and every single company can engage AI and go learn how to harness this capability for their application. Excellent. Number two, it is under-celebrated the amount of technology that is dedicated to inventing AI to keep AI safe. Yes. AIs to curate data, to curate information, to train an AI, AI created to align AI, synthetic data generation, AI to expand the knowledge of AI, to cause it to hallucinate less.

All of the AIs that are being created for vectorization or graphing or whatever it is, to inform an AI, guardrailing AI, AIs to monitor other AIs. The system of AIs. To create safe AI is under-celebrated. Right.

That we've already built. That we're building everybody all over the industry. The methodologies, the red teaming, the process, the model cards, the evaluation systems, the benchmarking systems.

All of the harnesses that are being built at the velocity that's been built is incredible. I wonder if they under celebrated. Do you guys understand? Yes.

And there's no, there's no government regulation saying you have to do this. Yeah. This is the actors in the space today who are building these AIs are taking seriously and coordinating around best practices with respect to these critical matters. That's right. Exactly.

And so, so that's under celebrated, under understood. Yes. Somebody needs to, to, to, well, everybody needs to start talking about AI as a system. of AIs and system of engineered systems, engineered systems that are well engineered, built from first principles, well tested, so on and so forth. Regulation.

Remember, remember AI is a capability that can be applied. And it's necessary to have regulation for important technologies. But it's also don't, don't.

Don't overreach to the point where some of the regulation ought to be done, most of the regulation ought to be done at the applications. Right. The FAA, NHTSA, FDA, you name it.

Right. All of the different ecosystems that already regulate applications of technology now have to regulate the application of technology that is now infused with AI. Right. And so I thinkā€¦ There's, don't, don't, don't misunderstand, don't overlook the overwhelming amount of regulation in the world that are going to have to be activated for AI.

And don't rely on just one universal galactic. AI council that's going to possibly be able to do this. Because there's a reason why all of these different agencies were created.

There's a reason why all these different regulatory bodies were created. We'll go back to first principles again. I'd get in trouble by my partner, Bill Gurley, if I didn't go back to the open source point. You guys launched a very important, very large, very capable open source model recently. Obviously...

Meta is making significant contributions to open source. I find when I read Twitter, you have this open versus closed, a lot of chatter about it. How do you feel about your own open source models, ability to keep up with Frontier? That would be the first question.

The second question would be, is that having that open source model and also having closed, source models that are powering commercial operations. Is that what you see into the future and do those two things, does that create the healthy tension for safety? Open source versus closed source is related to safety, but not only about safety.

So for example, there's absolutely nothing wrong with having closed source models that are the engines of an economic model. Exactly. necessary to sustain innovation.

Right. Okay. I celebrate that wholeheartedly. Right.

It is, I believe, wrong-minded to be closed versus open. Right. It should be closed and open.

Plus open. Yeah, right. Because open is necessary for many industries to be activated. Right now, if we didn't have open source, how would all these different fields of science be able to be activated on AI? Right.

Because they have to develop their own domain-specific AIs, and they have to develop their own, using open-source models, create domain-specific AIs. They're related, again, not the same. Just because you have an open-source model doesn't mean you have an AI.

And so you have to have that open-source model to enable the creation of AIs. So financial services, healthcare, transportation, the list of industries, fields of science, that has now been enabled as a result of open-source. Unbelievable.

Are you seeing a lot of demand for your open source models? Our open source models, so first of all, Lama. Downloads, right? Obviously, yeah, Mark and the work that they've done, incredible.

Off the charts! And it completely activated and engaged every single industry, every single field of science. The reason why we did Nemotron was for synthetic data generation. Intuitively, the idea that one AI would somehow sit there and loop and generate data to learn itself, it sounds brittle.

And how many times you can go around that infinite loop, that loop, you know, questionable. However, my mental image is kind of like you get a super smart person, put him into a padded room, close the door for about a month. You know, what comes out is probably not a smarter person.

But the idea that you could have two or three people sit around and we have different AIs. We have different distributions of knowledge, and we can go QA back and forth. All three of us can come out smarter.

And so the idea that you can have AI models exchanging, interacting, going back and forth, debating, reinforcement learning, synthetic data generation, for example, kind of intuitively suggests it makes sense. And so our model, Nemotron 350B, is... 340B is the best model in the world for reward systems.

And so it is the best critique. Okay. Interesting.

Yeah. And so a fantastic model for enhancing everybody else's models. Irrespective of how great somebody else's model is, I'd recommend using Nemotron 340B to enhance and make it better.

And we've already seen it made Lama better, made all the other models better. Well. We're coming to the end. Thank goodness. As somebody who delivered DGX-1 in 2016, it's really been an incredible journey.

Your journey is unlikely and incredible at the same time. Thank you. You survived. Just surviving the early days was pretty extraordinary.

You delivered the first DGX-1 in 2016. We had this Cambrian moment. 2022. And so I'm going to ask you the question I often get asked, which is, how long can you sustain what you're doing today? With 60 direct reports, you're everywhere. You're driving this revolution.

Are you having fun? And is there something else that you would rather be doing? Is this a question about the last hour and a half?

The answer is I had a great time. I had a great time. I couldn't imagine anything else I'd rather be doing. Let's see. I don't think it's right to leave the impression that our job...

It's fun all the time. My job isn't fun all the time. Nor do I expect it to be fun all the time.

Was that ever an expectation that it was fun all the time? I think it's important all the time. I don't take myself too seriously. I take the work very seriously.

I take our responsibility very seriously. I take our contribution and our moment in time very seriously. Is that always fun? No.

Yeah. But do I always love it? Yes. Yeah.

Like all things. You know, whether it's family, friends, children, is it always fun? No. Do we always love it?

Absolutely, deeply. And so I think the, how long can I do this? The real question is, how long can I be relevant?

And that only matters, that piece of information, that question can only be answered with, how am I going to continue to learn? I am a lot more optimistic today. I'm not saying this simply because of our topic today.

I'm a lot more optimistic about my ability to stay relevant and continue to learn because of AI. I use it. I don't know, but I'm sure you guys do. I use it literally every day.

There's not one piece of research that I don't involve AI with. There's not one question that even if I know the answer, I double check on it with AI. And surprisingly, you know, the next two or three questions I ask it reveals something I didn't know. You pick your topic.

You pick your topic. And I think that AI as a tutor, AI as an assistant. AI as a partner to brainstorm with, double check my work. Boy, you guys, it's completely revolutionary.

And that's just, I'm an information worker. My output is information. And so I think the contributions that I'll have on society is pretty extraordinary.

So I think if that's the case, if I could stay relevant like this. and I can continue to make a contribution. I know that the work is important enough for me to want to continue to pursue it.

And my quality of life is incredible. So I'll say, I can't imagine, you and I have been at this for a few decades. I can't imagine missing this moment.

It's the most consequential moment of our careers. We're deeply grateful for the partnership. Don't miss the next 10 years.

For the thought partnership. You make us smarter. Thank you.

And I think you're really important as part of the leadership, right, that's going to optimistically and safely lead this forward. So thank you for being with us. Really enjoyed it.

Thanks, Brad. Thanks, Clark. Good job. As a reminder to everybody, just our opinions, not investment advice.