Keynotes
Loading
Summary
Loading
Transcripts
[00:00]: At some point, you have to believe something.Â
We've reinvented computing as we know it. What
[00:03]: is the vision for what you see coming next? WeÂ
asked ourselves, if it can do this, how far can
[00:08]: it go? How do we get from the robots thatÂ
we have now to the future world that you
[00:13]: see? Cleo, everything that moves will beÂ
robotic someday and it will be soon. We
[00:17]: invested tens of billions of dollars beforeÂ
it really happened. No that's very good, you
[00:22]: did some research! But the big breakthroughÂ
I would say is when we...
[00:28]: That's Jensen Huang, and whether you know it or not
his decisions are shaping your future. He's the CEO of
[00:36]: NVIDIA, the company that skyrocketed over the past few
years to become one of the most valuable companies in
[00:41]: the world because they led a fundamental shiftÂ
in how computers work unleashing this current
[00:46]: explosion of what's possible with technology.Â
"NVIDIA's done it again!" We found ourselves being
[00:51]: one of the most important technology companies inÂ
the world and potentially ever. A huge amount of
[00:56]: the most futuristic tech that you're hearing about
in AI and robotics and gaming and self-driving
[01:01]: cars and breakthrough medical research relies onÂ
new chips and software designed by him and his
[01:06]: company. During the dozens of background interviewsÂ
that I did to prepare for this what struck me most
[01:10]: was how much Jensen Huang has already influencedÂ
all of our lives over the last 30 years, and how
[01:16]: many said it's just the beginning of somethingÂ
even bigger. We all need to know what he's building
[01:22]: and why and most importantly what he's tryingÂ
to build next. Welcome to Huge Conversations...
[01:36]: Thank you so much for doing this. I'm so happy to doÂ
it. Before we dive in, I wanted to tell you
[01:42]: how this interview is going to be a little bitÂ
different than other interviews I've seen you
[01:45]: do recently. Okay! I'm not going to ask you anyÂ
questions about - you could ask - company finances,
[01:51]: thank you! I'm not going to ask you questionsÂ
about your management style or why you don't
[01:55]: like one-on ones. I'm not going to ask youÂ
about regulations or politics. I think all
[02:01]: of those things are important but I think that ourÂ
audience can get them well covered elsewhere. Okay.
[02:06]: What we do on huge if true is we make optimisticÂ
explainer videos and we've covered - I'm the worst
[02:13]: person to be an explainer video. I think youÂ
might be the best and I think that's what I'm
[02:18]: really hoping that we can do together is make aÂ
joint explainer video about how can we actually
[02:25]: use technology to make the future better. Yeah. AndÂ
we do it because we believe that when people see
[02:30]: those better futures, they help build them. SoÂ
the people that you're going to be talking to
[02:33]: are awesome. They are optimists who want toÂ
build those better futures but because we
[02:39]: cover so many different topics, we've coveredÂ
supersonic planes and quantum computers and
[02:43]: particle colliders, it means that millionsÂ
of people come into every episode without
[02:48]: any prior knowledge whatsoever. You might beÂ
talking to an expert in their field who doesn't
[02:53]: know the difference between a CPU and a GPU or aÂ
12-year-old who might grow up one day to be you
[03:00]: but is just starting to learn. For my part,Â
I've now been preparing for this interview for
[03:06]: several months, including doing backgroundÂ
conversations with many members of your team
[03:11]: but I'm not an engineer. So my goal is to help thatÂ
audience see the future that you see so I'm going
[03:18]: to ask about three areas: The first is, how did weÂ
get here? What were the key insights that led to
[03:23]: this big fundamental shift in computing that we'reÂ
in now? The second is, what's actually happening
[03:29]: right now? How did those insights lead to the worldÂ
that we're now living in that seems like so much
[03:34]: is going on all at once? And the third is, what isÂ
the vision for what you see coming next? In order
[03:42]: to talk about this big moment we're in with AIÂ
I think we need to go back to video games in the
[03:48]: '90s. At the time I know game developers wantedÂ
to create more realistic looking graphics but
[03:56]: the hardware couldn't keep up with all of thatÂ
necessary math. NVIDIA came up with
[04:02]: a solution that would change not just gamesÂ
but computing itself. Could you take us back
[04:09]: there and explain what was happening and whatÂ
were the insights that led you and the NVIDIA
[04:15]: team to create the first modern GPU? So in theÂ
early '90s when we first started the company
[04:20]: we observed that in a software program insideÂ
it there are just a few lines of code, maybe
[04:27]: 10% of the code, does 99% % of the processingÂ
and that 99% of the processing could be done
[04:33]: in parallel. However the other 90% of the codeÂ
has to be done sequentially. It turns out that
[04:40]: the proper computer the perfect computer is oneÂ
that could do sequential processing and parallel
[04:45]: processing not just one or the other. That was theÂ
big observation and we set out to build a company
[04:52]: to solve computer problems that normal computersÂ
can't. And that's really the beginning of NVIDIA.
[05:00]: My favorite visual of why a CPU versus aÂ
GPU really matters so much is a 15-year-old
[05:05]: video on the NVIDIA YouTube channel where theÂ
Mythbusters, they use a little robot shooting
[05:11]: paintballs one by one to show solving problemsÂ
one at a time or sequential processing on a
[05:16]: CPU, but then they roll out this huge robotÂ
that shoots all of the paintballs at once
[05:24]: doing smaller problems all at the sameÂ
time or parallel processing on a GPU.
[05:30]: "3... 2... 1..." So Nvidia unlocks all of this new power
for video games. Why gaming first? The video games
[05:41]: requires parallel processing for processingÂ
3D graphics and we chose video games because,
[05:47]: one, we loved the application, it's a simulationÂ
of virtual worlds and who doesn't want to go to
[05:52]: virtual worlds and we had the good observationÂ
that video games has potential to be the largest
[05:58]: market for for entertainment ever. And it turnedÂ
out to be true. And having it being a large market
[06:04]: is important because the technology is complicatedÂ
and if we had a large market, our R&D budget could
[06:10]: be large, we could create new technology. And thatÂ
flywheel between technology and market and greater
[06:17]: technology was really the flywheel thatÂ
got NVIDIA to become one of the most important
[06:21]: technology companies in the world. It was allÂ
because of video games. I've heard you say that
[06:25]: GPUs were a time machine? Yeah. Could you tell meÂ
more about what you meant by that? A GPU is like a
[06:31]: time machine because it lets you see the futureÂ
sooner. One of the most amazing things anybody's
[06:37]: ever said to me was a quantum chemistryÂ
scientist. He said, Jensen, because of NVIDIA's work,
[06:46]: I can do my life's work in my lifetime. That's timeÂ
travel. He was able to do something that was beyond
[06:52]: his lifetime within his lifetime and this isÂ
because we make applications run so much faster
[07:00]: and you get to see the future. And so when you'reÂ
doing weather prediction for example, you're seeing
[07:05]: the future when you're doing a simulationÂ
a virtual city with virtual traffic and we're
[07:11]: simulating our self-driving car throughÂ
that virtual city, we're doing time travel. So
[07:17]: parallel processing takes off in gaming and it'sÂ
allowing us to create worlds in computers that
[07:24]: we never could have before and and gaming is sortÂ
of this this first incredible cas Cas of parallel
[07:30]: processing unlocking a lot more power and thenÂ
as you said people begin to use that power across
[07:37]: many different industries. The case of the of theÂ
quantum chemistry researcher, when I've heard you
[07:42]: tell that story it's that he was running molecularÂ
simulations in a way where it was much faster to
[07:49]: run in parallel on NVIDIA GPUs even then than itÂ
was to run them on the supercomputer with the CPU
[07:56]: that he had been using before. Yeah that's true. SoÂ
oh my god it's revolutionizing all of these other
[08:00]: industries as well, it's beginning to changeÂ
how we see what's possible with computers and my
[08:07]: understanding is that in the early 2000s youÂ
see this and you realize that actually doing
[08:14]: that is a little bit difficult because what thatÂ
researcher had to do is he had to sort of trick
[08:18]: the GPUs into thinking that his problem was aÂ
graphics problem. That's exactly right, no that's
[08:23]: very good, you did some research. So you createÂ
a way to make that a lot easier. That's right
[08:29]: Specifically it's a platform called CUDA whichÂ
lets programmers tell the GPU what to do using
[08:34]: programming languages that they already know likeÂ
C and that's a big deal because it gives way more
[08:39]: people easier access to all of this computingÂ
power. Could you explain what the vision was that
[08:44]: led you to create CUDA? Partly researchersÂ
discovering it, partly internal inspiration and
[08:53]: and partly solving a problem. And you know aÂ
lot of interesting interesting ideas come out
[09:00]: of that soup. You know some of it is aspirationÂ
and inspiration, some of it is just desperation you
[09:06]: know. And so in the case of CUDA is veryÂ
much this the same way and probably the first
[09:13]: external ideas of using our GPUs for parallelÂ
processing emerged out of some interesting work
[09:19]: in medical imaging a couple of researchersÂ
at Mass General were using it to do CT
[09:26]: reconstruction. They were using our graphicsÂ
processors for that reason and it inspired us.
[09:32]: Meanwhile the problem that we're trying to solveÂ
inside our company has to do with the fact that
[09:36]: when you're trying to create these virtual worldsÂ
for video games, you would like it to be beautiful
[09:41]: but also dynamic. Water should flow like water andÂ
explosions should be like explosions. So there's
[09:50]: particle physics you want to do, fluid dynamicsÂ
you want to do and that is much harder to do if
[09:56]: your pipeline is only able to do computer graphics.Â
And so we have a natural reason to want to do it
[10:02]: in the market that we were serving. SoÂ
researchers were also horsing around with using
[10:08]: our GPUs for general purpose uh acceleration andÂ
and so there there are multiple multiple factors
[10:13]: that were coming together in that soup, weÂ
just when the time came and we decided
[10:20]: to do something proper and created a CUDA asÂ
a result of that. Fundamentally the reason why
[10:25]: I was certain that CUDA was going to be successfulÂ
and we put the whole company behind it was
[10:31]: because fundamentally our GPU was going to beÂ
the highest volume parallel processors built in
[10:38]: the world because the market of video games was soÂ
large and so this architecture has a good chance
[10:43]: of reaching many people. It has seemed to me likeÂ
creating CUDA was this incredibly optimistic "huge
[10:51]: if true" thing to do where you were saying, if weÂ
create a way for many more people to use much
[10:58]: more computing power, they might create incredibleÂ
things. And then of course it came true. They did.
[11:04]: In 2012, a group of three researchers submits anÂ
entry to a famous competition where the goal is
[11:09]: to create computer systems that could recognizeÂ
images and label them with categories. And their
[11:14]: entry just crushes the competition. It gets wayÂ
fewer answers wrong. It was incredible. It blows
[11:20]: everyone away. It's called AlexNet, and it's a kindÂ
of AI called the neural network. My understanding
[11:24]: is one reason it was so good is that they usedÂ
a huge amount of data to train that system
[11:29]: and they did it on NVIDIA GPUs. All of a sudden,Â
GPUs weren't just a way to make computers faster
[11:35]: and more efficient they're becoming the enginesÂ
of a whole new way of computing. We're moving from
[11:40]: instructing computers with step-by-step directionsÂ
to training computers to learn by showing them a
[11:47]: huge number of examples. This moment in 2012 reallyÂ
kicked off this truly seismic shift that we're
[11:54]: all seeing with AI right now. Could you describeÂ
what that moment was like from your perspective
[11:59]: and what did you see it would mean for all ofÂ
our futures? When you create something new like
[12:06]: CUDA, if you build it, they might not come. And
that's always the cynic's perspective
[12:14]: however the optimist's perspective would say, butÂ
if you don't build it, they can't come. And that's
[12:20]: usually how we look at the world. You know weÂ
have to reason about intuitively why this would
[12:25]: be very useful. And in fac, in 2012 Ilya Sutskever,Â
and Alex Krizhevsky and Geoff Hinton in the University
[12:33]: of Toronto the lab that they were at they reachedÂ
out to a gForce GTX 580 because they learned about
[12:39]: CUDA and that CUDA might be able to to be used asÂ
a parallel processor for training AlexNet and
[12:45]: so our inspiration that GeForce could be the theÂ
vehicle to bring out this parallel architecture
[12:51]: into the world and that researchers would somehowÂ
find it someday was a good was a good strategy. It
[12:57]: was a strategy based on hope, but it was alsoÂ
reasoned hope. The thing that really caught
[13:03]: our attention was simultaneously we were tryingÂ
to solve the computer vision problem inside the
[13:07]: company and we were trying to get CUDA toÂ
be a good computer vision processor and we
[13:13]: were frustrated by a whole bunch of earlyÂ
developments internally with respect to our
[13:19]: computer vision effort and getting CUDA to beÂ
able to do it. And all of a sudden we saw AlexNet,
[13:25]: this new algorithm that is completelyÂ
different than computer vision algorithms before
[13:31]: it, take a giant leap in terms of capabilityÂ
for computer vision. And when we saw that it was
[13:38]: partly out of interest but partly because we wereÂ
struggling with something ourselves. And so we were
[13:43]: we were highly interested to want to see it work.Â
And so when we when we looked at AlexNet we were
[13:49]: inspired by that. But the big breakthrough IÂ
would say is when we when we saw AlexNet, we
[13:57]: asked ourselves you know, how far can AlexNetÂ
go? If it can do this with computer vision, how
[14:04]: far can it go? And if it if it could go to theÂ
limits of what we think it could go, the type
[14:11]: of problems it could solve, what would it mean forÂ
the computer industry? And what would it mean for
[14:16]: the computer architecture? And we were,Â
we rightfully reasoned that if machine learning,
[14:25]: if the deep learning architecture can scale,Â
the vast majority of machine learning problems
[14:30]: could be represented with deep neural networks. AndÂ
the type of problems we could solve with machine
[14:36]: learning is so vast that it has the potential ofÂ
reshaping the computer industry altogether,
[14:42]: which prompted us to re-engineer the entireÂ
computing stack which is where DGX came from
[14:49]: and this little baby DGX sitting here, all ofÂ
this came from from that observation that we ought
[14:56]: to reinvent the entire computing stack layer byÂ
layer by layer. You know computers, after 65 years
[15:03]: since IBM System 360 introduced modern generalÂ
purpose computing, we've reinvented computing as we
[15:09]: know it. To think about this as a whole story, soÂ
parallel processing reinvents modern gaming and
[15:16]: revolutionizes an entire industry then that wayÂ
of computing that parallel processing begins to
[15:22]: be used across different industries. You investÂ
in that by building CUDA and then CUDA and the
[15:29]: use of GPUs allows for a a step change in neuralÂ
networks and machine learning and begins a sort
[15:38]: of revolution that we're now seeing onlyÂ
increase in importance today... All of a sudden
[15:45]: computer vision is solved. All of a sudden speechÂ
recognition is solved. All of a sudden language
[15:50]: understanding is solved. These incredibleÂ
problems associated with intelligence one
[15:54]: by one by one by one where we had no solutionsÂ
for in past, desperate desire to have solutions
[16:01]: for, all of a sudden one after another get solvedÂ
you know every couple of years. It's incredible.
[16:07]: Yeah so you're seeing that, in 2012 you'reÂ
looking ahead and believing that that's
[16:12]: the future that you're going to be living in now,Â
and you're making bets that get you there, really
[16:17]: big bets that have very high stakes. And then myÂ
perception as a lay person is that it takes a
[16:22]: pretty long time to get there. You make these bets -Â
8 years, 10 years - so my question is:
[16:30]: If AlexNet that happened in 2012 and this audienceÂ
is probably seeing and hearing so much more about
[16:36]: AI and NVIDIA specifically 10 years later,Â
why did it take a decade and also because you
[16:43]: had placed those bets, what did the middleÂ
of that decade feel like for you? Wow that's
[16:48]: a good question. It probably felt like today. YouÂ
know to me, there's always some problem and
[16:55]: then there's some reason to be to beÂ
impatient. There's always some reason to be
[17:03]: happy about where you are and there's alwaysÂ
many reasons to carry on. And so I think as I
[17:09]: was reflecting a second ago, that sounds like thisÂ
morning! So but I would say that in all things that
[17:16]: we pursue, first you have to have core beliefs. YouÂ
have to reason from your best principles
[17:25]: and ideally you're reasoning from it from principlesÂ
of either physics or deep understanding of
[17:32]: the industry or deep understanding of theÂ
science, wherever you're reasoning from, you
[17:38]: reason from first principles. And at some point youÂ
have to believe something. And if those principles
[17:45]: don't change and the assumptions don't change,
then you, there's no reason to change your
[17:50]: core beliefs. And then along the way there's alwaysÂ
some evidence of you know of success and
[17:59]: and that you're leading in the rightÂ
direction and sometimes you know you go a
[18:04]: long time without evidence of success and youÂ
might have to course correct a little but
[18:08]: the evidence comes. And if you feel like you'reÂ
going in the right direction, we just keep on going.
[18:12]: The question of why did we stay so committed forÂ
so long, the answer is actually the opposite: There
[18:19]: was no reason to not be committed because we are,Â
we believed it. And I've believed in NVIDIA
[18:28]: for 30 plus years and I'm still here workingÂ
every single day. There's no fundamental
[18:34]: reason for me to change my belief system andÂ
I fundamentally believe that the
[18:39]: work we're doing in revolutionizing computingÂ
is as true today, even more true today than it
[18:43]: was before. And so we'll stickÂ
with it you know until otherwise. There's
[18:51]: of course very difficult times along the way. YouÂ
know when you're investing in something and nobody
[18:58]: else believes in it and cost a lot of money andÂ
you know maybe investors or or others would rather
[19:05]: you just keep the profit or you know whatever itÂ
is improve the share price or whatever it is.
[19:11]: But you have to believe in your future. You have toÂ
invest in yourself. And we believe this so
[19:17]: deeply that we invested you know tensÂ
of billions of dollars before it really
[19:25]: happened. And yeah it was, it was 10 longÂ
years. But it was fun along the way.
[19:32]: How would you summarize those core beliefs? WhatÂ
is it that you believe about the way computers
[19:38]: should work and what they can do for us that keepsÂ
you not only coming through that decade but also
[19:44]: doing what you're doing now, making bets I'm sureÂ
you're making for the next few decades? The first
[19:50]: core belief was our first discussion, was aboutÂ
accelerated computing. Parallel computing versus
[19:56]: general purpose computing. We would addÂ
two of those processors together and we would do
[20:00]: accelerated computing. And I continue to believeÂ
that today. The second was the recognition
[20:06]: that these deep learning networks, these DNNs, thatÂ
came to the public during 2012, these deep neural
[20:13]: networks have the ability to learn patterns andÂ
relationships from a whole bunch of different
[20:18]: types of data. And that it can learn more andÂ
more nuanced features if it could be larger
[20:24]: and larger. And it's easier to make them larger andÂ
larger, make them deeper and deeper or wider and
[20:29]: wider, and so the scalability of the architectureÂ
is empirically true. The fact
[20:40]: that model size and the data size being largerÂ
and larger, you can learn more knowledge is
[20:47]: also true, empirically true. And so if that'sÂ
the case, you could you know, what what are the
[20:55]: limits? There not, unless there's a physical limitÂ
or an architectural limit or mathematical limit
[21:00]: and it was never found, and so we believe that youÂ
could scale it. Then the question, the only other
[21:05]: question is: What can you learn from data? WhatÂ
can you learn from experience? Data is basically
[21:11]: digital versions of human experience. And so whatÂ
can you learn? You obviously can learn object
[21:17]: recognition from images. You can learn speechÂ
from just listening to sound. You can learn
[21:22]: even languages and vocabulary and syntax andÂ
grammar and all just by studying a whole bunch
[21:27]: of letters and words. So we've now demonstratedÂ
that AI or deep learning has the ability to learn
[21:33]: almost any modality of data and it can translateÂ
to any modality of data. And so what does that mean?
[21:42]: You can go from text to text, right, summarize aÂ
paragraph. You can go from text to text, translate
[21:49]: from language to language. You can go from textÂ
to images, that's image generation. You can go from
[21:55]: images to text, that's captioning. You can even goÂ
from amino acid sequences to protein structures.
[22:03]: In the future, you'll go from protein to words: "WhatÂ
does this protein do?" or "Give me an example of a
[22:11]: protein that has these properties." You knowÂ
identifying a drug target. And so you could
[22:17]: just see that all of these problems are aroundÂ
the corner to be solved. You can go from words
[22:24]: to video, why can't you go from words to actionÂ
tokens for a robot? You know from the computer's
[22:33]: perspective how is it any different? And so itÂ
it opened up this universe of opportunities and
[22:40]: universe of problems that we can go solve. AndÂ
that gets us quite excited. It feels like
[22:48]: we are on the cusp of this truly enormous change.Â
When I think about the next 10 years, unlike the
[22:56]: last 10 years, I know we've gone through a lot ofÂ
change already but I don't think I can predict
[23:02]: anymore how I will be using the technology that isÂ
currently being developed. That's exactly right. I
[23:07]: think the last 10, the reason why you feel that wayÂ
is, the last 10 years was really about the science
[23:12]: of AI. The next 10 years we're going to have plentyÂ
of science of AI but the next 10 years is going to
[23:18]: be the application science of AI. The fundamentalÂ
science versus the application science. And so the
[23:24]: the applied research, the application side of AIÂ
now becomes: How can I apply AI to digital biology?
[23:31]: How can I apply AI to climate technology? How canÂ
I apply AI to agriculture, to fishery, to robotics,
[23:39]: to transportation, optimizing logistics? How canÂ
I apply AI to you know teaching? How do I apply AI
[23:47]: to you know podcasting right? I'd love toÂ
choose a couple of those to help people see how
[23:53]: this fundamental change in computing that we'veÂ
been talking about is actually going to change
[23:58]: their experience of their lives, how they'reÂ
actually going to use technology that is based
[24:02]: on everything we just talked about. One of theÂ
things that I've now heard you talk a lot about
[24:07]: and I have a particular interest in is physicalÂ
AI. Or in other words, robots - "my friends!" - meaning
[24:16]: humanoid robots but also robots like self-drivingÂ
cars and smart buildings or autonomous warehouses
[24:23]: or autonomous lawnmowers or more. From whatÂ
I understand, we might be about to see a huge
[24:29]: leap in what all of these robots are capable ofÂ
because we're changing how we train them. Up until
[24:37]: recently you've either had to train your robot inÂ
the real world where it could get damaged or wear
[24:43]: down or you could get data from fairly limitedÂ
sources like humans in motion capture suits. But
[24:50]: that means that robots aren't getting as manyÂ
examples as they'd need to learn more quickly.
[24:56]: But now we're starting to train robots in digitalÂ
worlds, which means way more repetitions a day, way
[25:03]: more conditions, learning way faster. So we couldÂ
be in a big bang moment for robots right now and
[25:11]: NVIDIA is building tools to make that happen. YouÂ
have Omniverse and my understanding is this is 3D
[25:19]: worlds that help train robotic systems so thatÂ
they don't need to train in the physical world.
[25:26]: That's exactly right. You just just announcedÂ
Cosmos which is ways to make that 3D universe
[25:34]: much more realistic. So you can get all kindsÂ
of different, if we're training something on
[25:39]: this table, many different kinds of lighting on theÂ
table, many different times of day, many different
[25:44]: you know experiences for the robot to go throughÂ
so that it can get even more out of Omniverse. As
[25:52]: a kid who grew up loving Data on Star Trek, IsaacÂ
Asimov's books and just dreaming about a future with
[26:00]: robots, how do we get from the robots that we haveÂ
now to the future world that you see of robotics?
[26:08]: Yeah let me use language models maybe ChatGPTÂ
as a reference for understanding Omniverse and
[26:17]: Cosmos. So first of all when ChatGPT firstÂ
came out it, it was extraordinary and
[26:24]: it has the ability to do to basically fromÂ
your prompt, generate text. However, as amazing as
[26:32]: it was, it has the tendency to hallucinate if
it goes on too long or if it pontificates about
[26:40]: a topic it you know is not informed about, it'llÂ
still do a good job generating plausible answers.
[26:46]: It just wasn't grounded in the truth. And so
people called it hallucination. And
[26:55]: so the next generation shortly it was, it hadÂ
the ability to be conditioned by context, so
[27:03]: you could upload your PDF and now it's groundedÂ
by the PDF. The PDF becomes the ground truth. It
[27:09]: could be it could actually look up search andÂ
then the search becomes its ground truth. And
[27:14]: between that it could reason about what is howÂ
to produce the answer that you're asking for. And
[27:21]: so the first part is a generative AI and theÂ
second part is ground truth. Okay and so now let's
[27:28]: come into the the physical world. The
world model, we need a foundation model just like
[27:35]: we need ChatGPT had a core foundation modelÂ
that was the breakthrough in order for robotics
[27:41]: to to be smart about the physical world. It has toÂ
understand things like gravity, friction, inertia,
[27:50]: geometric and spatial awareness. It has to uhÂ
understand that an object is sitting there even
[27:57]: when I looked away when I come back it's stillÂ
sitting there, object permanence. It has to
[28:02]: understand cause and effect. If I tip it, it'llÂ
fall over. And so these kind of physical
[28:08]: common sense if you will has to be captured orÂ
encoded into a world foundation model so that
[28:16]: the AI has world common sense. Okay and so weÂ
have to go, somebody has to go create that, and
[28:23]: that's what we did with Cosmos. We created a worldÂ
language model. Just like ChatGPT was a language model,
[28:29]: this is a world model. The second thing we have toÂ
go do is we have to do the same thing that we did
[28:35]: with PDFs and context and grounding it withÂ
ground truth. And so the way we augment Cosmos
[28:42]: with ground truth is with physical simulations,Â
because Omniverse uses physics simulation which
[28:49]: is based on principled solvers. The mathematicsÂ
is Newtonian physics is the, right, it's the math we
[28:56]: know, all of the the fundamental laws ofÂ
physics we've understood for a very long
[29:02]: time. And it's encoded into, captured into Omniverse.Â
That's why Omniverse is a simulator. And using the
[29:09]: simulator to ground or to condition Cosmos, we canÂ
now generate an infinite number of stories of the
[29:19]: future. And they're grounded on physical truth. JustÂ
like between PDF or search plus ChatGPT, we can
[29:30]: generate an infinite amount of interesting things,Â
answer a whole bunch of interesting questions. The
[29:37]: combination of Omniverse plus Cosmos, you couldÂ
do that for the physical world. So to illustrate
[29:43]: this for the audience, if you had a robot in aÂ
factory and you wanted to make it learn every
[29:49]: route that it could take, instead of manuallyÂ
going through all of those routes, which could
[29:53]: take days and could be a lot of wear and tear onÂ
the robot, we're now able to simulate all of them
[29:59]: digitally in a fraction of the time and in manyÂ
different situations that the robot might face -
[30:04]: it's dark, it's blocked it's etc - so the robotÂ
is now learning much much faster. It seems to
[30:10]: me like the future might look very different thanÂ
today. If you play this out 10 years, how do you see
[30:17]: people actually interacting with this technologyÂ
in the near future? Cleo, everything that moves
[30:22]: will be robotic someday and it will be soon. YouÂ
know the the idea that you'll be pushing around
[30:28]: a lawn mower is already kind of silly. You knowÂ
maybe people do it because because it's fun but
[30:35]: but there's no need to. And every car isÂ
going to be robotic. Humanoid robots, the technology
[30:44]: necessary to make it possible, is just aroundÂ
the corner. And so everything that moves will be
[30:50]: robotic and they'll learn how to beÂ
a robot in Omniverse Cosmos and we'll generate
[30:59]: all these plausible, physically plausible futuresÂ
and the the robots will learn from them and
[31:05]: then they'll come into the physical world and youÂ
know it's exactly the same. A future where
[31:11]: you're just surrounded by robots is for certain.Â
And I'm just excited about having my own R2-D2.
[31:18]: And of course R2-D2 wouldn't be quite the can thatÂ
it is and roll around. It'll be you know R2-D2
[31:25]: yeah, it'll probably be a different physicalÂ
embodiment, but it's always R2. You know so my R2
[31:32]: is going to go around with me. Sometimes it's in myÂ
smart glasses, sometimes it's in my phone, sometimes
[31:36]: it's in my PC. It's in my car. So R2 is with meÂ
all the time including you know when I get home
[31:43]: you know where I left a physical version of R2. AndÂ
you know whatever that version happens to
[31:49]: be you know, we'll interact with R2. And so IÂ
think the idea that we'll have our own R2-D2 for
[31:55]: our entire life and it grows up with us, that'sÂ
a certainty now yeah. I think a lot of news media
[32:05]: when they talk about futures like this they focusÂ
on what could go wrong. And that makes sense. There
[32:10]: is a lot that could go wrong. We should talk aboutÂ
what could go wrong so we could keep it from from
[32:14]: going wrong. Yeah that's the approach that we likeÂ
to take on the show is, what are the big challenges
[32:19]: so that we can overcome them? Yeah. What buckets doÂ
you think about when you're worrying about this
[32:24]: future? Well there's a whole bunch of theÂ
stuff that everybody talks about: Bias or toxicity
[32:30]: or just hallucination. You know speaking withÂ
great confidence about something it knows nothing
[32:37]: about and as a result we rely on that information.Â
Generating, that's a version of generating
[32:45]: fake information, fake news or fake imagesÂ
or whatever it is. Of course impersonation.
[32:50]: It does such a good job pretending to be aÂ
human, it could be it could do an incredibly good
[32:56]: job pretending to be a specific human. And so
the spectrum of areas we
[33:05]: have to be concerned about is fairly clear andÂ
there's a lot of people who are
[33:11]: working on it. There's a some of the stuff,Â
some of the stuff related to AI safety requires
[33:18]: deep research and deep engineering andÂ
that's simply, it wants to do the right thing it
[33:24]: just didn't perform it right and as a result hurtÂ
somebody. You know for example self-driving car
[33:29]: that wants to drive nicely and drive properlyÂ
and just somehow the sensor broke down or it
[33:36]: didn't detect something. Or you know made itÂ
too aggressive turn or whatever it is. It did
[33:41]: it poorly. It did it wrongly. And so that's
a whole bunch of engineering that has to
[33:47]: be done to to make sure that AI safety is upheldÂ
by making sure that the product functions properly.
[33:54]: And then and then lastly you know whatever whatÂ
happens if the system, the AI wants to do a good
[34:00]: job but the system failed. Meaning the AI wantedÂ
to stop something from happening
[34:07]: and it turned out just when it wanted to doÂ
it, the machine broke down. And so this is
[34:13]: no different than a flight computer insideÂ
a plane having three versions of them and then
[34:19]: so there's triple redundancy inside theÂ
system inside autopilots and then you have two
[34:25]: pilots and then you have air traffic controlÂ
and then you have other pilots watching out for
[34:31]: these pilots. And so that the AI safetyÂ
systems has to be architected as a community
[34:38]: such that such that these AIs one, work,
function properly. When they don't
[34:47]: function properly, they don't put people in harm'sÂ
way. And that they're sufficient safety and
[34:52]: security systems all around them to make sureÂ
that we keep AI safe. And so there's
[34:58]: this spectrum of conversation is gigantic and andÂ
you know we have to take the parts, take the
[35:05]: parts apart and and build them as engineers. OneÂ
of the incredible things about this moment that
[35:11]: we're in right now is that we no longer have aÂ
lot of the technological limits that we had in a
[35:17]: world of CPUs and sequential processing. And we'veÂ
unlocked not only a new way to do computing and
[35:28]: and but also a way to continue to improve. ParallelÂ
processing has a a different kind of physics to it
[35:35]: than the improvements that we were able to makeÂ
on CPUs. I'm curious, what are the scientific or
[35:42]: technological limitations that we face now inÂ
the current world that you're thinking a lot
[35:47]: about? Well everything in the end is about how muchÂ
work you can get done within the limitations of
[35:54]: the energy that you have. And so that'sÂ
a physical limit and the laws of
[36:02]: physics about transporting information andÂ
transporting bits, flipping bits and transporting
[36:11]: bits, at the end of the day the energy it takesÂ
to do that limits what we can get done. And the
[36:18]: amount of energy that we have limits what we canÂ
get done. We're far from having any fundamental
[36:23]: limits that keep us from advancing. In the meantime,Â
we seek to build better and more energy efficient
[36:29]: computers. This little computer, the theÂ
big version of it was $250,000 - Pick up? - Yeah
[36:38]: Yeah that's little baby DIGITS yeah. This isÂ
an AI supercomputer. The version that I delivered,
[36:46]: this is just a prototype so it's a mockup.
The very first version was DGX 1, I
[36:52]: delivered to Open AI in 2016 and that was $250,000.Â
10,000 times more power, more energy necessary
[37:03]: than this version and this version has six timesÂ
more performance. I know, it's incredible. We're
[37:09]: in a whole in the world. And it's only since 2016Â
and so eight years later we've in increased the
[37:16]: energy efficiency of computing by 10,000 times.Â
And imagine if we became 10,000 times more energy
[37:25]: efficient or if a car was 10,000 times moreÂ
energy efficient or electric light bulb was
[37:31]: 10,000 times more energy efficient. Our lightÂ
bulb would be right now instead of 100 Watts,
[37:38]: 10,000 times less producing the same illumination.Â
Yeah and so the energy efficiency of
[37:45]: computing particularly for AI computing that we'veÂ
been working on has advanced incredibly and that's
[37:51]: essential because we want to create youÂ
know more intelligent systems and and we want to
[37:56]: use more computation to be smarter and soÂ
energy efficiency to do the work is our number one
[38:03]: priority. When I was preparing for this interview, IÂ
spoke to a lot of my engineering friends and this
[38:09]: is a question that they really wanted me to ask. SoÂ
you're really speaking to your people here. You've
[38:15]: shown a value of increasing accessibilityÂ
and abstraction, with CUDA and allowing more
[38:21]: people to use more computing power in all kinds ofÂ
other ways. As applications of technology get more
[38:28]: specific, I'm thinking of transformers in AI forÂ
example... For the audience, a transformer is a very
[38:35]: popular more recent structure of AI that's nowÂ
used in a huge number of the tools that you've
[38:40]: seen. The reason that they're popular is becauseÂ
transformers are structured in a way that helps
[38:45]: them pay "attention" to key bits of information andÂ
give much better results. You could build chips
[38:51]: that are perfectly suited for just one kind of AIÂ
model, but if you do that then you're making them
[38:56]: less able to do other things. So as these specificÂ
structures or architectures of AI get more popular,
[39:03]: my understanding is there's a debate between howÂ
much you place these bets on "burning them into the
[39:09]: chip" or designing hardware that is very specificÂ
to a certain task versus staying more general and
[39:15]: so my question is, how do you make those bets? HowÂ
do you think about whether the solution is a car
[39:22]: that could go anywhere or it's really optimizingÂ
a train to go from A to B? You're making bets
[39:28]: with huge stakes and I'm curious how you thinkÂ
about that. Yeah and that now comes back
[39:33]: to exactly your question, what are yourÂ
core beliefs? And the question, the core
[39:41]: belief either one, that transformer is the last AIÂ
algorithm, AI architecture that any researcher will
[39:52]: ever discover again, or that transformersÂ
is a stepping stone towards evolutions of
[40:01]: transformers that are uh barely recognizable as aÂ
transformer years from now. And we believe the
[40:08]: latter. And the reason for that is because youÂ
just have to go back in history and ask yourself,
[40:14]: in the world of computer algorithms, inÂ
the world of software, in the world of
[40:20]: engineering and innovation, has one idea stayedÂ
along that long? And the answer is no. And so that's
[40:27]: kind of the beauty, that's in factÂ
the essential beauty of a computer that it's able
[40:34]: to do something today that no one even imaginedÂ
possible 10 years ago. And if you would have, if
[40:41]: you would have turned that computer 10 years agoÂ
into a microwave, then why would the applications
[40:48]: keep coming? And so we believe, we believe in theÂ
richness of innovation and the
[40:54]: richness of invention and we want to create anÂ
architecture that let inventors and innovators
[40:59]: and software programmers and AI researchersÂ
swim in the soup and come up with some amazing
[41:05]: ideas. Look at transformers. The fundamentalÂ
characteristic of a transformer is this idea
[41:10]: called "attention mechanism" and it basically saysÂ
the transformer is going to understand the meaning
[41:16]: and the relevance of every single word with everyÂ
other word. So if you had 10 words, it has to figure
[41:22]: out the relationship across 10 of them. But if youÂ
have a 100,000 words or if your context is
[41:27]: now as large as, read a PDF and that read a wholeÂ
bunch of PDFs, and the context window is now like
[41:35]: a million tokens, the processing all of it acrossÂ
all of it is just impossible. And so the way you
[41:42]: solve that problem is there all kinds of new ideas,Â
flash attention or hierarchical attention or you
[41:49]: know all the, wave attention I just read aboutÂ
the other day. The number of different types of
[41:54]: attention mechanisms that have been inventedÂ
since the transformer is quite extraordinary.
[42:00]: And so I think that that's going to continueÂ
and we believe it's going to continue and that
[42:06]: that computer science hasn't ended and that AIÂ
research have not all given up and we haven't
[42:12]: given up anyhow and that having aÂ
computer that enables the flexibility of
[42:21]: of research and innovation and new ideas isÂ
fundamentally the most important thing. One of the
[42:29]: things that I am just so curious about, you designÂ
the chips. There are companies that assemble the
[42:37]: chips. There are companies that design hardware toÂ
make it possible to work at nanometer scale. When
[42:44]: you're designing tools like this, how do you thinkÂ
about design in the context of what's physically
[42:51]: possible right now to make? What are the thingsÂ
that you're thinking about with sort of pushing
[42:56]: that limit today? The way we do it is evenÂ
though even though we have things made like for
[43:05]: example our chips are made by TSMC. Even thoughÂ
we have them made by TSMC, we assume that we need
[43:13]: to have the deep expertise that TSMC has. And soÂ
we have people in our company who are incredibly
[43:19]: good at semiconductive physics so that we have aÂ
feeling for, we have an intuition for, what are the
[43:25]: limits of what today's semiconductor physicsÂ
can do. And then we work very closely with them to
[43:32]: discover the limits because we're trying to pushÂ
the limits and so we discover the limits together.
[43:36]: Now we do the same thing in system engineering andÂ
cooling systems. It turns out plumbing is really
[43:41]: important to us because of liquid cooling.Â
And maybe fans are really important to us
[43:44]: because of air cooling and we're trying to designÂ
these fans in a way almost like you know they're
[43:49]: aerodynamically sound so that we could pass theÂ
highest volume of air, make the least amount of
[43:54]: noise. So we have aerodynamics engineers in our
company. And so even though even though we don't
[44:01]: make 'em, we design them and we have to deepÂ
expertise of knowing how to have them made. And
[44:09]: and from that we try to push theÂ
limits. One of the themes of this conversation is
[44:18]: that you are a person who makes big bets on theÂ
future and time and time again you've been right
[44:25]: about those bets. We've talked about GPUs, we'veÂ
talked about CUDA, we've talked about bets you've
[44:30]: made in AI - self-driving cars, and we're going toÂ
be right on robotics and - this is my question. What
[44:37]: are the bets you're making now? the latest bet we
just described at the CES and I'm very very proud
[44:43]: of it and I'm very excited about it is theÂ
fusion of Omniverse and Cosmos so that we have
[44:50]: this new type of generative world generationÂ
system, this multiverse generation system. I
[44:59]: think that's going to be profoundly important inÂ
the future of robotics and physical systems.
[45:06]: Of course the work that we're doing with humanÂ
robots, developing the tooling systems and the
[45:11]: training systems and the human demonstrationÂ
systems and all of this stuff that that you've
[45:17]: already mentioned, we're just seeing theÂ
beginnings of that work and I think the
[45:23]: next 5 years are going to be very interesting inÂ
the world of human robotics. Of course the work
[45:28]: that we're doing in digital biology so thatÂ
we can understand the language of molecules and
[45:34]: understand the language of cells and just asÂ
we understand the language of physics and the
[45:39]: physical world we'd like to understand the languageÂ
of the human body and understand the language of
[45:44]: biology. And so if we can learn that, and we canÂ
predict it. Then all of a sudden our ability to
[45:50]: have a digital twin of the human is plausible.Â
And so I'm very excited about that work. I love
[45:56]: the work that we're doing in climate scienceÂ
and be able to, from weather predictions, understand
[46:03]: and predict the high resolution regional climates,Â
the weather patterns within a kilometer above
[46:10]: your head. That we can somehow predict that withÂ
great accuracy, its implications is really quite
[46:17]: profound. And so the number of things thatÂ
we're working on is really cool. You know we
[46:24]: we're fortunate that we've created thisÂ
this instrument that is a time machine and
[46:37]: we need time machines in all of these areas thatÂ
we just talked about so that we can see
[46:43]: the future. And if we could see the future andÂ
we can predict the future then we have a better
[46:48]: chance of making that future the best versionÂ
of it. And that's the reason why scientists
[46:53]: want to predict the future. That's the reason why,Â
that's the reason why we try to predict the future
[46:58]: and everything that we try to design so that weÂ
can optimize for the best version. So if
[47:05]: someone is watching this and maybe they came intoÂ
this video knowing that NVIDIA is an incredibly
[47:12]: important company but not fully understanding whyÂ
or how it might affect their life and they're now
[47:18]: hopefully better understanding a big shift thatÂ
we've gone through over the last few decades in
[47:23]: computing, this very exciting, very sort of strangeÂ
moment that we're in right now, where we're sort
[47:30]: of on the precipice of so many different things.Â
If they would like to be able to look into the
[47:36]: future a little bit, how would you advise them toÂ
prepare or to think about this moment that they're
[47:42]: in personally with respect to how these toolsÂ
are actually going to affect them? Well there are
[47:49]: several ways to reason about the future thatÂ
we're creating. One way to reason about it is,
[47:57]: suppose the work that you do continues toÂ
be important but the effort by which you
[48:04]: do it went from you know being a week longÂ
to almost instantaneous. You know that the
[48:15]: effort of drudgery basically goes to zero.Â
What is the implication of that? This is, this
[48:23]: is very similar to what would change if allÂ
of a sudden we had highways in this country?
[48:30]: And that kind of happened you know in the lastÂ
Industrial Revolution, all of a sudden we have
[48:34]: interstate highways and when you have interstateÂ
highways what happens? Well you know suburbs start
[48:40]: to be created and and all of a sudden you knowÂ
distribution of goods from east to west is
[48:48]: no longer a concern and all of a sudden gasÂ
stations start cropping up on highways and
[48:55]: and fast food restaurants show up and youÂ
know someone, some motels show up because people
[49:03]: you know traveling across the state, across theÂ
country and just wanted to stay somewhere for a
[49:07]: few hours or overnight, and so all of a suddenÂ
new economies and new capabilities, new economies.
[49:13]: What would happen if a video conference madeÂ
it possible for us to see each other without
[49:19]: having to travel anymore? All of a suddenÂ
it's actually okay to work further away from
[49:24]: home and from work, work and liveÂ
further away. And so you ask yourself kind of
[49:32]: these questions. You know what would happenÂ
if I have a software programmer with me
[49:40]: all the time and whatever it is I can dream up,Â
the software programmer could write for me. You
[49:46]: know what would, what would happenÂ
if I just had a seed of an idea and
[49:54]: and I rough it out and all of sudden a you knowÂ
a prototype of a production was put in front
[50:01]: of me? And what how would that change my life andÂ
how would that change my opportunity? And you
[50:07]: know what does it free me to be able to do andÂ
and so on so forth. And so I think that the next
[50:13]: the next decade intelligence, not for everythingÂ
but for for some things, would basically become
[50:22]: superhuman. But I can tellÂ
you exactly what that feels like. I'm surrounded
[50:31]: by superhuman people, super intelligence fromÂ
my perspective because they're the best in the
[50:38]: world at what they do and they do what theyÂ
do way better than I can do it. and I'm
[50:46]: surrounded by thousands of them and yet what itÂ
it never one day caused me to to think all of a
[50:56]: son I'm no longer necessary. It actually empowersÂ
me and gives me the confidence to go tackle more
[51:05]: and more ambitious things. And so suppose,Â
suppose now everybody is surrounded by these
[51:13]: super AIs that are very good at specific thingsÂ
or good at some of the things. What would that
[51:20]: make you feel? Well it's going to empower you,Â
it's going to make you feel confident and
[51:25]: and I'm pretty sure you probably use ChatGPT andÂ
AI and I feel more empowered today, more
[51:32]: confident to learn something today. The knowledgeÂ
of almost any particular field, the barriers to
[51:38]: that understanding, it has been reduced and I haveÂ
a personal tutor with me all of the time. And
[51:44]: so I think that that feeling should be universal.
If there's one thing that I would
[51:50]: encourage everybody to do is to go get yourselfÂ
an AI tutor right away. And that AI tutor could
[51:56]: of course just teach your things, anything youÂ
like, help you program, help you write,
[52:03]: help you analyze, help you think, help you reason,Â
you know all of those things is going to
[52:10]: really make you feel empowered and and I thinkÂ
that going to be our future. We're
[52:16]: going to become, we're going to become super humans,Â
not because we have super, we're going to become
[52:21]: super humans because we have super AIs. Could youÂ
tell us a little bit about each of these objects?
[52:27]: This is a new GeForce graphics card and yes andÂ
this is the RTX 50 Series. It is essentially
[52:39]: a supercomputer that you put into your PC and weÂ
use it for gaming, of course people today use it
[52:45]: for design and creative arts and it does amazingÂ
AI. The real breakthrough here and this is
[52:52]: this is truly an amazing thing, GeForceÂ
enabled AI and it enabled Geoff Hinton, Ilya Sutskever,
[52:59]: Alex Krizhevsky to be able to train AlexNet. We
discovered AI and we advanced AI then AI came back
[53:07]: to GeForce to help computer graphics. And so here'sÂ
the amazing thing: Out of 8 million pixels or so in
[53:16]: a 4K display we are computing, we're processingÂ
only 500,000 of them. The rest of them we use AI
[53:24]: to predict. The AI guessed it and yet the image isÂ
perfect. We inform it by the 500,000 pixels that we
[53:32]: computed and we ray traced every single one and it'sÂ
all beautiful. It's perfect. And then we tell the
[53:38]: AI, if these are the 500,000 perfect pixels in thisÂ
screen, what are the other 8 million? And it goes it
[53:44]: fills in the rest of the screen and it's perfect.
And if you only have to do fewer pixels, are you
[53:50]: able to invest more in doing that because you haveÂ
fewer to do so then the quality is better so the
[53:58]: extrapolation that the AI does... Exactly. BecauseÂ
whatever computing, whatever attention you have,
[54:03]: whatever resources you have, you can place it intoÂ
500,000 pixels. Now this is a perfect example of
[54:11]: why AI is going to make us all superhuman, becauseÂ
all of the other things that it can do, it'll do
[54:17]: for us, allows us to take our time and energy andÂ
focus it on the really really valuable things that
[54:23]: we do. And so we'll take our own resource which isÂ
you know energy intensive, attention intensive, and
[54:33]: we'll dedicated to the few 100,000 pixels andÂ
use AI to superres, upres it you know to
[54:39]: everything else. And so this this graphics cardÂ
is now powered mostly by AI and the computer
[54:47]: graphics technology inside is incredible asÂ
well. And then this next one, as I mentioned
[54:52]: earlier, in 2016 I built the first one for AIÂ
researchers and we delivered the first one to Open AI
[54:58]: and Elon was there to receive it and thisÂ
version I built a mini mini version and the
[55:06]: reason for that is because AI has now gone from AIÂ
researchers to every engineer, every student, every
[55:15]: AI scientist. And AI is going to be everywhere.Â
And so instead of these $250,000 versions we're
[55:21]: going to make these $3,000 versions and schoolsÂ
can have them, you know students can have them, and
[55:28]: you set it next to your PC or Mac and all ofÂ
a sudden you have your own AI supercomputer. And
[55:36]: you could develop and build AIs. Build your ownÂ
AI, build your own R2-D2. What do you feel like is
[55:42]: important for this audience to know that I haven'tÂ
asked? One of the most important things I would
[55:48]: advise is for example if I were a student todayÂ
the first thing I would do is to learn AI. How do
[55:54]: I learn to interact with ChatGPT, how do I learnÂ
to interact with Gemini Pro, and how do I learn
[56:00]: to interact with Grok? Learning how to
interact with with AI is not unlike being
[56:10]: someone who is really good at asking questions.Â
You're incredibly good at asking questions and
[56:17]: and prompting AI is very very similar.
You can't just randomly ask a bunch of questions
[56:23]: and so asking an AI to be assistantÂ
to you requires some expertise and
[56:30]: artistry and how to prompt it. And so if I were,Â
if I were a student today, irrespective whether
[56:35]: it's for for math or for science or chemistryÂ
or biology or doesn't matter what field of science
[56:40]: I'm going to go into or what profession, I'mÂ
going to ask myself, how can I use AI to do my job
[56:46]: better? If I want to be a lawyer, how can I useÂ
AI to be a better lawyer? If I want to be a better
[56:50]: do doctor, how can I use AI to be a better doctor?Â
If I want to be a chemist, how do I use AI to be
[56:55]: a better chemist? If I want to be a biologist, I howÂ
do I use AI to be a better biologist? That question
[57:02]: should be persistent across everybody. And just asÂ
my generation grew up as the first generation
[57:10]: that has to ask ourselves, how can we use computersÂ
to do our jobs better? Yeah the generation before
[57:17]: us had no computers, my generation was the firstÂ
generation that had to ask the question, how do I
[57:23]: use computers to do my job better? Remember I cameÂ
into the industry before Windows 95 right, 1984
[57:32]: there were no computers in offices. And after that,Â
shortly after that, computers started to emerge and
[57:38]: so we had to ask ourselves how do we use computersÂ
to do our jobs better? The next generation doesn't
[57:45]: have to ask that question but it has to askÂ
obviously next question, how can I use AI to
[57:49]: do my job better? That is start and finish I thinkÂ
for everybody. It's a really exciting and scary and
[57:59]: therefore worthwhile question I think for everyone.Â
I think it's going to be incredibly fun. AI is
[58:04]: obviously a word that people are just learningÂ
now but it's just you know, it's
[58:10]: made your computer so much more accessible. It isÂ
easier to prompt ChatGPT to ask it anything you
[58:15]: like than to go do the research yourself. And soÂ
we've lowered a barrier of understanding, we've
[58:22]: lowered a barrier of knowledge, we'veÂ
lowered a barrier of intelligence, and
[58:26]: and everybody really had to just go tryÂ
it. You know the thing that's really really crazy
[58:32]: is if I put a computer in front of somebody andÂ
they've never used a computer there is no chance
[58:37]: they're going to learn that computer in a day.
There's just no chance. Somebody really has to
[58:43]: show it to you and yet with ChatGPT if youÂ
don't know how to use it, all you have to do is
[58:49]: type in "I don't know how to use ChatGPT, tellÂ
me," and it would come back and give you some
[58:54]: examples and so that's the amazing thing.
You know the amazing thing about intelligence is
[59:02]: it'll help you along the way and make you uhÂ
superhuman you know along the way. All right I have
[59:08]: one more question if you have a second. This isÂ
not something that I planned to ask you but on the
[59:13]: way here, I'm a little bit afraid of planes,Â
which is not my most reasonable quality, and
[59:21]: the flight here was a little bit bumpy mhm veryÂ
bumpy and I'm sitting there and it's moving and
[59:30]: I'm thinking about what they're going to say at myÂ
funeral and after - She asked good questions, that's
[59:37]: what the tombstone's going to say - IÂ
hope so! Yeah. And after I loved my husband and my
[59:44]: friends and my family, the thing that I hoped thatÂ
they would talk about was optimism. I hope that
[59:49]: they would recognize what I'm trying to do here.Â
And I'm very curious for you, you've you've been
[59:56]: doing this a long time, it feels like there'sÂ
so much that you've described in this vision
[01:00:00]: ahead, what would the theme be that you wouldÂ
want people to say about what you're trying to do?
[01:00:14]: Very simply, they made an extraordinary impact.Â
I think that we're fortunate because of some
[01:00:23]: core beliefs a long time ago and sticking withÂ
those core beliefs and building upon them
[01:00:32]: we found ourselves today being one ofÂ
the most, one of the many most important and
[01:00:42]: consequential technology companies in
the world and potentially ever. And so
[01:00:49]: we take that responsibility very seriously.
We work hard to make sure that
[01:00:56]: the capabilities that we've created areÂ
available to large companies as well as
[01:01:03]: individual researchers and developers, acrossÂ
every field of science no matter profitable or
[01:01:10]: not, big or small, famous or otherwise.
And it's because of this understanding of
[01:01:21]: the consequential work that we're doing and theÂ
potential impact it has on so many people
[01:01:27]: that we want to make make this capabilityÂ
as pervasively as possible and I
[01:01:37]: do think that when we look back in a fewÂ
years, and I do hope that what the
[01:01:47]: next generation realized is as they, wellÂ
first of all they're going to know us because of
[01:01:53]: all the you know gaming technology we create.
I do think that we'll look back and the whole
[01:01:59]: field of digital biology and life sciences hasÂ
been transformed. Our whole understanding of of
[01:02:06]: material sciences has completely beenÂ
revolutionized. That robots are helping
[01:02:13]: us do dangerous and mundane things all over theÂ
place. That if we wanted to drive we can drive
[01:02:19]: but otherwise you know take a nap or enjoyÂ
your car like it's a home theater of yours,
[01:02:26]: you know read from work to home and at thatÂ
point you're hoping that you live far
[01:02:31]: away and so you could be in a car for longer.Â
And you look back and
[01:02:37]: you realize that there's this company almost atÂ
the epicenter of all of that and happens
[01:02:43]: to be the company that
you grew up playing games with.
[01:02:46]: I hope for that to be
what the next generation learn.
[01:02:50]: Thank you so much for your time.
I enjoyed it, thank you! I'm glad!
Flashcards
1 / 3
Loading
Click to reveal answer
Loading
💡 Explanation
Loading
Quizzes
JavaScript Fundamentals Quiz
Test your knowledge of JavaScript basics
180s
Question 2
Which method is used to add an element to the end of an array?
A. push()
B. add()
C. append()
D. insert()
Unlock Interactive Quizzes
Create quizzes from your content in one clickUsers get instant feedback. You get clear accuracy, time, strength & weakness insights and VEGA's next-step recommendations.
🎯 Sign Up for Free to Create Quizzes✓ $10 free AI credit on signup✓ No credit card required
Keynotes
Loading
Summary
Loading
Transcripts
[00:00]: At some point, you have to believe something.Â
We've reinvented computing as we know it. What
[00:03]: is the vision for what you see coming next? WeÂ
asked ourselves, if it can do this, how far can
[00:08]: it go? How do we get from the robots thatÂ
we have now to the future world that you
[00:13]: see? Cleo, everything that moves will beÂ
robotic someday and it will be soon. We
[00:17]: invested tens of billions of dollars beforeÂ
it really happened. No that's very good, you
[00:22]: did some research! But the big breakthroughÂ
I would say is when we...
[00:28]: That's Jensen Huang, and whether you know it or not
his decisions are shaping your future. He's the CEO of
[00:36]: NVIDIA, the company that skyrocketed over the past few
years to become one of the most valuable companies in
[00:41]: the world because they led a fundamental shiftÂ
in how computers work unleashing this current
[00:46]: explosion of what's possible with technology.Â
"NVIDIA's done it again!" We found ourselves being
[00:51]: one of the most important technology companies inÂ
the world and potentially ever. A huge amount of
[00:56]: the most futuristic tech that you're hearing about
in AI and robotics and gaming and self-driving
[01:01]: cars and breakthrough medical research relies onÂ
new chips and software designed by him and his
[01:06]: company. During the dozens of background interviewsÂ
that I did to prepare for this what struck me most
[01:10]: was how much Jensen Huang has already influencedÂ
all of our lives over the last 30 years, and how
[01:16]: many said it's just the beginning of somethingÂ
even bigger. We all need to know what he's building
[01:22]: and why and most importantly what he's tryingÂ
to build next. Welcome to Huge Conversations...
[01:36]: Thank you so much for doing this. I'm so happy to doÂ
it. Before we dive in, I wanted to tell you
[01:42]: how this interview is going to be a little bitÂ
different than other interviews I've seen you
[01:45]: do recently. Okay! I'm not going to ask you anyÂ
questions about - you could ask - company finances,
[01:51]: thank you! I'm not going to ask you questionsÂ
about your management style or why you don't
[01:55]: like one-on ones. I'm not going to ask youÂ
about regulations or politics. I think all
[02:01]: of those things are important but I think that ourÂ
audience can get them well covered elsewhere. Okay.
[02:06]: What we do on huge if true is we make optimisticÂ
explainer videos and we've covered - I'm the worst
[02:13]: person to be an explainer video. I think youÂ
might be the best and I think that's what I'm
[02:18]: really hoping that we can do together is make aÂ
joint explainer video about how can we actually
[02:25]: use technology to make the future better. Yeah. AndÂ
we do it because we believe that when people see
[02:30]: those better futures, they help build them. SoÂ
the people that you're going to be talking to
[02:33]: are awesome. They are optimists who want toÂ
build those better futures but because we
[02:39]: cover so many different topics, we've coveredÂ
supersonic planes and quantum computers and
[02:43]: particle colliders, it means that millionsÂ
of people come into every episode without
[02:48]: any prior knowledge whatsoever. You might beÂ
talking to an expert in their field who doesn't
[02:53]: know the difference between a CPU and a GPU or aÂ
12-year-old who might grow up one day to be you
[03:00]: but is just starting to learn. For my part,Â
I've now been preparing for this interview for
[03:06]: several months, including doing backgroundÂ
conversations with many members of your team
[03:11]: but I'm not an engineer. So my goal is to help thatÂ
audience see the future that you see so I'm going
[03:18]: to ask about three areas: The first is, how did weÂ
get here? What were the key insights that led to
[03:23]: this big fundamental shift in computing that we'reÂ
in now? The second is, what's actually happening
[03:29]: right now? How did those insights lead to the worldÂ
that we're now living in that seems like so much
[03:34]: is going on all at once? And the third is, what isÂ
the vision for what you see coming next? In order
[03:42]: to talk about this big moment we're in with AIÂ
I think we need to go back to video games in the
[03:48]: '90s. At the time I know game developers wantedÂ
to create more realistic looking graphics but
[03:56]: the hardware couldn't keep up with all of thatÂ
necessary math. NVIDIA came up with
[04:02]: a solution that would change not just gamesÂ
but computing itself. Could you take us back
[04:09]: there and explain what was happening and whatÂ
were the insights that led you and the NVIDIA
[04:15]: team to create the first modern GPU? So in theÂ
early '90s when we first started the company
[04:20]: we observed that in a software program insideÂ
it there are just a few lines of code, maybe
[04:27]: 10% of the code, does 99% % of the processingÂ
and that 99% of the processing could be done
[04:33]: in parallel. However the other 90% of the codeÂ
has to be done sequentially. It turns out that
[04:40]: the proper computer the perfect computer is oneÂ
that could do sequential processing and parallel
[04:45]: processing not just one or the other. That was theÂ
big observation and we set out to build a company
[04:52]: to solve computer problems that normal computersÂ
can't. And that's really the beginning of NVIDIA.
[05:00]: My favorite visual of why a CPU versus aÂ
GPU really matters so much is a 15-year-old
[05:05]: video on the NVIDIA YouTube channel where theÂ
Mythbusters, they use a little robot shooting
[05:11]: paintballs one by one to show solving problemsÂ
one at a time or sequential processing on a
[05:16]: CPU, but then they roll out this huge robotÂ
that shoots all of the paintballs at once
[05:24]: doing smaller problems all at the sameÂ
time or parallel processing on a GPU.
[05:30]: "3... 2... 1..." So Nvidia unlocks all of this new power
for video games. Why gaming first? The video games
[05:41]: requires parallel processing for processingÂ
3D graphics and we chose video games because,
[05:47]: one, we loved the application, it's a simulationÂ
of virtual worlds and who doesn't want to go to
[05:52]: virtual worlds and we had the good observationÂ
that video games has potential to be the largest
[05:58]: market for for entertainment ever. And it turnedÂ
out to be true. And having it being a large market
[06:04]: is important because the technology is complicatedÂ
and if we had a large market, our R&D budget could
[06:10]: be large, we could create new technology. And thatÂ
flywheel between technology and market and greater
[06:17]: technology was really the flywheel thatÂ
got NVIDIA to become one of the most important
[06:21]: technology companies in the world. It was allÂ
because of video games. I've heard you say that
[06:25]: GPUs were a time machine? Yeah. Could you tell meÂ
more about what you meant by that? A GPU is like a
[06:31]: time machine because it lets you see the futureÂ
sooner. One of the most amazing things anybody's
[06:37]: ever said to me was a quantum chemistryÂ
scientist. He said, Jensen, because of NVIDIA's work,
[06:46]: I can do my life's work in my lifetime. That's timeÂ
travel. He was able to do something that was beyond
[06:52]: his lifetime within his lifetime and this isÂ
because we make applications run so much faster
[07:00]: and you get to see the future. And so when you'reÂ
doing weather prediction for example, you're seeing
[07:05]: the future when you're doing a simulationÂ
a virtual city with virtual traffic and we're
[07:11]: simulating our self-driving car throughÂ
that virtual city, we're doing time travel. So
[07:17]: parallel processing takes off in gaming and it'sÂ
allowing us to create worlds in computers that
[07:24]: we never could have before and and gaming is sortÂ
of this this first incredible cas Cas of parallel
[07:30]: processing unlocking a lot more power and thenÂ
as you said people begin to use that power across
[07:37]: many different industries. The case of the of theÂ
quantum chemistry researcher, when I've heard you
[07:42]: tell that story it's that he was running molecularÂ
simulations in a way where it was much faster to
[07:49]: run in parallel on NVIDIA GPUs even then than itÂ
was to run them on the supercomputer with the CPU
[07:56]: that he had been using before. Yeah that's true. SoÂ
oh my god it's revolutionizing all of these other
[08:00]: industries as well, it's beginning to changeÂ
how we see what's possible with computers and my
[08:07]: understanding is that in the early 2000s youÂ
see this and you realize that actually doing
[08:14]: that is a little bit difficult because what thatÂ
researcher had to do is he had to sort of trick
[08:18]: the GPUs into thinking that his problem was aÂ
graphics problem. That's exactly right, no that's
[08:23]: very good, you did some research. So you createÂ
a way to make that a lot easier. That's right
[08:29]: Specifically it's a platform called CUDA whichÂ
lets programmers tell the GPU what to do using
[08:34]: programming languages that they already know likeÂ
C and that's a big deal because it gives way more
[08:39]: people easier access to all of this computingÂ
power. Could you explain what the vision was that
[08:44]: led you to create CUDA? Partly researchersÂ
discovering it, partly internal inspiration and
[08:53]: and partly solving a problem. And you know aÂ
lot of interesting interesting ideas come out
[09:00]: of that soup. You know some of it is aspirationÂ
and inspiration, some of it is just desperation you
[09:06]: know. And so in the case of CUDA is veryÂ
much this the same way and probably the first
[09:13]: external ideas of using our GPUs for parallelÂ
processing emerged out of some interesting work
[09:19]: in medical imaging a couple of researchersÂ
at Mass General were using it to do CT
[09:26]: reconstruction. They were using our graphicsÂ
processors for that reason and it inspired us.
[09:32]: Meanwhile the problem that we're trying to solveÂ
inside our company has to do with the fact that
[09:36]: when you're trying to create these virtual worldsÂ
for video games, you would like it to be beautiful
[09:41]: but also dynamic. Water should flow like water andÂ
explosions should be like explosions. So there's
[09:50]: particle physics you want to do, fluid dynamicsÂ
you want to do and that is much harder to do if
[09:56]: your pipeline is only able to do computer graphics.Â
And so we have a natural reason to want to do it
[10:02]: in the market that we were serving. SoÂ
researchers were also horsing around with using
[10:08]: our GPUs for general purpose uh acceleration andÂ
and so there there are multiple multiple factors
[10:13]: that were coming together in that soup, weÂ
just when the time came and we decided
[10:20]: to do something proper and created a CUDA asÂ
a result of that. Fundamentally the reason why
[10:25]: I was certain that CUDA was going to be successfulÂ
and we put the whole company behind it was
[10:31]: because fundamentally our GPU was going to beÂ
the highest volume parallel processors built in
[10:38]: the world because the market of video games was soÂ
large and so this architecture has a good chance
[10:43]: of reaching many people. It has seemed to me likeÂ
creating CUDA was this incredibly optimistic "huge
[10:51]: if true" thing to do where you were saying, if weÂ
create a way for many more people to use much
[10:58]: more computing power, they might create incredibleÂ
things. And then of course it came true. They did.
[11:04]: In 2012, a group of three researchers submits anÂ
entry to a famous competition where the goal is
[11:09]: to create computer systems that could recognizeÂ
images and label them with categories. And their
[11:14]: entry just crushes the competition. It gets wayÂ
fewer answers wrong. It was incredible. It blows
[11:20]: everyone away. It's called AlexNet, and it's a kindÂ
of AI called the neural network. My understanding
[11:24]: is one reason it was so good is that they usedÂ
a huge amount of data to train that system
[11:29]: and they did it on NVIDIA GPUs. All of a sudden,Â
GPUs weren't just a way to make computers faster
[11:35]: and more efficient they're becoming the enginesÂ
of a whole new way of computing. We're moving from
[11:40]: instructing computers with step-by-step directionsÂ
to training computers to learn by showing them a
[11:47]: huge number of examples. This moment in 2012 reallyÂ
kicked off this truly seismic shift that we're
[11:54]: all seeing with AI right now. Could you describeÂ
what that moment was like from your perspective
[11:59]: and what did you see it would mean for all ofÂ
our futures? When you create something new like
[12:06]: CUDA, if you build it, they might not come. And
that's always the cynic's perspective
[12:14]: however the optimist's perspective would say, butÂ
if you don't build it, they can't come. And that's
[12:20]: usually how we look at the world. You know weÂ
have to reason about intuitively why this would
[12:25]: be very useful. And in fac, in 2012 Ilya Sutskever,Â
and Alex Krizhevsky and Geoff Hinton in the University
[12:33]: of Toronto the lab that they were at they reachedÂ
out to a gForce GTX 580 because they learned about
[12:39]: CUDA and that CUDA might be able to to be used asÂ
a parallel processor for training AlexNet and
[12:45]: so our inspiration that GeForce could be the theÂ
vehicle to bring out this parallel architecture
[12:51]: into the world and that researchers would somehowÂ
find it someday was a good was a good strategy. It
[12:57]: was a strategy based on hope, but it was alsoÂ
reasoned hope. The thing that really caught
[13:03]: our attention was simultaneously we were tryingÂ
to solve the computer vision problem inside the
[13:07]: company and we were trying to get CUDA toÂ
be a good computer vision processor and we
[13:13]: were frustrated by a whole bunch of earlyÂ
developments internally with respect to our
[13:19]: computer vision effort and getting CUDA to beÂ
able to do it. And all of a sudden we saw AlexNet,
[13:25]: this new algorithm that is completelyÂ
different than computer vision algorithms before
[13:31]: it, take a giant leap in terms of capabilityÂ
for computer vision. And when we saw that it was
[13:38]: partly out of interest but partly because we wereÂ
struggling with something ourselves. And so we were
[13:43]: we were highly interested to want to see it work.Â
And so when we when we looked at AlexNet we were
[13:49]: inspired by that. But the big breakthrough IÂ
would say is when we when we saw AlexNet, we
[13:57]: asked ourselves you know, how far can AlexNetÂ
go? If it can do this with computer vision, how
[14:04]: far can it go? And if it if it could go to theÂ
limits of what we think it could go, the type
[14:11]: of problems it could solve, what would it mean forÂ
the computer industry? And what would it mean for
[14:16]: the computer architecture? And we were,Â
we rightfully reasoned that if machine learning,
[14:25]: if the deep learning architecture can scale,Â
the vast majority of machine learning problems
[14:30]: could be represented with deep neural networks. AndÂ
the type of problems we could solve with machine
[14:36]: learning is so vast that it has the potential ofÂ
reshaping the computer industry altogether,
[14:42]: which prompted us to re-engineer the entireÂ
computing stack which is where DGX came from
[14:49]: and this little baby DGX sitting here, all ofÂ
this came from from that observation that we ought
[14:56]: to reinvent the entire computing stack layer byÂ
layer by layer. You know computers, after 65 years
[15:03]: since IBM System 360 introduced modern generalÂ
purpose computing, we've reinvented computing as we
[15:09]: know it. To think about this as a whole story, soÂ
parallel processing reinvents modern gaming and
[15:16]: revolutionizes an entire industry then that wayÂ
of computing that parallel processing begins to
[15:22]: be used across different industries. You investÂ
in that by building CUDA and then CUDA and the
[15:29]: use of GPUs allows for a a step change in neuralÂ
networks and machine learning and begins a sort
[15:38]: of revolution that we're now seeing onlyÂ
increase in importance today... All of a sudden
[15:45]: computer vision is solved. All of a sudden speechÂ
recognition is solved. All of a sudden language
[15:50]: understanding is solved. These incredibleÂ
problems associated with intelligence one
[15:54]: by one by one by one where we had no solutionsÂ
for in past, desperate desire to have solutions
[16:01]: for, all of a sudden one after another get solvedÂ
you know every couple of years. It's incredible.
[16:07]: Yeah so you're seeing that, in 2012 you'reÂ
looking ahead and believing that that's
[16:12]: the future that you're going to be living in now,Â
and you're making bets that get you there, really
[16:17]: big bets that have very high stakes. And then myÂ
perception as a lay person is that it takes a
[16:22]: pretty long time to get there. You make these bets -Â
8 years, 10 years - so my question is:
[16:30]: If AlexNet that happened in 2012 and this audienceÂ
is probably seeing and hearing so much more about
[16:36]: AI and NVIDIA specifically 10 years later,Â
why did it take a decade and also because you
[16:43]: had placed those bets, what did the middleÂ
of that decade feel like for you? Wow that's
[16:48]: a good question. It probably felt like today. YouÂ
know to me, there's always some problem and
[16:55]: then there's some reason to be to beÂ
impatient. There's always some reason to be
[17:03]: happy about where you are and there's alwaysÂ
many reasons to carry on. And so I think as I
[17:09]: was reflecting a second ago, that sounds like thisÂ
morning! So but I would say that in all things that
[17:16]: we pursue, first you have to have core beliefs. YouÂ
have to reason from your best principles
[17:25]: and ideally you're reasoning from it from principlesÂ
of either physics or deep understanding of
[17:32]: the industry or deep understanding of theÂ
science, wherever you're reasoning from, you
[17:38]: reason from first principles. And at some point youÂ
have to believe something. And if those principles
[17:45]: don't change and the assumptions don't change,
then you, there's no reason to change your
[17:50]: core beliefs. And then along the way there's alwaysÂ
some evidence of you know of success and
[17:59]: and that you're leading in the rightÂ
direction and sometimes you know you go a
[18:04]: long time without evidence of success and youÂ
might have to course correct a little but
[18:08]: the evidence comes. And if you feel like you'reÂ
going in the right direction, we just keep on going.
[18:12]: The question of why did we stay so committed forÂ
so long, the answer is actually the opposite: There
[18:19]: was no reason to not be committed because we are,Â
we believed it. And I've believed in NVIDIA
[18:28]: for 30 plus years and I'm still here workingÂ
every single day. There's no fundamental
[18:34]: reason for me to change my belief system andÂ
I fundamentally believe that the
[18:39]: work we're doing in revolutionizing computingÂ
is as true today, even more true today than it
[18:43]: was before. And so we'll stickÂ
with it you know until otherwise. There's
[18:51]: of course very difficult times along the way. YouÂ
know when you're investing in something and nobody
[18:58]: else believes in it and cost a lot of money andÂ
you know maybe investors or or others would rather
[19:05]: you just keep the profit or you know whatever itÂ
is improve the share price or whatever it is.
[19:11]: But you have to believe in your future. You have toÂ
invest in yourself. And we believe this so
[19:17]: deeply that we invested you know tensÂ
of billions of dollars before it really
[19:25]: happened. And yeah it was, it was 10 longÂ
years. But it was fun along the way.
[19:32]: How would you summarize those core beliefs? WhatÂ
is it that you believe about the way computers
[19:38]: should work and what they can do for us that keepsÂ
you not only coming through that decade but also
[19:44]: doing what you're doing now, making bets I'm sureÂ
you're making for the next few decades? The first
[19:50]: core belief was our first discussion, was aboutÂ
accelerated computing. Parallel computing versus
[19:56]: general purpose computing. We would addÂ
two of those processors together and we would do
[20:00]: accelerated computing. And I continue to believeÂ
that today. The second was the recognition
[20:06]: that these deep learning networks, these DNNs, thatÂ
came to the public during 2012, these deep neural
[20:13]: networks have the ability to learn patterns andÂ
relationships from a whole bunch of different
[20:18]: types of data. And that it can learn more andÂ
more nuanced features if it could be larger
[20:24]: and larger. And it's easier to make them larger andÂ
larger, make them deeper and deeper or wider and
[20:29]: wider, and so the scalability of the architectureÂ
is empirically true. The fact
[20:40]: that model size and the data size being largerÂ
and larger, you can learn more knowledge is
[20:47]: also true, empirically true. And so if that'sÂ
the case, you could you know, what what are the
[20:55]: limits? There not, unless there's a physical limitÂ
or an architectural limit or mathematical limit
[21:00]: and it was never found, and so we believe that youÂ
could scale it. Then the question, the only other
[21:05]: question is: What can you learn from data? WhatÂ
can you learn from experience? Data is basically
[21:11]: digital versions of human experience. And so whatÂ
can you learn? You obviously can learn object
[21:17]: recognition from images. You can learn speechÂ
from just listening to sound. You can learn
[21:22]: even languages and vocabulary and syntax andÂ
grammar and all just by studying a whole bunch
[21:27]: of letters and words. So we've now demonstratedÂ
that AI or deep learning has the ability to learn
[21:33]: almost any modality of data and it can translateÂ
to any modality of data. And so what does that mean?
[21:42]: You can go from text to text, right, summarize aÂ
paragraph. You can go from text to text, translate
[21:49]: from language to language. You can go from textÂ
to images, that's image generation. You can go from
[21:55]: images to text, that's captioning. You can even goÂ
from amino acid sequences to protein structures.
[22:03]: In the future, you'll go from protein to words: "WhatÂ
does this protein do?" or "Give me an example of a
[22:11]: protein that has these properties." You knowÂ
identifying a drug target. And so you could
[22:17]: just see that all of these problems are aroundÂ
the corner to be solved. You can go from words
[22:24]: to video, why can't you go from words to actionÂ
tokens for a robot? You know from the computer's
[22:33]: perspective how is it any different? And so itÂ
it opened up this universe of opportunities and
[22:40]: universe of problems that we can go solve. AndÂ
that gets us quite excited. It feels like
[22:48]: we are on the cusp of this truly enormous change.Â
When I think about the next 10 years, unlike the
[22:56]: last 10 years, I know we've gone through a lot ofÂ
change already but I don't think I can predict
[23:02]: anymore how I will be using the technology that isÂ
currently being developed. That's exactly right. I
[23:07]: think the last 10, the reason why you feel that wayÂ
is, the last 10 years was really about the science
[23:12]: of AI. The next 10 years we're going to have plentyÂ
of science of AI but the next 10 years is going to
[23:18]: be the application science of AI. The fundamentalÂ
science versus the application science. And so the
[23:24]: the applied research, the application side of AIÂ
now becomes: How can I apply AI to digital biology?
[23:31]: How can I apply AI to climate technology? How canÂ
I apply AI to agriculture, to fishery, to robotics,
[23:39]: to transportation, optimizing logistics? How canÂ
I apply AI to you know teaching? How do I apply AI
[23:47]: to you know podcasting right? I'd love toÂ
choose a couple of those to help people see how
[23:53]: this fundamental change in computing that we'veÂ
been talking about is actually going to change
[23:58]: their experience of their lives, how they'reÂ
actually going to use technology that is based
[24:02]: on everything we just talked about. One of theÂ
things that I've now heard you talk a lot about
[24:07]: and I have a particular interest in is physicalÂ
AI. Or in other words, robots - "my friends!" - meaning
[24:16]: humanoid robots but also robots like self-drivingÂ
cars and smart buildings or autonomous warehouses
[24:23]: or autonomous lawnmowers or more. From whatÂ
I understand, we might be about to see a huge
[24:29]: leap in what all of these robots are capable ofÂ
because we're changing how we train them. Up until
[24:37]: recently you've either had to train your robot inÂ
the real world where it could get damaged or wear
[24:43]: down or you could get data from fairly limitedÂ
sources like humans in motion capture suits. But
[24:50]: that means that robots aren't getting as manyÂ
examples as they'd need to learn more quickly.
[24:56]: But now we're starting to train robots in digitalÂ
worlds, which means way more repetitions a day, way
[25:03]: more conditions, learning way faster. So we couldÂ
be in a big bang moment for robots right now and
[25:11]: NVIDIA is building tools to make that happen. YouÂ
have Omniverse and my understanding is this is 3D
[25:19]: worlds that help train robotic systems so thatÂ
they don't need to train in the physical world.
[25:26]: That's exactly right. You just just announcedÂ
Cosmos which is ways to make that 3D universe
[25:34]: much more realistic. So you can get all kindsÂ
of different, if we're training something on
[25:39]: this table, many different kinds of lighting on theÂ
table, many different times of day, many different
[25:44]: you know experiences for the robot to go throughÂ
so that it can get even more out of Omniverse. As
[25:52]: a kid who grew up loving Data on Star Trek, IsaacÂ
Asimov's books and just dreaming about a future with
[26:00]: robots, how do we get from the robots that we haveÂ
now to the future world that you see of robotics?
[26:08]: Yeah let me use language models maybe ChatGPTÂ
as a reference for understanding Omniverse and
[26:17]: Cosmos. So first of all when ChatGPT firstÂ
came out it, it was extraordinary and
[26:24]: it has the ability to do to basically fromÂ
your prompt, generate text. However, as amazing as
[26:32]: it was, it has the tendency to hallucinate if
it goes on too long or if it pontificates about
[26:40]: a topic it you know is not informed about, it'llÂ
still do a good job generating plausible answers.
[26:46]: It just wasn't grounded in the truth. And so
people called it hallucination. And
[26:55]: so the next generation shortly it was, it hadÂ
the ability to be conditioned by context, so
[27:03]: you could upload your PDF and now it's groundedÂ
by the PDF. The PDF becomes the ground truth. It
[27:09]: could be it could actually look up search andÂ
then the search becomes its ground truth. And
[27:14]: between that it could reason about what is howÂ
to produce the answer that you're asking for. And
[27:21]: so the first part is a generative AI and theÂ
second part is ground truth. Okay and so now let's
[27:28]: come into the the physical world. The
world model, we need a foundation model just like
[27:35]: we need ChatGPT had a core foundation modelÂ
that was the breakthrough in order for robotics
[27:41]: to to be smart about the physical world. It has toÂ
understand things like gravity, friction, inertia,
[27:50]: geometric and spatial awareness. It has to uhÂ
understand that an object is sitting there even
[27:57]: when I looked away when I come back it's stillÂ
sitting there, object permanence. It has to
[28:02]: understand cause and effect. If I tip it, it'llÂ
fall over. And so these kind of physical
[28:08]: common sense if you will has to be captured orÂ
encoded into a world foundation model so that
[28:16]: the AI has world common sense. Okay and so weÂ
have to go, somebody has to go create that, and
[28:23]: that's what we did with Cosmos. We created a worldÂ
language model. Just like ChatGPT was a language model,
[28:29]: this is a world model. The second thing we have toÂ
go do is we have to do the same thing that we did
[28:35]: with PDFs and context and grounding it withÂ
ground truth. And so the way we augment Cosmos
[28:42]: with ground truth is with physical simulations,Â
because Omniverse uses physics simulation which
[28:49]: is based on principled solvers. The mathematicsÂ
is Newtonian physics is the, right, it's the math we
[28:56]: know, all of the the fundamental laws ofÂ
physics we've understood for a very long
[29:02]: time. And it's encoded into, captured into Omniverse.Â
That's why Omniverse is a simulator. And using the
[29:09]: simulator to ground or to condition Cosmos, we canÂ
now generate an infinite number of stories of the
[29:19]: future. And they're grounded on physical truth. JustÂ
like between PDF or search plus ChatGPT, we can
[29:30]: generate an infinite amount of interesting things,Â
answer a whole bunch of interesting questions. The
[29:37]: combination of Omniverse plus Cosmos, you couldÂ
do that for the physical world. So to illustrate
[29:43]: this for the audience, if you had a robot in aÂ
factory and you wanted to make it learn every
[29:49]: route that it could take, instead of manuallyÂ
going through all of those routes, which could
[29:53]: take days and could be a lot of wear and tear onÂ
the robot, we're now able to simulate all of them
[29:59]: digitally in a fraction of the time and in manyÂ
different situations that the robot might face -
[30:04]: it's dark, it's blocked it's etc - so the robotÂ
is now learning much much faster. It seems to
[30:10]: me like the future might look very different thanÂ
today. If you play this out 10 years, how do you see
[30:17]: people actually interacting with this technologyÂ
in the near future? Cleo, everything that moves
[30:22]: will be robotic someday and it will be soon. YouÂ
know the the idea that you'll be pushing around
[30:28]: a lawn mower is already kind of silly. You knowÂ
maybe people do it because because it's fun but
[30:35]: but there's no need to. And every car isÂ
going to be robotic. Humanoid robots, the technology
[30:44]: necessary to make it possible, is just aroundÂ
the corner. And so everything that moves will be
[30:50]: robotic and they'll learn how to beÂ
a robot in Omniverse Cosmos and we'll generate
[30:59]: all these plausible, physically plausible futuresÂ
and the the robots will learn from them and
[31:05]: then they'll come into the physical world and youÂ
know it's exactly the same. A future where
[31:11]: you're just surrounded by robots is for certain.Â
And I'm just excited about having my own R2-D2.
[31:18]: And of course R2-D2 wouldn't be quite the can thatÂ
it is and roll around. It'll be you know R2-D2
[31:25]: yeah, it'll probably be a different physicalÂ
embodiment, but it's always R2. You know so my R2
[31:32]: is going to go around with me. Sometimes it's in myÂ
smart glasses, sometimes it's in my phone, sometimes
[31:36]: it's in my PC. It's in my car. So R2 is with meÂ
all the time including you know when I get home
[31:43]: you know where I left a physical version of R2. AndÂ
you know whatever that version happens to
[31:49]: be you know, we'll interact with R2. And so IÂ
think the idea that we'll have our own R2-D2 for
[31:55]: our entire life and it grows up with us, that'sÂ
a certainty now yeah. I think a lot of news media
[32:05]: when they talk about futures like this they focusÂ
on what could go wrong. And that makes sense. There
[32:10]: is a lot that could go wrong. We should talk aboutÂ
what could go wrong so we could keep it from from
[32:14]: going wrong. Yeah that's the approach that we likeÂ
to take on the show is, what are the big challenges
[32:19]: so that we can overcome them? Yeah. What buckets doÂ
you think about when you're worrying about this
[32:24]: future? Well there's a whole bunch of theÂ
stuff that everybody talks about: Bias or toxicity
[32:30]: or just hallucination. You know speaking withÂ
great confidence about something it knows nothing
[32:37]: about and as a result we rely on that information.Â
Generating, that's a version of generating
[32:45]: fake information, fake news or fake imagesÂ
or whatever it is. Of course impersonation.
[32:50]: It does such a good job pretending to be aÂ
human, it could be it could do an incredibly good
[32:56]: job pretending to be a specific human. And so
the spectrum of areas we
[33:05]: have to be concerned about is fairly clear andÂ
there's a lot of people who are
[33:11]: working on it. There's a some of the stuff,Â
some of the stuff related to AI safety requires
[33:18]: deep research and deep engineering andÂ
that's simply, it wants to do the right thing it
[33:24]: just didn't perform it right and as a result hurtÂ
somebody. You know for example self-driving car
[33:29]: that wants to drive nicely and drive properlyÂ
and just somehow the sensor broke down or it
[33:36]: didn't detect something. Or you know made itÂ
too aggressive turn or whatever it is. It did
[33:41]: it poorly. It did it wrongly. And so that's
a whole bunch of engineering that has to
[33:47]: be done to to make sure that AI safety is upheldÂ
by making sure that the product functions properly.
[33:54]: And then and then lastly you know whatever whatÂ
happens if the system, the AI wants to do a good
[34:00]: job but the system failed. Meaning the AI wantedÂ
to stop something from happening
[34:07]: and it turned out just when it wanted to doÂ
it, the machine broke down. And so this is
[34:13]: no different than a flight computer insideÂ
a plane having three versions of them and then
[34:19]: so there's triple redundancy inside theÂ
system inside autopilots and then you have two
[34:25]: pilots and then you have air traffic controlÂ
and then you have other pilots watching out for
[34:31]: these pilots. And so that the AI safetyÂ
systems has to be architected as a community
[34:38]: such that such that these AIs one, work,
function properly. When they don't
[34:47]: function properly, they don't put people in harm'sÂ
way. And that they're sufficient safety and
[34:52]: security systems all around them to make sureÂ
that we keep AI safe. And so there's
[34:58]: this spectrum of conversation is gigantic and andÂ
you know we have to take the parts, take the
[35:05]: parts apart and and build them as engineers. OneÂ
of the incredible things about this moment that
[35:11]: we're in right now is that we no longer have aÂ
lot of the technological limits that we had in a
[35:17]: world of CPUs and sequential processing. And we'veÂ
unlocked not only a new way to do computing and
[35:28]: and but also a way to continue to improve. ParallelÂ
processing has a a different kind of physics to it
[35:35]: than the improvements that we were able to makeÂ
on CPUs. I'm curious, what are the scientific or
[35:42]: technological limitations that we face now inÂ
the current world that you're thinking a lot
[35:47]: about? Well everything in the end is about how muchÂ
work you can get done within the limitations of
[35:54]: the energy that you have. And so that'sÂ
a physical limit and the laws of
[36:02]: physics about transporting information andÂ
transporting bits, flipping bits and transporting
[36:11]: bits, at the end of the day the energy it takesÂ
to do that limits what we can get done. And the
[36:18]: amount of energy that we have limits what we canÂ
get done. We're far from having any fundamental
[36:23]: limits that keep us from advancing. In the meantime,Â
we seek to build better and more energy efficient
[36:29]: computers. This little computer, the theÂ
big version of it was $250,000 - Pick up? - Yeah
[36:38]: Yeah that's little baby DIGITS yeah. This isÂ
an AI supercomputer. The version that I delivered,
[36:46]: this is just a prototype so it's a mockup.
The very first version was DGX 1, I
[36:52]: delivered to Open AI in 2016 and that was $250,000.Â
10,000 times more power, more energy necessary
[37:03]: than this version and this version has six timesÂ
more performance. I know, it's incredible. We're
[37:09]: in a whole in the world. And it's only since 2016Â
and so eight years later we've in increased the
[37:16]: energy efficiency of computing by 10,000 times.Â
And imagine if we became 10,000 times more energy
[37:25]: efficient or if a car was 10,000 times moreÂ
energy efficient or electric light bulb was
[37:31]: 10,000 times more energy efficient. Our lightÂ
bulb would be right now instead of 100 Watts,
[37:38]: 10,000 times less producing the same illumination.Â
Yeah and so the energy efficiency of
[37:45]: computing particularly for AI computing that we'veÂ
been working on has advanced incredibly and that's
[37:51]: essential because we want to create youÂ
know more intelligent systems and and we want to
[37:56]: use more computation to be smarter and soÂ
energy efficiency to do the work is our number one
[38:03]: priority. When I was preparing for this interview, IÂ
spoke to a lot of my engineering friends and this
[38:09]: is a question that they really wanted me to ask. SoÂ
you're really speaking to your people here. You've
[38:15]: shown a value of increasing accessibilityÂ
and abstraction, with CUDA and allowing more
[38:21]: people to use more computing power in all kinds ofÂ
other ways. As applications of technology get more
[38:28]: specific, I'm thinking of transformers in AI forÂ
example... For the audience, a transformer is a very
[38:35]: popular more recent structure of AI that's nowÂ
used in a huge number of the tools that you've
[38:40]: seen. The reason that they're popular is becauseÂ
transformers are structured in a way that helps
[38:45]: them pay "attention" to key bits of information andÂ
give much better results. You could build chips
[38:51]: that are perfectly suited for just one kind of AIÂ
model, but if you do that then you're making them
[38:56]: less able to do other things. So as these specificÂ
structures or architectures of AI get more popular,
[39:03]: my understanding is there's a debate between howÂ
much you place these bets on "burning them into the
[39:09]: chip" or designing hardware that is very specificÂ
to a certain task versus staying more general and
[39:15]: so my question is, how do you make those bets? HowÂ
do you think about whether the solution is a car
[39:22]: that could go anywhere or it's really optimizingÂ
a train to go from A to B? You're making bets
[39:28]: with huge stakes and I'm curious how you thinkÂ
about that. Yeah and that now comes back
[39:33]: to exactly your question, what are yourÂ
core beliefs? And the question, the core
[39:41]: belief either one, that transformer is the last AIÂ
algorithm, AI architecture that any researcher will
[39:52]: ever discover again, or that transformersÂ
is a stepping stone towards evolutions of
[40:01]: transformers that are uh barely recognizable as aÂ
transformer years from now. And we believe the
[40:08]: latter. And the reason for that is because youÂ
just have to go back in history and ask yourself,
[40:14]: in the world of computer algorithms, inÂ
the world of software, in the world of
[40:20]: engineering and innovation, has one idea stayedÂ
along that long? And the answer is no. And so that's
[40:27]: kind of the beauty, that's in factÂ
the essential beauty of a computer that it's able
[40:34]: to do something today that no one even imaginedÂ
possible 10 years ago. And if you would have, if
[40:41]: you would have turned that computer 10 years agoÂ
into a microwave, then why would the applications
[40:48]: keep coming? And so we believe, we believe in theÂ
richness of innovation and the
[40:54]: richness of invention and we want to create anÂ
architecture that let inventors and innovators
[40:59]: and software programmers and AI researchersÂ
swim in the soup and come up with some amazing
[41:05]: ideas. Look at transformers. The fundamentalÂ
characteristic of a transformer is this idea
[41:10]: called "attention mechanism" and it basically saysÂ
the transformer is going to understand the meaning
[41:16]: and the relevance of every single word with everyÂ
other word. So if you had 10 words, it has to figure
[41:22]: out the relationship across 10 of them. But if youÂ
have a 100,000 words or if your context is
[41:27]: now as large as, read a PDF and that read a wholeÂ
bunch of PDFs, and the context window is now like
[41:35]: a million tokens, the processing all of it acrossÂ
all of it is just impossible. And so the way you
[41:42]: solve that problem is there all kinds of new ideas,Â
flash attention or hierarchical attention or you
[41:49]: know all the, wave attention I just read aboutÂ
the other day. The number of different types of
[41:54]: attention mechanisms that have been inventedÂ
since the transformer is quite extraordinary.
[42:00]: And so I think that that's going to continueÂ
and we believe it's going to continue and that
[42:06]: that computer science hasn't ended and that AIÂ
research have not all given up and we haven't
[42:12]: given up anyhow and that having aÂ
computer that enables the flexibility of
[42:21]: of research and innovation and new ideas isÂ
fundamentally the most important thing. One of the
[42:29]: things that I am just so curious about, you designÂ
the chips. There are companies that assemble the
[42:37]: chips. There are companies that design hardware toÂ
make it possible to work at nanometer scale. When
[42:44]: you're designing tools like this, how do you thinkÂ
about design in the context of what's physically
[42:51]: possible right now to make? What are the thingsÂ
that you're thinking about with sort of pushing
[42:56]: that limit today? The way we do it is evenÂ
though even though we have things made like for
[43:05]: example our chips are made by TSMC. Even thoughÂ
we have them made by TSMC, we assume that we need
[43:13]: to have the deep expertise that TSMC has. And soÂ
we have people in our company who are incredibly
[43:19]: good at semiconductive physics so that we have aÂ
feeling for, we have an intuition for, what are the
[43:25]: limits of what today's semiconductor physicsÂ
can do. And then we work very closely with them to
[43:32]: discover the limits because we're trying to pushÂ
the limits and so we discover the limits together.
[43:36]: Now we do the same thing in system engineering andÂ
cooling systems. It turns out plumbing is really
[43:41]: important to us because of liquid cooling.Â
And maybe fans are really important to us
[43:44]: because of air cooling and we're trying to designÂ
these fans in a way almost like you know they're
[43:49]: aerodynamically sound so that we could pass theÂ
highest volume of air, make the least amount of
[43:54]: noise. So we have aerodynamics engineers in our
company. And so even though even though we don't
[44:01]: make 'em, we design them and we have to deepÂ
expertise of knowing how to have them made. And
[44:09]: and from that we try to push theÂ
limits. One of the themes of this conversation is
[44:18]: that you are a person who makes big bets on theÂ
future and time and time again you've been right
[44:25]: about those bets. We've talked about GPUs, we'veÂ
talked about CUDA, we've talked about bets you've
[44:30]: made in AI - self-driving cars, and we're going toÂ
be right on robotics and - this is my question. What
[44:37]: are the bets you're making now? the latest bet we
just described at the CES and I'm very very proud
[44:43]: of it and I'm very excited about it is theÂ
fusion of Omniverse and Cosmos so that we have
[44:50]: this new type of generative world generationÂ
system, this multiverse generation system. I
[44:59]: think that's going to be profoundly important inÂ
the future of robotics and physical systems.
[45:06]: Of course the work that we're doing with humanÂ
robots, developing the tooling systems and the
[45:11]: training systems and the human demonstrationÂ
systems and all of this stuff that that you've
[45:17]: already mentioned, we're just seeing theÂ
beginnings of that work and I think the
[45:23]: next 5 years are going to be very interesting inÂ
the world of human robotics. Of course the work
[45:28]: that we're doing in digital biology so thatÂ
we can understand the language of molecules and
[45:34]: understand the language of cells and just asÂ
we understand the language of physics and the
[45:39]: physical world we'd like to understand the languageÂ
of the human body and understand the language of
[45:44]: biology. And so if we can learn that, and we canÂ
predict it. Then all of a sudden our ability to
[45:50]: have a digital twin of the human is plausible.Â
And so I'm very excited about that work. I love
[45:56]: the work that we're doing in climate scienceÂ
and be able to, from weather predictions, understand
[46:03]: and predict the high resolution regional climates,Â
the weather patterns within a kilometer above
[46:10]: your head. That we can somehow predict that withÂ
great accuracy, its implications is really quite
[46:17]: profound. And so the number of things thatÂ
we're working on is really cool. You know we
[46:24]: we're fortunate that we've created thisÂ
this instrument that is a time machine and
[46:37]: we need time machines in all of these areas thatÂ
we just talked about so that we can see
[46:43]: the future. And if we could see the future andÂ
we can predict the future then we have a better
[46:48]: chance of making that future the best versionÂ
of it. And that's the reason why scientists
[46:53]: want to predict the future. That's the reason why,Â
that's the reason why we try to predict the future
[46:58]: and everything that we try to design so that weÂ
can optimize for the best version. So if
[47:05]: someone is watching this and maybe they came intoÂ
this video knowing that NVIDIA is an incredibly
[47:12]: important company but not fully understanding whyÂ
or how it might affect their life and they're now
[47:18]: hopefully better understanding a big shift thatÂ
we've gone through over the last few decades in
[47:23]: computing, this very exciting, very sort of strangeÂ
moment that we're in right now, where we're sort
[47:30]: of on the precipice of so many different things.Â
If they would like to be able to look into the
[47:36]: future a little bit, how would you advise them toÂ
prepare or to think about this moment that they're
[47:42]: in personally with respect to how these toolsÂ
are actually going to affect them? Well there are
[47:49]: several ways to reason about the future thatÂ
we're creating. One way to reason about it is,
[47:57]: suppose the work that you do continues toÂ
be important but the effort by which you
[48:04]: do it went from you know being a week longÂ
to almost instantaneous. You know that the
[48:15]: effort of drudgery basically goes to zero.Â
What is the implication of that? This is, this
[48:23]: is very similar to what would change if allÂ
of a sudden we had highways in this country?
[48:30]: And that kind of happened you know in the lastÂ
Industrial Revolution, all of a sudden we have
[48:34]: interstate highways and when you have interstateÂ
highways what happens? Well you know suburbs start
[48:40]: to be created and and all of a sudden you knowÂ
distribution of goods from east to west is
[48:48]: no longer a concern and all of a sudden gasÂ
stations start cropping up on highways and
[48:55]: and fast food restaurants show up and youÂ
know someone, some motels show up because people
[49:03]: you know traveling across the state, across theÂ
country and just wanted to stay somewhere for a
[49:07]: few hours or overnight, and so all of a suddenÂ
new economies and new capabilities, new economies.
[49:13]: What would happen if a video conference madeÂ
it possible for us to see each other without
[49:19]: having to travel anymore? All of a suddenÂ
it's actually okay to work further away from
[49:24]: home and from work, work and liveÂ
further away. And so you ask yourself kind of
[49:32]: these questions. You know what would happenÂ
if I have a software programmer with me
[49:40]: all the time and whatever it is I can dream up,Â
the software programmer could write for me. You
[49:46]: know what would, what would happenÂ
if I just had a seed of an idea and
[49:54]: and I rough it out and all of sudden a you knowÂ
a prototype of a production was put in front
[50:01]: of me? And what how would that change my life andÂ
how would that change my opportunity? And you
[50:07]: know what does it free me to be able to do andÂ
and so on so forth. And so I think that the next
[50:13]: the next decade intelligence, not for everythingÂ
but for for some things, would basically become
[50:22]: superhuman. But I can tellÂ
you exactly what that feels like. I'm surrounded
[50:31]: by superhuman people, super intelligence fromÂ
my perspective because they're the best in the
[50:38]: world at what they do and they do what theyÂ
do way better than I can do it. and I'm
[50:46]: surrounded by thousands of them and yet what itÂ
it never one day caused me to to think all of a
[50:56]: son I'm no longer necessary. It actually empowersÂ
me and gives me the confidence to go tackle more
[51:05]: and more ambitious things. And so suppose,Â
suppose now everybody is surrounded by these
[51:13]: super AIs that are very good at specific thingsÂ
or good at some of the things. What would that
[51:20]: make you feel? Well it's going to empower you,Â
it's going to make you feel confident and
[51:25]: and I'm pretty sure you probably use ChatGPT andÂ
AI and I feel more empowered today, more
[51:32]: confident to learn something today. The knowledgeÂ
of almost any particular field, the barriers to
[51:38]: that understanding, it has been reduced and I haveÂ
a personal tutor with me all of the time. And
[51:44]: so I think that that feeling should be universal.
If there's one thing that I would
[51:50]: encourage everybody to do is to go get yourselfÂ
an AI tutor right away. And that AI tutor could
[51:56]: of course just teach your things, anything youÂ
like, help you program, help you write,
[52:03]: help you analyze, help you think, help you reason,Â
you know all of those things is going to
[52:10]: really make you feel empowered and and I thinkÂ
that going to be our future. We're
[52:16]: going to become, we're going to become super humans,Â
not because we have super, we're going to become
[52:21]: super humans because we have super AIs. Could youÂ
tell us a little bit about each of these objects?
[52:27]: This is a new GeForce graphics card and yes andÂ
this is the RTX 50 Series. It is essentially
[52:39]: a supercomputer that you put into your PC and weÂ
use it for gaming, of course people today use it
[52:45]: for design and creative arts and it does amazingÂ
AI. The real breakthrough here and this is
[52:52]: this is truly an amazing thing, GeForceÂ
enabled AI and it enabled Geoff Hinton, Ilya Sutskever,
[52:59]: Alex Krizhevsky to be able to train AlexNet. We
discovered AI and we advanced AI then AI came back
[53:07]: to GeForce to help computer graphics. And so here'sÂ
the amazing thing: Out of 8 million pixels or so in
[53:16]: a 4K display we are computing, we're processingÂ
only 500,000 of them. The rest of them we use AI
[53:24]: to predict. The AI guessed it and yet the image isÂ
perfect. We inform it by the 500,000 pixels that we
[53:32]: computed and we ray traced every single one and it'sÂ
all beautiful. It's perfect. And then we tell the
[53:38]: AI, if these are the 500,000 perfect pixels in thisÂ
screen, what are the other 8 million? And it goes it
[53:44]: fills in the rest of the screen and it's perfect.
And if you only have to do fewer pixels, are you
[53:50]: able to invest more in doing that because you haveÂ
fewer to do so then the quality is better so the
[53:58]: extrapolation that the AI does... Exactly. BecauseÂ
whatever computing, whatever attention you have,
[54:03]: whatever resources you have, you can place it intoÂ
500,000 pixels. Now this is a perfect example of
[54:11]: why AI is going to make us all superhuman, becauseÂ
all of the other things that it can do, it'll do
[54:17]: for us, allows us to take our time and energy andÂ
focus it on the really really valuable things that
[54:23]: we do. And so we'll take our own resource which isÂ
you know energy intensive, attention intensive, and
[54:33]: we'll dedicated to the few 100,000 pixels andÂ
use AI to superres, upres it you know to
[54:39]: everything else. And so this this graphics cardÂ
is now powered mostly by AI and the computer
[54:47]: graphics technology inside is incredible asÂ
well. And then this next one, as I mentioned
[54:52]: earlier, in 2016 I built the first one for AIÂ
researchers and we delivered the first one to Open AI
[54:58]: and Elon was there to receive it and thisÂ
version I built a mini mini version and the
[55:06]: reason for that is because AI has now gone from AIÂ
researchers to every engineer, every student, every
[55:15]: AI scientist. And AI is going to be everywhere.Â
And so instead of these $250,000 versions we're
[55:21]: going to make these $3,000 versions and schoolsÂ
can have them, you know students can have them, and
[55:28]: you set it next to your PC or Mac and all ofÂ
a sudden you have your own AI supercomputer. And
[55:36]: you could develop and build AIs. Build your ownÂ
AI, build your own R2-D2. What do you feel like is
[55:42]: important for this audience to know that I haven'tÂ
asked? One of the most important things I would
[55:48]: advise is for example if I were a student todayÂ
the first thing I would do is to learn AI. How do
[55:54]: I learn to interact with ChatGPT, how do I learnÂ
to interact with Gemini Pro, and how do I learn
[56:00]: to interact with Grok? Learning how to
interact with with AI is not unlike being
[56:10]: someone who is really good at asking questions.Â
You're incredibly good at asking questions and
[56:17]: and prompting AI is very very similar.
You can't just randomly ask a bunch of questions
[56:23]: and so asking an AI to be assistantÂ
to you requires some expertise and
[56:30]: artistry and how to prompt it. And so if I were,Â
if I were a student today, irrespective whether
[56:35]: it's for for math or for science or chemistryÂ
or biology or doesn't matter what field of science
[56:40]: I'm going to go into or what profession, I'mÂ
going to ask myself, how can I use AI to do my job
[56:46]: better? If I want to be a lawyer, how can I useÂ
AI to be a better lawyer? If I want to be a better
[56:50]: do doctor, how can I use AI to be a better doctor?Â
If I want to be a chemist, how do I use AI to be
[56:55]: a better chemist? If I want to be a biologist, I howÂ
do I use AI to be a better biologist? That question
[57:02]: should be persistent across everybody. And just asÂ
my generation grew up as the first generation
[57:10]: that has to ask ourselves, how can we use computersÂ
to do our jobs better? Yeah the generation before
[57:17]: us had no computers, my generation was the firstÂ
generation that had to ask the question, how do I
[57:23]: use computers to do my job better? Remember I cameÂ
into the industry before Windows 95 right, 1984
[57:32]: there were no computers in offices. And after that,Â
shortly after that, computers started to emerge and
[57:38]: so we had to ask ourselves how do we use computersÂ
to do our jobs better? The next generation doesn't
[57:45]: have to ask that question but it has to askÂ
obviously next question, how can I use AI to
[57:49]: do my job better? That is start and finish I thinkÂ
for everybody. It's a really exciting and scary and
[57:59]: therefore worthwhile question I think for everyone.Â
I think it's going to be incredibly fun. AI is
[58:04]: obviously a word that people are just learningÂ
now but it's just you know, it's
[58:10]: made your computer so much more accessible. It isÂ
easier to prompt ChatGPT to ask it anything you
[58:15]: like than to go do the research yourself. And soÂ
we've lowered a barrier of understanding, we've
[58:22]: lowered a barrier of knowledge, we'veÂ
lowered a barrier of intelligence, and
[58:26]: and everybody really had to just go tryÂ
it. You know the thing that's really really crazy
[58:32]: is if I put a computer in front of somebody andÂ
they've never used a computer there is no chance
[58:37]: they're going to learn that computer in a day.
There's just no chance. Somebody really has to
[58:43]: show it to you and yet with ChatGPT if youÂ
don't know how to use it, all you have to do is
[58:49]: type in "I don't know how to use ChatGPT, tellÂ
me," and it would come back and give you some
[58:54]: examples and so that's the amazing thing.
You know the amazing thing about intelligence is
[59:02]: it'll help you along the way and make you uhÂ
superhuman you know along the way. All right I have
[59:08]: one more question if you have a second. This isÂ
not something that I planned to ask you but on the
[59:13]: way here, I'm a little bit afraid of planes,Â
which is not my most reasonable quality, and
[59:21]: the flight here was a little bit bumpy mhm veryÂ
bumpy and I'm sitting there and it's moving and
[59:30]: I'm thinking about what they're going to say at myÂ
funeral and after - She asked good questions, that's
[59:37]: what the tombstone's going to say - IÂ
hope so! Yeah. And after I loved my husband and my
[59:44]: friends and my family, the thing that I hoped thatÂ
they would talk about was optimism. I hope that
[59:49]: they would recognize what I'm trying to do here.Â
And I'm very curious for you, you've you've been
[59:56]: doing this a long time, it feels like there'sÂ
so much that you've described in this vision
[01:00:00]: ahead, what would the theme be that you wouldÂ
want people to say about what you're trying to do?
[01:00:14]: Very simply, they made an extraordinary impact.Â
I think that we're fortunate because of some
[01:00:23]: core beliefs a long time ago and sticking withÂ
those core beliefs and building upon them
[01:00:32]: we found ourselves today being one ofÂ
the most, one of the many most important and
[01:00:42]: consequential technology companies in
the world and potentially ever. And so
[01:00:49]: we take that responsibility very seriously.
We work hard to make sure that
[01:00:56]: the capabilities that we've created areÂ
available to large companies as well as
[01:01:03]: individual researchers and developers, acrossÂ
every field of science no matter profitable or
[01:01:10]: not, big or small, famous or otherwise.
And it's because of this understanding of
[01:01:21]: the consequential work that we're doing and theÂ
potential impact it has on so many people
[01:01:27]: that we want to make make this capabilityÂ
as pervasively as possible and I
[01:01:37]: do think that when we look back in a fewÂ
years, and I do hope that what the
[01:01:47]: next generation realized is as they, wellÂ
first of all they're going to know us because of
[01:01:53]: all the you know gaming technology we create.
I do think that we'll look back and the whole
[01:01:59]: field of digital biology and life sciences hasÂ
been transformed. Our whole understanding of of
[01:02:06]: material sciences has completely beenÂ
revolutionized. That robots are helping
[01:02:13]: us do dangerous and mundane things all over theÂ
place. That if we wanted to drive we can drive
[01:02:19]: but otherwise you know take a nap or enjoyÂ
your car like it's a home theater of yours,
[01:02:26]: you know read from work to home and at thatÂ
point you're hoping that you live far
[01:02:31]: away and so you could be in a car for longer.Â
And you look back and
[01:02:37]: you realize that there's this company almost atÂ
the epicenter of all of that and happens
[01:02:43]: to be the company that
you grew up playing games with.
[01:02:46]: I hope for that to be
what the next generation learn.
[01:02:50]: Thank you so much for your time.
I enjoyed it, thank you! I'm glad!