HomeVideos

Jensen Huang: Nvidia's Future, Physical AI, Rise of the Agent, Inference Explosion, AI PR Crisis

Now Playing

Jensen Huang: Nvidia's Future, Physical AI, Rise of the Agent, Inference Explosion, AI PR Crisis

Transcript

1882 segments

0:00

special episode this week. We've

0:02

preempted the weekly show and there's

0:05

only three people we preempt the show

0:06

for. President Trump, Jesus, and Jensen.

0:11

And uh I'll let you pick which order we

0:13

do that. Uh but what an amazing run

0:16

you've had and a great event. Uh

0:19

>> every industry is here. Every tech

0:21

company is here. Every AI company is

0:23

here. Incredible. Incredible.

0:25

>> Extraordinary. And one of the great

0:27

announcements of the past year has been

0:29

Grock. When you made the purchase of

0:31

Grock, did you realize how insufferable

0:33

Cha Chimath would become?

0:35

>> I had I had an inkling that that that

0:39

>> we're his friends. We have to deal with

0:40

him every week.

0:41

>> I know it.

0:42

>> You had to deal with him for the six

0:43

week close.

0:44

>> I know it's like two weeks. Two weeks.

0:46

>> It's all coming back to me now. It's

0:48

it's making me rather uncomfortable. The

0:50

the thing is uh many of our strategies

0:54

are are presented in in broad daylight

0:58

at GTC years in advance of when we do

1:01

it. Two and a half years ago, I

1:04

introduced the operating system of the

1:06

AI factory and it's called Dynamo.

1:08

Dynamo, as you know, is a piece of

1:11

instrument, a machine that was created

1:14

by Seammens to turn essentially water

1:17

into electricity. and Dynamo uh powered

1:22

the factory of the last industrial

1:23

revolution. So I thought it was the

1:25

perfect name for the operating system of

1:27

the next industrial revolution, the

1:29

factory of that. And so inside Dynamo,

1:32

the fundamental technology is

1:34

disagregated inference.

1:37

>> Jason, I I know you're you're super

1:40

technical.

1:40

>> Absolutely.

1:41

>> I know it.

1:41

>> I'll let you take this one. Go ahead and

1:43

define it for the audience. I don't want

1:44

to step on you.

1:45

>> Yeah, thank you. I I I know you wanted

1:46

to jump in there for a second, but it's

1:48

it's this aggregating inference, which

1:50

means the the pipeline, the processing

1:53

pipeline of inference is extremely

1:55

complicated. In fact, it is the most

1:57

complicated computing problem today.

1:59

Incredible scale, lots of mathematics of

2:02

different shapes and sizes. And we came

2:04

up came up with the idea that you would

2:07

change you would you would disagregate

2:09

parts of the processing such that some

2:12

of it can run on some GPUs rest of it

2:15

can run on different GPUs and that led

2:18

to us realizing that maybe even

2:21

disagregated computing could make sense

2:24

that we could have different

2:25

heterogeneous nature of computing that

2:27

same sensibility led us to melanox

2:31

>> you know today Nvidia's computing is

2:34

spread across GPUs, CPUs, switches,

2:37

scale up switches, scale out switches,

2:39

networking processors, and now we're

2:41

going to add Gro to that and we're going

2:43

to put the right workload on the right

2:45

chips. You know, we just really evolved

2:47

from a GPU company to an AI factory

2:49

company.

2:49

>> I mean, I think that was probably the

2:51

biggest takeaway that I had. You're

2:52

seeing this fundamental disagregation

2:54

where we've gone from a GPU and now you

2:57

have this complexion of all these

2:59

different options that will eventually

3:01

exist. The thing that you guys said on

3:03

stage or you said on stage was I I would

3:06

like the high value inference people to

3:09

take a listen to this and 25% of your

3:11

data center space you said should be

3:13

allocated to this gro lpu GPU combo

3:16

>> grock to about 25% of the ver rubins in

3:19

the in the data center. So can you tell

3:22

us about how the industry looks at this

3:24

idea of now basically creating this next

3:27

generation form of disagregated prefill

3:29

decode disag?

3:32

>> Yeah and take a step back and at the

3:35

time that we added this we went from

3:38

large language model processing

3:41

to agentic processing. Now when you're

3:45

running an agent you're accessing

3:48

working memory you're accessing

3:49

long-term memory. You're using tools.

3:52

You're really beating up on storage

3:54

really hard. You have agents working

3:56

with other agents. Some of the agents

3:58

are very large models. Some of them are

4:00

smaller models. Some of them are

4:01

diffusion models. Some of them are auto

4:04

reggressive models. And so there are all

4:06

kinds of different types of models

4:08

inside this data center. We created Vera

4:10

Rubin to be able to run this

4:12

extraordinarily diverse workload. My

4:14

sense is and so we added what used to be

4:18

a one rack company. we now added four

4:20

more racks,

4:21

>> right?

4:22

>> So, Nvidia's TAM, if you will, increased

4:25

from what whatever it was to probably

4:28

something, call it, you know, 33% 50%

4:31

higher. Now, part of that 33% or 50% a

4:34

lot of it is going to be storage

4:36

processors. It's called Blue Field. Some

4:38

of it will be a lot of it, I'm hoping,

4:40

will be Grock processors, and some of it

4:43

will be CPUs. And they're all and a lot

4:46

of it's going to be networking

4:48

processors. And so all of this is going

4:50

to be running basically the computer of

4:54

the AI revolution called agents, right?

4:56

>> The operating system of of um modern

4:59

modern industry.

5:00

>> What about embedded applications? So you

5:02

know my daughter's teddy bear at home

5:04

wants to talk to her. What goes in

5:06

there? Is it a custom ASIC or does there

5:09

end up becoming much more kind of a

5:11

broader set of TAM with developing tools

5:14

that are maybe different for different

5:15

use cases at the edge and an embedded

5:17

application? We think that there's three

5:19

computers in the problem at the at the

5:21

largest at the largest scale when you

5:23

stick take a step back. There's one

5:25

computer that's really about training

5:27

the AI model, developing creating the

5:29

AI, another computer for evaluating it.

5:32

Depending on the type of problem you're

5:34

having, like for example, you look

5:36

around, there's all kinds of robots and

5:37

cars and things like that. You have to

5:39

evaluate these robots inside a virtual

5:44

gym that represents the physical world.

5:47

So it has to be software that obeys the

5:49

laws of physics. And that's a second

5:52

computer. We call that omniverse. The

5:54

third computer is the computer at the

5:56

edge, the robotics computer.

5:58

>> That robotics computer, one of them

5:59

could be self-driving car. Another one's

6:01

a robot. Another one could be a teddy

6:02

bear. Little tiny one for a teddy bear.

6:05

One of the most important ones is one

6:07

that we're working on that basically

6:09

turns the telecommunications base

6:11

stations into part of the AI

6:14

infrastructure. So now all of the it's a

6:17

$2 trillion industry. All of that in

6:20

time will be transformed into an

6:22

extension of the AI infrastructure. And

6:24

so radios radios will become a edge

6:27

devices, factories, warehouses, you name

6:30

it. And so so there are three these

6:32

three basic computers.

6:34

All of them, you know, are going to be

6:36

necessary.

6:36

>> Jensen, last uh last year, I think you

6:39

were ahead of the the rest of the world

6:41

in in in saying inference isn't going to

6:43

a thousand.

6:44

>> Just last year,

6:45

>> yes.

6:46

Is it is it going to 1 millionx is going

6:49

to 1 billionx? Yeah.

6:50

>> Right. And I think people at the time

6:52

thought it was pretty hyperbolic because

6:53

the world was still focused on

6:55

pre-scaling, on training. Here we are

6:58

now. Inference has exploded. We're

6:59

inference constrained. um you announced

7:02

an inference factory that I think is

7:04

leading edge that's going to be 10x

7:06

better in terms of throughput to the

7:08

next factory but yet if you if I listen

7:10

to what the chatter is out there it's

7:13

that your inference factory is going to

7:14

cost 40 or 50 billion and the

7:16

alternatives the custom AS6 AMD others

7:19

are going to cost 25 to 30 billion and

7:22

you're going to lose share so why don't

7:24

you talk to us what are you seeing how

7:25

do you think about share and does it

7:28

make sense for all these folks to pay

7:29

something that's a 2x X premium to what

7:32

others are marketing.

7:33

>> The big takeaway, the big idea is that

7:39

you should not equate the price of the

7:43

factory and the price of the tokens, the

7:46

cost of the tokens. It is very likely

7:49

that the $50 billion factory, and in

7:52

fact, I can prove it that the $50

7:54

billion factory will generate for you

7:56

the lowest cost tokens. And the reason

8:00

for that is because we produce these

8:02

tokens at extraordinary efficiency

8:05

10 times you know the difference between

8:08

50 billion. Now it turns out 20 billion

8:11

is just land power and shell. Right.

8:12

>> Right.

8:13

>> And then on top of that you have storage

8:15

anyways networking anyways you got CPUs

8:17

anyways you got servers anyways you got

8:20

cooling anyways. The difference between

8:22

that GPU being 1x price or halfx price

8:26

>> is not between 50 billion and 30

8:28

billion. Pick your favorite number, but

8:30

let's say between 50 billion and 40

8:32

billion.

8:33

>> That is not a large percentage when the

8:35

$50 billion

8:37

>> data center is actually 10 times the

8:39

throughput,

8:40

>> right? Jess, that's the reason why I

8:42

said that even for most chips,

8:45

>> if you can't keep up with the state of

8:47

the technology and the pace that we're

8:49

running, even when the chips are free,

8:51

it's not cheap enough.

8:52

>> Yeah.

8:52

>> Can I can I just ask a general strategy

8:54

question?

8:55

>> Yeah.

8:56

>> I mean, you're running the most valuable

8:58

company in the world. This thing is

8:59

going to do 350 plus billion of revenue

9:02

next year, 200 billion of free cash

9:04

flow. It's compounding at these crazy

9:06

rates. How do you decide what to do?

9:09

like how do you actually get the

9:11

information? I mean it's famous now

9:13

these sort of emails that are people are

9:15

meant to send you but how do you really

9:17

decide to get an intuition of how to

9:20

shape the market where to really double

9:21

down where to maybe pull back where to

9:23

actually go into a green field how how

9:25

does that information get to you how do

9:26

you decide these things

9:27

>> in a final analysis that's the job of

9:28

the CEO yeah

9:29

>> and our job is to define the strategy

9:32

define the vision define the strategy

9:34

we're informed of course by amazing

9:36

computer scientists amazing

9:37

technologists great people all over the

9:39

company but we have to shape that future

9:41

well Part of it has to do with is this

9:44

something that's insanely hard to do? If

9:46

it's not hard to do, we should back away

9:48

from it. And the reason for that is if

9:50

if it's easy to do, obviously, um,

9:52

>> lots of competitors,

9:53

>> a lot of competitors. Is this something

9:54

that has never been done before that's

9:57

insanely hard to do? And that somehow

9:59

taps into

10:00

>> the special superpowers of our company.

10:03

>> And so I have to find this confluence of

10:05

things to that meets the standard. And

10:09

in the end, we also know that a lot of

10:11

pain and suffering is going to go into

10:12

it. Yeah.

10:13

>> There no great things that are invented

10:15

because it was just easy to do and just

10:17

like first try, here we are. And so if

10:20

it's super hard to do, nobody's ever

10:22

done it before, it's very likely that

10:23

you're going to have a lot of pain and

10:24

suffering and so you better enjoy it. So

10:26

can you can you just look at maybe three

10:28

or four of the more longtail things you

10:30

announced and just talk about the

10:32

long-term viability of whether it's the

10:34

data centers in space or whether it's

10:36

what you're trying to do with ADAS in

10:38

autos or you know what you're trying to

10:39

do on the biology side just give us a

10:41

sense of like how you see some of these

10:43

curves inflecting upwards in some of

10:44

these longer tail businesses.

10:46

>> Excellent. uh physical AI large category

10:50

we believe and I just mentioned we have

10:52

three computing systems all the software

10:54

platforms on top of it physical AI as a

10:56

large category

10:59

it's technology industry's first

11:02

opportunity

11:03

to address a $50 trillion industry that

11:07

has largely been you know void of

11:09

technology until now and so we need to

11:12

invent all of the technology necessary

11:13

to do that I felt that that was a

11:15

10-year journey

11:16

We started 10 years ago. We're seeing it

11:18

inflecting now. It is a multi-billion

11:21

dollar business for us. It's close to10

11:23

billion a year now. And so it's a big

11:25

business and it's growing exponentially.

11:27

And so that's number one. I think in the

11:29

case of digital biology, I think we are

11:31

literally near the chat GPT moment of

11:34

digital biology. We're about to

11:35

understand how to represent genes,

11:38

proteins, cells. We already know how to

11:40

understand chemicals. And so the ability

11:43

for us to represent and understand the

11:46

dynamics of the building blocks of

11:48

biology that's a couple of two three

11:51

five years from now in 5 years time I

11:54

completely believe that the healthcare

11:55

industry where digital biology is going

11:57

to inflect and so these are a couple of

11:59

the really great ones and you could see

12:01

they're all around us

12:02

>> agriculture

12:03

>> agriculture

12:04

>> reflecting now

12:05

>> no question yeah

12:06

>> Jensen I want to take you from the data

12:08

center to the desktop uh the company was

12:11

built in large part on hobbyists, video

12:14

gamers and and all those graphic cards

12:16

in the beginning. And you mentioned in

12:18

front of I think 10,000 people here just

12:21

clawed open claw clawed code and what a

12:25

revolution agents have become and

12:27

specifically the hobbyists who are

12:30

really where a lot of energy um we see

12:33

you know a lot of the innovation breaks

12:35

want desktops. You announced one here uh

12:38

I believe it's the Dell 6800. Uh this is

12:41

a very powerful workstation to run local

12:43

models. 750 gigs of RAM. Obviously the

12:46

the Mac uh studio sold out everywhere in

12:49

my company. We're moving to OpenClaw

12:51

everything. Freeberg just got clawed.

12:54

You got claw pelled I understand. And

12:55

you're obsessed with these.

12:57

>> What is this from the streets movement

13:00

of creating open-source agents and using

13:03

open source on the desktop mean to you?

13:06

Great. Where is that going?

13:07

>> Yeah. So great. First of all, let's take

13:08

a step back. Um in the last two years we

13:11

saw basically three inflection points.

13:14

The first one was generative chat GPT

13:19

brought AI to the common

13:21

everybody to our awareness. But the fact

13:24

of the matter is the technology sat in

13:26

plain sight months before GPT. It wasn't

13:29

until chat GPT put a user interface

13:33

around it made it easy for us to use

13:35

that generative AI took off. Now

13:36

generative AI as you know generates

13:39

tokens for internal consumption as well

13:41

as external consumption. Internal

13:43

consumption is thinking which led to

13:46

reasoning. 01 and 03

13:49

continue that wave of chat GPT grounded

13:51

information made AI not only answer

13:54

questions but answer questions in a more

13:56

grounded way useful.

13:58

We started seeing the revenues and the e

14:01

the economic model of open AI start to

14:03

inflect. Then the third one was only

14:07

inside the industry that we saw clock

14:09

code the first agentic system that was

14:12

very useful really revolutionary stuff

14:15

but but claw code was only available for

14:18

enterprises. Most people outside never

14:21

saw anything about cloud code until open

14:24

claw. Open claw basically put into the

14:28

po popular consciousness what an AI

14:31

agent can do. Mhm.

14:33

>> That's the reason why open claw is so

14:35

important from a cultural perspective.

14:37

Now the second second reason why it's so

14:39

important is that open claw is open but

14:43

it formulates

14:46

it structures a type of computing model

14:49

that is basically reinventing computing

14:52

all together. It has a memory system. It

14:55

scratch is a short-term memory file

14:57

system. It has it has it has scales. Did

15:01

you say skills or scales?

15:02

>> Skills.

15:03

>> Oh, skills.

15:03

>> They do have scales theoretically. Yeah.

15:05

>> Yeah. Skills.

15:06

>> So, the first thing first thing it it,

15:08

you know, it has resources. It it

15:09

manages resources. It's it does

15:12

scheduling.

15:12

>> Yep.

15:13

>> Right. And it cron jobs. It could it

15:16

could spawn off agents. It could, you

15:18

know, it could decompose a task and and

15:20

cause and solve problems as does

15:22

scheduling. It has IO subsystems. It

15:25

could, you know, input. It has output.

15:26

It connect to WhatsApp. And also it has

15:29

a API that allows it to run multiple

15:33

types of applications called skills.

15:35

>> Yeah.

15:35

>> These four elements fundamentally define

15:38

a computer.

15:39

>> Yeah.

15:40

>> And therefore what do we have? We have a

15:43

personal artificial intelligence

15:47

computer for the very first time.

15:49

>> Open source.

15:50

>> It's open source. It runs literally

15:52

everywhere. And so this is now the this

15:54

is the op this is basically the

15:56

blueprint the operating system of modern

15:58

computing.

15:59

>> Yeah.

15:59

>> And it's going to run literally

16:00

everywhere. Now of course one of the

16:02

things that we had to help it do is

16:04

whenever you have agentic software you

16:07

have to make sure that and agentic

16:08

software has access to sensitive

16:10

information. It execute code. It could

16:12

communicate externally. We have to make

16:14

sure that all of it has to be governed.

16:16

all of it has to be secure and that we

16:19

have policies that that gives these

16:21

agents two of the three things but not

16:24

all three things at the same time

16:25

>> and so the governance part of it we

16:27

contributed to Peter Peter Steinberger

16:30

was here and and so we've got a mountain

16:31

of great engineers working with him to

16:33

help secure and keep that thing so that

16:35

it could protect our privacy protect our

16:37

security

16:38

>> Jensen that paradigm shift makes some of

16:40

the AI legislation that has passed

16:43

around the country to regulate AI and a

16:46

lot of the proposed legislation

16:47

effectively moot, doesn't it? Can you

16:49

just comment for a second on how quickly

16:50

the paradigm shift kind of obiates a lot

16:53

of the models for regulatory oversight

16:56

of AI, which is becoming a very hot

16:58

topic in politics right now.

17:00

>> Well, this is this is the part that that

17:02

we just with policy makers, we need to

17:05

we need to always get in front of them

17:06

and Brad, you do a great job doing this.

17:08

We had to get in front of them and

17:09

inform them about the state of the

17:11

technology, what it is, what it is not.

17:14

It is not a biological being. It is not

17:19

alien. It is not conscious.

17:23

Um it is computer software.

17:26

>> Yeah. Exactly.

17:26

>> And and it is not something that um we

17:30

say things like we don't understand it

17:32

at all.

17:33

>> It is not true. We don't understand at

17:35

all. We understand a lot of things about

17:36

this technology. and and so so I think

17:39

one we have to make sure that we

17:40

continue to inform the policy makers and

17:43

not affect not allow dumerism and

17:46

extremism to affect how policy makers

17:49

think and understand about this

17:51

technology. However, however, we still

17:52

have to recognize is technology is

17:54

moving really fast and don't get policy

17:56

ahead of the technology too quickly. And

17:59

the risk that we we run as a nation, our

18:03

greatest source of national security

18:05

concern with respect to AI is that other

18:07

countries adopt this technology while we

18:10

are so angry at it or afraid of it or

18:14

somehow paranoid of it that our

18:16

industries, our society don't take

18:18

advantage of AI. So I'm just mostly

18:21

worried about the diffusion of AI here

18:22

in United States.

18:23

>> Can you just double click if you were in

18:25

the seat in the boardroom of anthropic

18:28

over that whole scuttlebutt with the

18:30

department of war? It sort of builds on

18:32

this idea of people didn't know what to

18:34

think. It's sort of added to this layer

18:37

of either resentment or fear or just

18:40

general mistrust that people have

18:42

sometimes at the software levels of AI.

18:44

What would do you think you would have

18:46

told Daario and that team to do maybe

18:48

differently to try to change some of

18:50

this outcome and some of this

18:51

perception?

18:51

>> The first thing that I I would I would

18:53

say about Anthropic is first of all the

18:54

technology is incredible. We are a large

18:57

consumer of anthropic technology. Really

18:59

admire their focus on security. Really

19:02

admires their focus on safety. Um the

19:04

the the the culture by which we they

19:07

went about it. The the technology

19:09

excellence by which they went about it

19:11

really fantastic. Um I I would say that

19:14

that the the desire to warn people about

19:18

the capability of the technology is is

19:20

also uh really terrific. We just have to

19:23

make sure that we understand that the

19:25

world has a spectrum and that that

19:28

warning is good, scaring is less good,

19:31

>> right?

19:31

>> Um and because this technology is too

19:34

important to us,

19:35

>> right? And and I think that it is fine

19:38

to uh predict the future but we need to

19:42

be a little bit more circumspect. We

19:44

need to have a little bit more humility

19:46

that in fact we can't completely predict

19:49

the future and the abil and to say

19:52

things that that are quite extreme quite

19:54

catastrophic that there's no evidence of

19:57

it happening um could be more damaging

20:00

than people think. And and of course we

20:03

are technology leaders. Uh there were

20:06

there was a time when nobody listened to

20:08

us. Yeah.

20:08

>> Um but now because technology is so

20:11

important in the social fabric such an

20:14

important industry, so important to

20:16

national security, our words do matter.

20:18

And I think we have to be much more

20:20

circumspect. We have to be more

20:21

moderate. We have to be more balanced.

20:23

We have to be more for more thoughtful.

20:25

>> Well, I you know I would nominate you. I

20:28

think the industry's got to get

20:29

together. 17% popularity of AI in the

20:32

United States. I mean, we see what

20:34

happened to nuclear, right? We basically

20:36

shut down the entire nuclear industry

20:38

and now we have a 100 fision reactors

20:40

being built in China and zero in the

20:42

United States. Um, we hear about

20:44

moratoriums on data centers. So, I think

20:45

we have to be a lot more proactive about

20:47

that. But, but I want to go back to this

20:49

agentic explosion that you're seeing

20:51

inside your company, the efficiencies,

20:53

the productivity gains inside your

20:55

company. There's a lot of debate whether

20:56

or not we're seeing ROI, right? and you

20:59

and I entering into into this year, the

21:01

big question was, are the revenues going

21:03

to show up? Are the revenues going to

21:05

scale like intelligence? And then we had

21:08

this kind of Oenheimer moment, a five6

21:10

billion month by Anthropic in February.

21:13

Um, do you think as you look ahead, you

21:16

announced a trillion dollar, you know,

21:18

visibility into a trillion dollars of

21:20

just Blackwell and Vera Rubin over the

21:22

course of the next couple years. When

21:24

you see this happening at Anthropic and

21:26

Open AI, do you think we're on that

21:28

curve now where we're going to see

21:29

revenues scale in the way that

21:31

intelligence is scaling?

21:32

>> When you look around when you I'll

21:34

answer this a couple different ways.

21:35

When you look around this audience, you

21:37

will see that anthropic and open AI is

21:39

represented here. But in fact, everybody

21:42

99% of everything that is here is all AI

21:45

and it's not anthropic and open AI.

21:46

>> Right. Right.

21:47

>> And the reason for that is because AI is

21:49

very diverse.

21:51

>> I would say that the second most popular

21:54

model as a category is open models.

21:58

>> Number one is yeah open open source open

22:00

ways open source.

22:02

>> Open AI is number one. Open source is

22:04

number two. Very distant. Third is

22:06

anthropic. And that tells you something

22:08

about the scale of all of the AI

22:10

companies that are here. And so, so it's

22:13

important to recognize recognize that.

22:16

Um, let me let me come back and say a

22:18

couple things. One, when we went from

22:20

generative to reasoning, the amount of

22:23

computation we needed was about a

22:25

hundred times. Right?

22:26

>> When we went from reasoning to agentic,

22:29

the computation is probably another 100

22:32

times. Now we're looking at in just two

22:35

years, computation went up by a fact

22:39

10,000x.

22:41

Meanwhile,

22:43

people pay for information, but people

22:47

mostly pay for work.

22:48

>> Yes.

22:50

>> Talking to a chatbot and getting an

22:52

answer is super great,

22:54

>> right? helping me do some research.

22:56

Unbelievable. But getting work done,

22:58

I'll pay fordeed.

23:00

>> And so that's where we are. Agentic

23:02

systems get work done. They're helping

23:04

our software engineers get work done.

23:06

And and so then you take that,

23:08

>> you got 10,000x more compute.

23:11

>> You get probably at this point 100x more

23:14

consumption now.

23:15

>> Yes.

23:15

>> Yeah.

23:15

>> And we haven't even started scaling yet.

23:18

We are absolutely at a millionx

23:20

>> which is I think a great place to talk

23:22

about the number of

23:24

have 20 30,000 at the company something

23:26

like that.

23:26

>> We have 43,000 employees. You know, I

23:29

would say 38,000

23:31

are engineers.

23:32

>> The conversation we've had on the pod a

23:35

number of times is, "Oh my god, look at

23:36

the token usage in our companies. It is

23:39

growing massively." And some people are

23:41

asking, "Hey, when I join a company, how

23:43

many tokens do I get cuz I want to be an

23:45

effective employee?" And you postulated,

23:48

I believe, during your 2 and 1/2 hour

23:50

keynote, pretty long keynote. Well done

23:54

that you were spending

23:56

>> if it was well done it would be shorter.

23:58

I just want

23:58

>> you didn't have time to do a time to

24:00

write an hour 45.

24:02

>> So you guys so you guys know so you guys

24:04

know there is no practice

24:06

>> and so it's a gripping and ripping

24:07

>> rip and rip. Yeah.

24:08

>> So so I just want to let you know I was

24:10

writing the speech while I was giving

24:12

the speech. Okay. So you never know

24:15

>> but

24:16

>> does that mean if we do back

24:17

>> I apologize. back of the envelope math

24:19

$75,000 in tokens for each engineer or

24:22

something like that. So are you spending

24:24

in Nvidia a billion2 billion on tokens

24:26

for your engineering team right now?

24:28

>> We're trying to Let me give you a

24:29

thought experiment. Let's say you have a

24:31

software engineer or AI researcher and

24:33

you pay them $500,000 a year. We do that

24:36

all the time.

24:37

>> Okay, this is happening all over the

24:39

time. um that $500,000 engineer at the

24:42

end of the year I'm going to ask him how

24:44

many tok how much did you spend in

24:46

tokens and that person said $5,000 I

24:49

will go ape something else

24:50

>> yes

24:50

>> right

24:51

>> if that if that $500,000 engineer did

24:54

not consume at least $250,000 worth of

24:57

tokens I am going to be deeply alarmed

25:01

okay and this is no different than one

25:04

of our chip designers who says guess

25:06

what I'm just going to use paper and

25:08

pencil I don't think I'm going to need

25:10

any CAD tools.

25:11

>> This is a real paradigm shift to start

25:13

thinking about these all-star employees.

25:15

It almost reminds me of of what we

25:17

learned in the NBA when LeBron James

25:19

started spending a million dollars a

25:20

year just on his health of his body like

25:23

and maintaining it.

25:24

>> That's right.

25:24

>> Here he is at age 41 still playing.

25:26

>> It really is, hey, if these are

25:28

incredible knowledge workers, why

25:30

wouldn't we give them

25:32

>> superhuman abilities?

25:33

>> That's exactly

25:34

>> where does that go? If we if we

25:36

extrapolate out two or three years from

25:38

now, what is the efficiency of that

25:40

allstar at an Nvidia and what they're

25:42

able to accomplish? What do they look

25:44

like?

25:44

>> Well, first of all, things that that

25:46

that um wow, this is too hard. That

25:50

thought is gone. Uh this is going to

25:52

take a long time. That thought is gone.

25:54

Uh we're going to need a lot of people.

25:55

That thought is gone. This is no

25:57

different than in this in the last

25:59

industrial re revolution. Somebody goes,

26:01

"Boy, that building really looks heavy."

26:04

Nobody says that. Nobody. Wow, that

26:06

mountain looks too big. Nobody says

26:07

that. Right.

26:08

>> Everything that's too big, too heavy,

26:11

takes too long,

26:12

>> those thought, those ideas are all gone.

26:14

>> You're reduced to creativity. That's

26:15

right. What can you come up with?

26:17

>> Exactly. Which means now the question is

26:19

how do you how do you work with these

26:21

agents? Well, it's just a new way of

26:23

doing computer programming. In the f in

26:25

the past, we code. In the future, we're

26:27

going we're going to write ideas,

26:29

architectures, specifications.

26:32

We're going to organize teams. We're

26:33

going to give them We're going to help

26:35

them define how to evaluate the

26:37

definition of good versus bad. What's

26:39

the what does it look like when

26:41

something is a great outcome? How to

26:43

iterate with you, how to brainstorm.

26:45

That's really what you're looking for.

26:46

And I'm I think that every engineer is

26:48

going to have hundred a hundred agents.

26:51

>> Back to the PR problem the industry has

26:53

right now. You have executives uh like

26:57

David Freeberg with Oho who's looking at

27:00

literally taking through the use of

27:02

technology your technology and AI the

27:05

number of calories produced and making

27:07

high quality cal calories what is the

27:10

factor you think you can bring the cost

27:12

down for what impact does this vision

27:15

have for what you're doing

27:17

>> zero shot genomic modeling and it works

27:20

>> and you have that moment and you're like

27:22

holy

27:24

>> honestly like and and that's after

27:26

people are replacing entire enterprise

27:29

software stacks in a night. I did

27:30

something in 90 minutes I was telling

27:32

the guys about replaced a whole software

27:34

stack and like a whole bunch of workload

27:36

90 minutes on cloud ran this agentic

27:38

system built the whole thing deployed it

27:40

and we got we were on a Sunday night

27:41

>> on a Sunday night 10 p.m. I was done at

27:43

11:30. I went to bed.

27:44

>> As the CEO, you replaced

27:46

>> Yeah. And everyone on my management team

27:48

had to do a similar exercise over the

27:50

weekend. What we saw on Monday, I was

27:52

like, it's over. But the technical

27:55

stuff, the science stuff, we did

27:57

something in 30 minutes using auto

27:59

research, and I'd love your view on auto

28:00

research and what that tells us about

28:02

how far we still have to go in terms of

28:04

efficiency. But using auto research and

28:06

a chunk of data, something was published

28:09

internally that we said, "Oh my god."

28:11

And that would normally be a PhD thesis

28:13

that would take seven years. It would be

28:14

one of the most celebrated PhD thesis

28:16

we've ever seen in this field and it

28:18

would be in the journal science and it

28:19

was done in 30 minutes on a desktop

28:21

computer running on auto research with

28:23

all the data we just ingested. We got it

28:25

on Friday and we're like, "Hey, let's

28:26

try it." Try booted up, went to GitHub,

28:28

downloaded Auto Research and ran it. And

28:30

you see everyone's face just go like and

28:33

then the potential of what this is

28:35

unlocking for us is like the kind of

28:37

thing that would take seven years and it

28:39

happened in 30 minutes and we're

28:41

experiencing it in genomics and we're

28:42

like this is unbelievable. So I I think

28:45

like the acceleration is widening the

28:48

aperture for everyone in a way that like

28:50

you didn't imagine a few years ago. But

28:53

just going back to the auto research

28:54

point, can you just comment on what you

28:56

think about the fact that this thing got

28:57

published with 600 lines of code in a

29:00

weekend and the capacity that it has to

29:02

run locally and achieve what it can

29:03

achieve with all of these diverse data

29:05

sets and what that tells us about the

29:07

early stages we are in terms of

29:09

optimization on algorithms and hardware.

29:11

The fundamental reason why Open Claw is

29:14

so incredible number one is it's com its

29:18

confluence its timing with the

29:20

breakthroughs in large language model.

29:23

>> Yeah,

29:23

>> its timing was perfect. It was

29:24

impeccable. Now, in a lot of ways, Peter

29:27

wouldn't have come up with it probably

29:29

if not for the fact that Claude and GPT

29:32

and chat GPT have reached a level that

29:34

is really very good,

29:35

>> right? It is also a new capability that

29:39

allows these models to tool use the

29:42

tools that we've created over time web

29:45

browsers and Excel spreadsheets and you

29:48

know in the case of chip design synopsis

29:50

and cadence and uh omniverse and blender

29:54

and autodesk and all of these tools are

29:57

going to continue to be used. There's

29:58

some some people say that that the

30:01

enterprise IT software industry is going

30:03

to get destroyed. There's it's there's a

30:07

let me give you the alternative view.

30:08

The enterprise software industry is

30:10

limited by butts and seats. It's about

30:13

to get a hundred times more agents

30:15

banging on those tools. They're going to

30:17

be agents banging on SQL. They're going

30:19

to be agents bang on vector databases,

30:21

agents bang on Blender, agents bang on

30:23

Photoshop. And the reason for that is

30:25

because those tools are first of all do

30:27

a very good job. Second, those tools are

30:30

the conduit between us in the final

30:34

analysis. When the work is done, it has

30:36

to be represented back to me in a way

30:38

that I can control.

30:39

>> Right?

30:40

>> And I know how to control those tools.

30:42

And so I need everything to be put back

30:44

into synopsis. I want everything to put

30:46

back into cadence because that's how I

30:48

control it. That's how I've ground

30:50

truth.

30:50

>> Let me ask you a question about open

30:52

source. So we have these closed source

30:54

models. They're excellent.

30:55

>> We have these openweight models. Many of

30:57

the Chinese models are incredible.

30:59

Absolutely incredible. Two days ago, you

31:02

may not have seen this because you were

31:03

busy on stage, but there was a training

31:06

run that happened in this crypto project

31:08

called Bit Tensor Subnet 3. They managed

31:11

to train a 4 billion parameter llama

31:13

model totally distributed with a bunch

31:15

of people contributing

31:18

excess compute, but they were able to do

31:20

it statefully and manage a training run,

31:22

which I thought was like a pretty crazy

31:24

technical accomplishment.

31:26

>> Yeah. Because it's like random people

31:27

and each person gets a little share.

31:29

>> Our our modern version of folding at

31:31

home.

31:32

>> Exactly. So what what do you think about

31:34

the end state of open source? Do you see

31:36

this decentralization of architecture as

31:39

well and decentralization of compute to

31:42

support open weights and a totally open-

31:45

source approach to making sure AI is

31:47

broadly available to every?

31:48

>> I believe we fundamentally need

31:51

models as a firstass product proprietary

31:55

product as well as models as open

31:58

source. These two things are not A or B.

32:01

It's A and B. There's no question about

32:03

it. And the reason for that is because

32:05

models is a technology, not a product.

32:08

Model is a technology, not a service.

32:10

For the vast majority of consumers, the

32:13

horizontal layer, the general

32:15

intelligence, I would really, really

32:17

love not to go fine-tune my own. I would

32:20

really love to keep using chat GPT. I'd

32:22

love to use Claude. I love to use

32:24

Gemini. I love to you use X. And they

32:26

all have their own personalities as you

32:28

know, which is kind of depends on my

32:29

mood and depends on what problem I'm

32:31

trying to solve. you know I might you

32:32

know do it on X or I might do it on on

32:34

chat GBT and so that that segment of the

32:37

of the industry is thriving it's going

32:39

to be great however there all these

32:43

industries their domain expertise their

32:46

specialization has to be channeled has

32:48

to be captured in a way that they can

32:51

control and that it can only come from

32:53

open models the open model industry

32:56

we're contributing tremendously to it is

32:58

near the frontier

33:00

and quite Quite frankly, even if it

33:03

reaches the frontier,

33:05

I think that products as a service,

33:08

worldclass products as as a models as a

33:10

product is going to continue to thrive.

33:12

>> Every startup we're investing in now is

33:16

open- source first and then going to the

33:18

proprietary models.

33:19

>> Yeah. And the beautiful thing is because

33:20

you have a great router you connect it

33:22

to by on on first day every single day

33:26

you're going to have access to the

33:27

world's best model and and then it gives

33:31

you time to cost reduce and fine-tune

33:33

and specialize and so you're going to

33:34

have worldclass capabilities out to

33:36

shoot every single time. Let J can I

33:39

question?

33:40

>> Nobody wants the US to win the global AI

33:42

race more than you, right? But a year

33:45

ago, the Biden era diffusion rule really

33:48

was an anti- American diffusion of AI

33:51

around the world. So here we are a year

33:53

into the new administration.

33:56

Give us a grade. Where is where are we

33:59

in terms of global diffusion and the

34:00

rate at which we're spreading US AI

34:03

technology around the world? Are we an

34:06

A? Are we a B? or we see what what's

34:07

working, what's not working.

34:09

>> Well, first of all, President Trump

34:12

wants American industry to lead. He

34:15

wants American technology industry to

34:17

lead. He wants American technology

34:19

industry to win. He wants us to spread

34:22

American technology around the world. He

34:24

wants the United States to be the

34:25

wealthiest country in the world. He

34:27

wants all of that. At the current

34:29

moment, as we speak,

34:32

Nvidia gave up a 95% market share in the

34:36

second largest market in the world, and

34:38

we're at 0%.

34:40

>> President Trump, That's right. President

34:42

Trump wants us to get back in there. And

34:45

and uh the first thing is uh to get

34:48

license licensed for the companies that

34:51

we're going to be able to sell to. We've

34:53

got many companies who have requested

34:55

for licenses. We've applied for licenses

34:58

for them and we've got approved licenses

35:00

from sec secretary lutnik. Uh now uh

35:03

we've we informed the Chinese companies

35:05

and many of them have given us purchase

35:07

orders and so we're going to we're going

35:09

to we're in the process of cranking up

35:10

our supply chain again to go ship. I

35:13

think at the highest level Brad um I

35:16

think one of the things that we should

35:17

acknowledge is this. Our national

35:19

security

35:21

is diminished when we don't have access

35:25

to miniature motors, rare earth

35:28

minerals. It's diminished when we don't

35:31

control our telecommunications networks.

35:33

It's diminished when we can't provide

35:35

for sustainable energy for our country.

35:38

It is fundamentally diminished. Every

35:40

single one of these industries is an

35:42

example of what I don't want the AI

35:45

industry to be.

35:46

>> Right? When we look forward in time and

35:49

we say what do we want? What is the what

35:51

does it look like when American

35:53

technology industry American AI industry

35:56

leads the world? We can all acknowledge

35:59

that there is no way that AI models is

36:03

one universally. It is we can all

36:06

acknowledge that that is an outcome that

36:08

makes no sense. However, we can all

36:10

imagine that the American tech stack

36:14

from chips to computing systems to the

36:18

platforms are used broadly by the world

36:21

where they build their own AI, they use

36:24

public AI, they use private AI whatever

36:26

and they can build their applications in

36:28

their society. I would love that the

36:30

American tech stack is 90% of the world.

36:33

Yes, I would love that. The alternative

36:36

if it looks like solar, rare earth,

36:40

magnets, motors, telecommunications, I

36:43

consider that a very bad outcome for

36:45

national security.

36:47

>> Great.

36:47

>> Yeah.

36:48

>> How much are you monitoring the

36:49

situation with the conflicts around the

36:52

world right now? And how much does it

36:53

worry you Jensen? So, China and Taiwan

36:56

and then helium availability coming out

36:58

of the Middle East, I understand, can be

36:59

a supply chain risk to semiconductor

37:01

manufacturing. How much do these

37:03

situations worry you? How much are you

37:05

spending on them?

37:06

>> Well, first of all, I think the in

37:07

Middle East, I have we have 6,000

37:10

families there.

37:10

>> Yeah.

37:11

>> Uh we have a lot of Iranians uh at

37:13

NVIDIA and their families are still in

37:15

Iran. And so so we have we have a lot of

37:17

families there. The first thing is is

37:19

they're quite anxious. They're quite

37:21

concerned, quite scared. Um we're

37:22

thinking about them all the time. Uh

37:24

we're monitoring and keeping an eye on

37:26

them all the time. They have 100% of our

37:28

support. Uh I've been asked several

37:29

times, are we still considering uh being

37:32

in Israel? We are 100% in Israel. We are

37:36

100% behind the families there. We are

37:38

100% in the Middle East. I was also

37:40

asked, you know, given what's happening

37:42

in the Middle East, uh is that an area

37:45

where we believe that we can expand

37:47

artificial intelligence to? Um I believe

37:51

that there's a reason we went to war and

37:53

I believe at the end of the war, Middle

37:55

East will be more stable than before.

37:58

And so if we were there, if we're

38:00

considering it before, we should

38:01

absolutely be considering it after. And

38:03

so I'm 100% in on that. With respect to

38:06

with with with respect to to Taiwan,

38:09

>> we have to do three things. One, we have

38:12

to make sure that we re-industrialize

38:14

the United States as fast as we can.

38:16

>> And whether it's the chip manufacturing

38:18

plants, the the computer manufacturing

38:20

plants, or the AI factories,

38:21

>> how are we doing on that? We're doing

38:23

excellent with by by gaining the

38:27

strategic support by gaining the

38:28

friendship of the supply chain of

38:31

Taiwan.

38:32

By gaining their friendship by gaining

38:35

their support, we were able to build

38:38

Arizona and Texas, California at

38:41

incredible rates. They're they are

38:44

genuinely a strategic partner. Um we we

38:47

we really they deserve our support. They

38:51

deserve our friendship. They deserve our

38:53

generosity and they're doing everything

38:56

they can to accelerate the manufacturing

38:58

process for us. And so, so I think

39:00

that's number one. Number two, we ought

39:02

to diversify the manufacturing supply

39:04

chain. And whether it's South Korea,

39:07

whether it's it's Japan, it's Europe, we

39:09

got to we got to diversify the supply

39:11

chain, make it more resilient. And

39:12

number three, let's be let's let's

39:16

demonstrate restraint. And while we're

39:20

reducing uh increasing our diversity and

39:23

resilience, let's not

39:26

press push um

39:28

>> unnecessary. We need to be patient.

39:31

>> Is helium a problem? A lot of reports,

39:34

>> you know, I think helium could be a

39:35

problem, but it's also the case that the

39:37

supply chain probably has a lot of

39:38

buffer in it.

39:39

>> These kind of things tend to have a lot

39:41

of buffer. Uh but but um you know yeah

39:45

>> you've um made massive progress in

39:48

self-driving. You made a big

39:50

announcement. You've added many more

39:52

partners including BYD. There was just a

39:54

video of you driving around in a

39:55

Mercedes and uh huge announcement uh

39:59

with Uber that you're going to have a

40:01

number of cars on the road from many

40:03

different manufacturers. your bet is I

40:07

believe that there's going to be an

40:08

Android

40:10

type open-source platform that you're

40:12

going to play a major part in with

40:15

dozens of uh car providers and then

40:17

maybe on the other side there could be

40:18

an iOS with Tesla or Whimo. What's your

40:22

strategy thinking there and how that

40:24

chessboard emerges because it feels like

40:27

you have a a pretty deep stack and in

40:30

some ways you're competing and in other

40:32

places you're collaborative. Yeah. Um,

40:36

it's taking a step back. We believe that

40:39

everything that moves will be autonomous

40:42

completely or partly

40:44

someday. Number one. Number two, we

40:46

don't want to build self-driving cars,

40:48

but we want to enable every car company

40:50

in the world to build self-driving cars.

40:52

And so, we built all three computers,

40:53

the training computer, the simulation

40:55

computer, the valu evaluation computer,

40:57

as well as the car computer. We develop

41:00

the world's safest driving operating

41:03

system. Uh we also created the world's

41:06

first reasoning autonomous vehicle so

41:09

that it could decompose complicated

41:11

scenarios into simpler scenarios that it

41:14

knows how to navigate through just like

41:16

us reasoning systems. And so that

41:18

reasoning system called Alpommyo has

41:21

enabled us to achieve incredible

41:22

results.

41:25

We

41:26

open this we ver we vertical

41:29

optimization. We horizontally innovate

41:31

and we let everybody decide. Do you want

41:34

to buy one computer from us? In the case

41:35

of Elon and Tesla, they buy our training

41:37

computers. Um, do they want to buy our

41:39

training computer and our simulation

41:41

computers or do you want to let us uh

41:43

work with us to do all three and even

41:45

put the car computer in your car. So, we

41:47

you know, our attitude is we want to

41:50

solve the problem.

41:52

We're not the solution provider

41:55

and we're delighted however you work

41:57

with us. Let me build on this question

41:59

because I think it's like it's so

42:00

fascinating. You actually do create this

42:02

platform. A thousand flowers are

42:04

blooming.

42:05

>> But it's also true that some of those

42:07

flowers want to now go back down in the

42:09

stack and try to compete with you a

42:11

little bit. Google has TPU, Amazon has

42:14

inferentia and tranium. You know,

42:16

everybody's sort of spinning up their

42:18

own version of I think I can out Nvidia

42:20

Nvidia

42:21

>> even though they also tend to be huge

42:23

customers.

42:24

>> How do you navigate that? And yeah, what

42:26

do you think happens over time and

42:29

>> where do those things play in the

42:30

complexion of this kind of vision?

42:32

>> Yeah, really great. You know, first of

42:33

all, um, we're the only AI company,

42:36

we're an AI company. We build foundation

42:39

models. We're at the frontier in many

42:40

different domains. We build every single

42:42

every single layer, every single stack.

42:45

Um, we're the only AI company in the

42:46

world that works with every AI company

42:48

in the world. They never show me what

42:50

they're building and I always show them

42:52

exactly what I'm building.

42:53

>> Right.

42:54

>> Yeah. And so so the confidence comes

42:57

from this one. Uh we are delighted to

43:01

compete on what is the best technology

43:04

and to the extent that to the extent

43:06

that we can continue to run fast I

43:08

believe that buying from Nvidia still is

43:11

one of the most economic things they

43:12

could do and that's just incredible

43:14

confidence there. Number one. Number two

43:16

we're the only architecture that could

43:17

be in every cloud and that gives us some

43:19

fundamental advantages. where the only

43:21

architecture you could take from a cloud

43:23

and put into onrem in the car in any

43:26

region

43:26

>> in space.

43:27

>> That's right. In space. And so there's a

43:29

whole whole part of our market about 40%

43:32

of our of our business most people don't

43:34

realize this 40% of our business unless

43:37

you have the CUDA stack unless you can

43:38

build an entire AI factory you have the

43:40

customers don't know what to do with

43:41

you. They're not trying to build chips.

43:44

They're not trying to buy chips. They're

43:46

trying to build AI infrastructure. And

43:48

so they want you to come in with the

43:49

full stack. And we've got the whole

43:50

stack. And so surprisingly, Nvidia is

43:53

gaining market share. If you look at

43:56

where we are today, we're gaining share.

43:57

>> Do you think what happens is these guys

43:59

try and they realize, oh my god, it's

44:01

too much. And then they come back. Is

44:02

that why the share grows?

44:04

>> Well, we're gaining share for several

44:05

reasons. One, um, our velocity has gone.

44:09

We help people realize it's not about

44:11

building the chip, it's about building

44:13

the system.

44:14

>> And that system is really hard to build.

44:16

uh and and so their their their business

44:18

with us is increasing. In the case of

44:20

AWS, I think they just announced, I

44:22

think it was yesterday, that they're

44:24

going to buy a a million chips uh in the

44:27

next couple years. I mean, that's a lot

44:29

of chips from from AWS. And that's on

44:31

top of all the chips they've already

44:32

bought. And so, we're delighted to do

44:34

that. But number one, we're gaining

44:36

share this last couple years because we

44:39

now have Anthropic coming to Nvidia.

44:42

Meta SL is coming to Nvidia. And the

44:46

growth of open models is incredible. And

44:49

that's all on Nvidia. And so we're

44:51

growing in share because of the number

44:53

of models. We're also growing in share

44:55

because out all of these companies are

44:58

outside of the cloud and they're growing

45:01

regionally in enterprise in industries

45:03

at the edge and that entire segment of

45:06

growth is you know really hard to do if

45:08

it's just building an as

45:09

>> Brad

45:10

>> related to that um and not to get in the

45:13

weeds on the numbers but analysts don't

45:15

seem to believe right so if you look at

45:17

the consensus forecast you said compute

45:20

could 1 millionx right and Yet they have

45:23

you growing next year at 30%, the year

45:25

after that at 20%. And in 2029, which is

45:28

supposed to be a monster year at 7%.

45:31

Right? So if you just if you take your

45:34

TAM and you apply their growth numbers,

45:36

it suggests that your share will

45:38

plummet. Do you see anything in your

45:41

future order book that would make that

45:43

correct?

45:44

>> Yeah. First of all, they just don't

45:46

understand the scale and the breadth of

45:48

AI.

45:49

>> Yes.

45:49

>> Yeah.

45:50

>> Yeah. I think that's true. Most people

45:52

think that AI is in the top five

45:54

hyperscalers,

45:55

>> right? That's right. There's also an

45:57

orthodoxy around these law of large

45:59

numbers where,

46:00

>> you know, they have to go back to their

46:02

investment banking risk committee and

46:03

show some model.

46:05

>> They're not going to believe in their

46:07

minds that 5 trillion goes to 15

46:09

trillion. They're like go to it can go

46:11

to seven or they can have a 10 trillion

46:14

company.

46:14

>> It's all just CIA stuff that I think

46:16

>> it's never happened before. So you can't

46:17

say it will

46:18

>> and and because because you have to

46:20

redefine what it is that you do. There

46:22

was somebody who made an observation

46:23

recently that Nvidia

46:26

Jensen, how can you be larger than Intel

46:29

in servers and the reason for that is

46:32

because the CPU market of the entire

46:35

data center was about $25 billion a

46:37

year,

46:37

>> right?

46:38

>> We do $25 billion a year as you guys

46:40

know in a very in the time that we were

46:42

sitting here.

46:43

>> And so obviously obviously

46:47

That was a joke.

46:48

>> No, it's but it's

46:49

>> all in podcast.

46:51

>> Don't worry. Everything on this show is

46:52

rough. Don't worry about it. It's all in

46:55

here. Anyways, that was not guidance.

46:58

But anyhow, anyhow, it the the point is

47:01

how big you can be

47:02

>> depends on what is it that you make,

47:05

>> right?

47:05

>> Nvidia is not making chips. Number one,

47:08

making chips does not help you solve the

47:10

AI infrastructure problem anymore. It's

47:12

too complicated. Number three, most

47:15

people think that AI is narrowly in the

47:17

things that they talk about and hear and

47:19

see.

47:20

>> It's AI is much open AI is incredible.

47:23

They're going to be enormous. Anthropic

47:25

is incredible. They're going to be

47:26

enormous. But AI is going to be much

47:29

much bigger than that.

47:30

>> Tell us

47:31

>> and we addressed that segment.

47:32

>> Tell us about data centers in space for

47:34

a second.

47:34

>> Yeah.

47:35

>> Um

47:36

>> we're already in space. How should the

47:38

layman think about what that business is

47:41

versus when you hear about these big

47:43

data center buildouts that's happening

47:44

in in on the ground?

47:47

>> Well, we should definitely work on the

47:48

ground first because we're already here

47:50

and number one. Number two, we should

47:52

prepare to be out in space and obviously

47:54

there's a lot of energy in space. Um the

47:57

challenge of course is that cooling

48:00

you can't take advantage of conduction

48:02

and convection and so you can only use

48:04

radiation and radiation requires very

48:07

large surfaces and so now that's not an

48:09

impossible thing to solve and there's a

48:11

lot of lot of space in space. Um but

48:13

nonetheless

48:15

the expense is still quite there is is

48:17

there uh we're going to go explore it.

48:19

We're already there. We're already

48:20

radiation hardened. Uh we have we have

48:23

uh uh uh CUDA in satellites around the

48:26

world. Um they're doing imaging, image

48:28

processing, AI imaging and um and that

48:31

kind of stuff ought to be done in space

48:33

instead of sending all the data back

48:34

here and do imaging down here. We ought

48:36

to just do imaging out in space. And so

48:38

there's a lot of things that we ought to

48:39

done do do in space. And in the

48:40

meantime, uh we're going to explore what

48:42

is the architecture of data centers look

48:44

like uh in space. And it'll take it'll

48:46

take years. It's okay. We got I got

48:48

plenty of time. I wanted to um double

48:50

click on healthcare. I know you've got a

48:52

big effort there. We're all of a certain

48:53

age where we're thinking about lifespan,

48:56

health span. I mean, we all look great.

48:58

I think

48:59

>> some better than others.

49:00

>> I think some better than others. I don't

49:01

know what your secret is, Jensen.

49:03

>> Pretty good these these

49:04

>> I mean what's what are you taking what's

49:06

off the menu? You got to talk to me when

49:08

we're backstage. I want to know in the

49:09

green room what you got going on.

49:11

>> Squat squats and push-ups and sits.

49:12

>> Perfect. Okay. Um but

49:15

>> that works. what you know in terms of

49:18

the buildout in healthcare

49:21

where is that going and what kind of

49:24

progress are we making? I was just using

49:26

Claude to do some analysis and saying

49:28

like where are all these billing codes?

49:29

We spend twice as much money in the US.

49:31

We get seem to get half as much. It

49:34

seemed like uh 15 to 25% of the dollar

49:37

spent were on these first GP visits. And

49:40

I think we all know like chat GBT and a

49:43

large language model does a better job

49:45

more consistently today at a first

49:47

visit. So what has to happen there to

49:50

kind of break through all that

49:52

regulation and have AI have a true

49:53

impact on the health care system?

49:55

>> There's several several areas that we're

49:57

involved in in um in healthcare. One is

50:02

uh AI

50:04

uh physics uh and and that's or AI

50:07

biology using AI to understand represent

50:11

predict biology behavior biological

50:13

behavior and so that's one that's very

50:15

important in drug discovery. There's

50:17

second which is AI agents and that's

50:20

where the assistance and helping

50:21

diagnosis and things like that. Open

50:23

evidence is a really good example.

50:25

Hypocratic is a really good example.

50:26

Love working with those companies. Um I

50:28

really think that this is an area uh

50:30

where agentic technology is going to

50:32

revolutionize how we interact with

50:35

doctors and how do we interact for

50:36

healthcare. The third part that we're in

50:38

involved in is physical AI. The first

50:40

one is AI physics using AI to predict

50:42

physics. The second one is physical AI.

50:45

AI that understand the properties of the

50:47

laws of physics and that's used for a uh

50:50

robotic surgery huge amounts of

50:53

activities there. Every single

50:54

instrument whether it's ultrasound or

50:57

you know CT or whatever instrument we

50:59

interact with in a hospital in the

51:01

future will be agentic.

51:02

>> Yeah.

51:02

>> You know open claw in a safe version

51:05

will be inside every single instrument.

51:07

And so in a lot of ways that instrument

51:09

is going to be interacting with patients

51:11

and nurses and doctors in a very unique

51:13

way. so much investment in AI weapons.

51:15

It would be wonderful to see some

51:17

investment in AI EMTs and paramedics and

51:21

saving lives, not just taking them,

51:23

which I think is a great segue into

51:25

robotics. You've got dozens of partners.

51:28

We have this very weird

51:30

>> I I don't know want to call a lost

51:31

decade or 20 years of Boston Dynamics.

51:34

Google bought a bunch of companies. They

51:36

then wound up selling them and spinning

51:37

them out where people just thought

51:40

robotics is just not ready for prime

51:41

time. And now here we have the world's

51:43

greatest entrepreneur at this time. Uh

51:45

tied with you, uh Elon Musk doing well,

51:48

that was a good save, I hope. Optimus,

51:51

uh pretty impressive. And then other

51:53

companies in China. How how close is

51:55

that to actually being in our lives

51:59

where we might see a chef, a robotic

52:03

chef, a robotic nurse, a robotic

52:05

housekeeper, you know, this humanoid

52:07

factor actually working in the real

52:09

world knowing what you know with those

52:12

partners and the fidelity, especially in

52:14

China where they seem to be doing as

52:15

good a job as we're doing here or maybe

52:16

better.

52:18

>> Um,

52:20

we invented the industry largely.

52:22

America invented. We c you could argue

52:24

we got into it too soon.

52:26

>> Yeah.

52:26

>> And and we got exhausted. We got tired

52:30

um about five years before the enabling

52:33

technology appeared.

52:34

>> The brain.

52:35

>> Yeah. Yeah. And we we just got tired of

52:37

it just a little too soon. Okay. That's

52:40

number one. But it's here now. Now the

52:42

question is how much longer? From the

52:44

point of high functioning existence

52:47

proof, high functioning exist existence

52:49

proof to reasonable products

52:53

technology never takes more than a

52:55

couple two three cycles. And so a couple

52:58

two three cycles basically be somewhere

53:00

around three years to 5 years. That's

53:02

it. 3 years to 5 years we're going to

53:03

have robots all over the place. Uh I

53:05

think I think um uh China is is uh

53:08

formidable and the reason for that is

53:10

because their micro electronics, their

53:14

uh motors, their rare earth, their

53:16

magnets, which is foundational to

53:17

robotics, they are the world's best. And

53:20

so in a lot of ways, our robotics

53:22

industry relies deeply on their

53:24

ecosystem and their supply chain. Um and

53:27

uh and and they're, you know, obviously

53:29

moving very quickly. Uh we're going to,

53:31

you know, our robotics industry will

53:33

have to rely a lot on it. the world's

53:35

robotics industry will have to rely on a

53:36

lot on it. And so so I think um you're

53:40

gonna see some fast fast movements here

53:42

>> ultimately one for one. Elon seems to

53:44

think we're going to have one robot for

53:46

every human. 7 billion for 7 billion, 8

53:48

billion for 8 billion.

53:49

>> Well, I'm hoping more. Yeah, I'm hoping

53:51

more. Yeah. Uh well, first of all,

53:53

there's a whole bunch of robots that are

53:55

going to be in factories working around

53:56

the clock. There's going to be a whole

53:58

bunch of fac that that don't move. They

54:01

move a little bit. Uh almost everything

54:03

will be robotic. What does the world

54:04

look like?

54:05

>> Sorry, let me I think like this is one

54:06

of the robotics for me is one of the

54:08

pieces that I think unlocks uh economic

54:11

mobility opportunities for every

54:13

individual. Everyone now like when

54:15

everyone got a car, they could now go

54:17

and do a lot of different jobs. When

54:18

everyone gets a robot, their robot can

54:21

do a lot of work for them. They can

54:22

stand up an Etsy store, a Shopify store.

54:25

They can create anything they want with

54:27

their robot. They could do things that

54:29

they independently cannot do. I think

54:31

the robot is going to end up being the

54:32

greatest unlock for prosperity for more

54:35

people on Earth than we've ever seen

54:37

with any technology before.

54:38

>> Yeah, no doubt. I mean, just a simp the

54:41

simple math at the moment is we're

54:42

millions of people short in labor today.

54:44

Right. Yeah.

54:45

>> Right. We're we're we're actually really

54:47

desperate in need of robotics and so

54:50

that all of these companies could grow

54:52

more if they had more labor. I mean,

54:55

we're we're number one. Some of the

54:57

things that you mentioned are super fun.

54:59

I mean, because of robots, we'll have

55:01

virtual presence. Uh, you know, I'll be

55:03

able to go into the robot of my house

55:07

and virtually operate it. I'm on a

55:10

business trip,

55:11

>> right?

55:12

>> Walk around the house.

55:13

>> Yeah. Walk the dog.

55:14

>> Yeah. Walk the dog.

55:14

>> Break the leaves.

55:15

>> Yeah. Exactly. Freak out the dog.

55:17

>> Maybe not quite that, but just, you

55:18

know, just, you know, wander around and

55:21

just see what's going on in the house.

55:22

You know, chat with the dogs, chat with

55:24

the kids. Yeah.

55:25

>> Yeah.

55:26

And time travel is also we're going to

55:27

be able to travel at the speed of light,

55:29

you know, and so, you know, clearly

55:31

we're going to send our robots ahead of

55:32

us.

55:33

>> Yeah.

55:33

>> Not going to send myself. I'm going to

55:35

send a robot, you know.

55:36

>> Check it out.

55:36

>> Yeah. Yeah. And then I'm going to upload

55:38

my AI.

55:39

>> Well, it's inevitable. It unlocks the

55:40

moon and it unlocks Mars as um targets

55:43

for for colonization, which gives us

55:45

>> infinite resources. Getting back from

55:47

the moon is effectively zero energy cost

55:49

to move material back because you can

55:51

use solar and accelerate. So you could

55:53

have factories that make everything the

55:54

world needs on the moon and the robots

55:56

are going to be the unlock for enabling.

55:58

>> That's right. Distance no longer

55:59

matters.

55:59

>> Distance doesn't matter. Yeah.

56:01

>> The more the more revenue we get out of

56:03

models and agents, the more we can

56:05

invest in building the infrastructure

56:07

which then unlocks more capabilities on

56:09

models and agents. Dario on Dwaresh's

56:12

podcast recently said by 2728 we'll have

56:15

hundreds of billions of dollars of

56:17

revenue out of the model companies and

56:18

the agent companies. and he forecasts a

56:20

trillion dollars by 2030, right? This is

56:23

non-infrastructure AI revenue. Um,

56:27

>> I think he I think he's he's being very

56:29

conservative. I believe Dario and

56:31

Anthropic is going to do way better than

56:33

that.

56:33

>> Wow.

56:34

>> Way better than that.

56:35

>> Wow. So, from 30 billion to a trillion.

56:37

>> Yeah. and not and and the reason for

56:38

that is the one part that he hasn't

56:40

considered is that I believe every

56:43

single enterprise software company will

56:46

also be a reseller

56:49

value added reseller of anthropic code

56:51

anthropics tokens value added reseller

56:54

open AI that's right and they're going

56:57

to that that that part of their

56:59

>> get this logarithmic expansion

57:01

>> yes

57:02

>> their go to market is going to expand

57:04

tremendously this year

57:06

>> what do you think in that world is the

57:08

moat what's left over. I mean you have

57:10

some moes that are frankly I think as

57:12

this scales almost insurmountable the

57:15

best one that nobody talks about is

57:17

probably CUDA which is just like an

57:19

incredible strategic advantage. But in

57:22

the future if a model can be used to

57:25

create something incredible then the

57:27

next spin of a model can be used to

57:29

maybe disrupt it. Sort of in your mind

57:31

what do you think for these companies

57:32

that are building at that application

57:34

layer? What's their moat? like how do

57:36

they differentiate themselves?

57:37

>> Deep specialization.

57:39

Deep specialization. I believe that um

57:43

these models they're going to have

57:45

general general models that are

57:46

connected into the software company's

57:49

agentic system,

57:51

>> right?

57:51

>> Many of those models are cloud models

57:54

and proprietary models, but many of

57:56

those models are specialized

57:59

sub aents that they've trained on their

58:02

own.

58:03

>> Right. All right. So, the call to arms

58:04

for you for entrepreneurs is look,

58:06

>> know your vertical.

58:07

>> That's right.

58:08

>> Know it as deep and as better than

58:10

everybody else.

58:11

>> That's right.

58:11

>> And then wait for these tools because

58:12

they're catching up to you and now you

58:14

can imbue it with your knowledge.

58:15

>> That's right. The sooner you connect

58:17

your agent,

58:18

>> the sooner you connect your agent with

58:20

customers,

58:21

>> that flywheel is going to cause your

58:23

agent to get

58:24

>> it very much is an inversion of what we

58:26

do today because today we build a piece

58:27

of software and we say what generalizes

58:30

>> and then let's try to sell it as broadly

58:32

as possible and then sell the

58:33

customization around it

58:34

>> and we in fact in fact exactly right we

58:37

we create a horizontal but notice there

58:41

are all these gsis and all of these

58:43

consultants who are specialists Yes.

58:45

>> Who then take your horizontal platform

58:47

and specializes it into

58:49

>> Exactly.

58:50

>> And that's arguably a five or six time

58:51

bigger industry is the customization.

58:53

>> It is absolutely the whole very much is

58:56

>> that's right. So I think that these

58:58

platform companies have an opportunity

59:00

to become that specialist to become that

59:02

vertical.

59:03

>> Yeah. Domain expert.

59:04

>> You know, I just want to give you your

59:05

flowers. I think it was three years ago

59:07

you said you're not going to lose your

59:08

job to AI. You're going to lose your job

59:10

to somebody using AI. And here we are.

59:12

The entire conversation has revolved

59:14

around this concept of agents making

59:17

people superhuman and the business

59:20

opportunity expanding and

59:21

entrepreneurship expanding. You actually

59:23

saw it pretty clearly. Yeah.

59:24

>> Have you changed your view?

59:26

>> I do. I'm not I'm not doomer. I do I do

59:31

have doomer.

59:31

>> No I you can hold space for I think two

59:33

ideas. One is there are going to be a

59:35

lot

59:35

>> that's spiral Jake we call it.

59:36

>> No there you can

59:37

>> but that's just because he doesn't hang

59:39

out with me enough. Well, we I mean we

59:41

a little bit. Be careful.

59:42

>> We don't talk about it.

59:44

>> He will show your breakfast. HE'LL

59:46

FOLLOW YOU AROUND.

59:46

>> I'm not asking for it. He'll follow you

59:48

around. I'm not asking for it.

59:50

>> You can come with me and Tucker. We ski

59:52

in Japan every January. Love it. Me and

59:54

Tucker go road trip. Um there is going

59:57

to be job displacement. And then the

59:59

question becomes,

60:01

>> you know, do those people have the

60:03

fortitude, the resolve to then go

60:05

embrace these,

60:06

>> you know, technologies? We're we're

60:08

going to see 100% of driving go away by

60:10

humans. That's just it's that's a

60:12

beautiful thing in the lives saved, but

60:14

we have to recognize that's 15 million

60:16

people in the United States, 10 to 15

60:18

million who are employed in that way.

60:20

And and so that is going to happen. Yes,

60:22

>> I I think I think that jobs will change.

60:24

For example, um there are many

60:26

chauffeers today uh who drives the car.

60:29

I believe that though many of those

60:31

chauffeers will actually be in the car

60:34

sitting behind the drive the steering

60:36

wheel while the car is driving by

60:38

itself. And the reason for that is

60:40

because remember what a chauffeur does

60:42

in the end. These chauffeers they're

60:44

helping you they're your assistants.

60:45

They're helping you with your luggage.

60:47

They're helping you. I mean, they're

60:48

helping you with a lot of things and and

60:50

so I wouldn't be surprised actually if

60:53

the chauffeers of the future becomes

60:55

your mobility assistant and they are

60:57

helping you do on a whole bunch of other

60:59

stuff to the hotel.

61:00

>> Yeah. And the car is driving by itself.

61:02

>> The autopilot in planes created a lot

61:05

more pilots and didn't take any of the

61:07

pilots out of the cockpit even though

61:09

the autopilot is flying the plane 90% of

61:11

the time. And by the way, while that car

61:12

is driving itself, that chauffeur is

61:14

going to be doing a bunch of other work

61:16

on his phone and he's going to be

61:17

>> arranging, for example, coordinating a

61:20

bunch of things for you, getting, you

61:21

know,

61:21

>> it's all the pie just grows in a way

61:23

that

61:24

>> one of the things that that that

61:26

yes, every job will be will be

61:28

transformed. Um, some jobs will be

61:30

eliminated. However, we also know that

61:32

many many jobs will be recre will be

61:34

created. The one thing that I will say

61:36

to young people who are coming out of

61:37

school who are concerned who are anxious

61:39

about AI be the expert of using AI

61:43

>> how much look we all want our employees

61:46

to be expert at using AI and it's not

61:49

not

61:51

>> not trivial not trivial and so knowing

61:54

how to specify not to overprescribe

61:58

leaving enough room for the AI to

62:00

innovate and create while we guide it to

62:03

the outcome we want. it. All of that

62:06

requires artistry.

62:07

>> You had you had this great advice to

62:09

when you were at Stanford, I think it

62:10

was, which is I wish to you pain and

62:12

suffering. Do you remember that?

62:13

>> Yeah.

62:13

>> Fantastic.

62:14

>> What's your advice to young people

62:16

around what they should be studying? So,

62:18

if they're sort of about to leave high

62:20

school because now those are the kids

62:22

that are at this like really native,

62:24

they haven't made a decision about

62:25

college, what to study, if at all go to

62:27

college. How do you guide those kids?

62:30

What would you tell them? I I still

62:32

believe that deep science, deep math, um

62:36

language skills, you know, as you know,

62:39

language is the programming language of

62:42

AI, the ultimate program.

62:44

>> And so, as it turns out, it it could be

62:46

that the English major could be the most

62:47

successful. Yeah.

62:48

>> And and so so I think I think um I I

62:52

would just advise whatever whatever

62:54

education you get, just make sure that

62:56

you're deeply deeply expert in using

62:58

AIs. One of the things that I wanted to

63:00

say with respect to jobs and I want

63:02

everybody to hear it that in fact at the

63:04

beginning of the deep learning

63:06

revolution, one of the the finest

63:09

computer scientists in the world deeply

63:10

deeply I deeply uh deeply uh um respect

63:15

uh predicted that computer vision will

63:17

completely eliminate radiologists

63:20

and and that the one at the one field he

63:23

advises everybody to not go into is

63:25

radiology. 10 years later, his

63:29

prediction was at 100% right. Computer

63:32

vision has been integrated into all of

63:34

the radiology technologies and radiology

63:36

platforms in the world 100%. The

63:39

surprising outcome is the number of

63:41

radiologists actually went up and the

63:43

demand for radiologists is skyrocketed.

63:47

The reason for that is because

63:48

everybody's job

63:51

has a purpose and its task. The task

63:55

that you do is studying the scans,

63:59

>> but your purpose is to diag helping the

64:01

doctors, helping the patient diagnose

64:03

disease.

64:04

>> And so what's surprising is because the

64:07

scans are now being done so quickly,

64:09

>> they could do more scans, improving

64:12

healthcare.

64:12

>> Yes.

64:13

>> But doing more scans more quickly allows

64:15

patients to

64:17

>> be

64:18

onboarded a lot more quick, treated a

64:21

lot more quickly. And as it turns out,

64:23

because hospitals enjoy making money,

64:25

too.

64:26

>> Yeah.

64:27

>> Right.

64:27

>> They're doing more scans.

64:29

>> They're treating more customers.

64:32

The revenues go up. And guess what?

64:34

Perfect. And and a country that grows

64:37

faster, productivity increases. A

64:39

wealthier country can put more teachers

64:41

in the classroom, not less teachers in

64:43

the classroom. That's right. You just

64:44

give every one of those teachers a

64:46

personalized curriculum for every

64:47

student in the room. It makes them all

64:49

bionic and leads to a lot more. Every

64:51

single student will be assisted by AI,

64:54

but every single student will need great

64:56

teachers.

64:56

>> Yeah. Yeah. Amazing. Uh Jensen,

64:59

congratulations. I know your success and

65:01

really this is an incredibly positive,

65:02

uplifting discussion. We really

65:04

appreciate you taking the time for us.

65:06

He is the steward we need.

65:08

>> You are you are you need to be more

65:09

vocal. I'm being very vocal about the

65:12

positive side of it. I think there's too

65:13

much dumerism is

65:14

>> but I also think it takes the humility

65:16

to have this level of success and be

65:18

humble about we're making software guys.

65:21

Yeah.

65:22

>> And I think that that's actually really

65:24

healthy for people to hear. We have done

65:26

this before. We have invented categories

65:28

and industries before.

65:29

>> We don't need to go to this

65:31

>> scaremongering place. It does nothing.

65:34

>> And we get to choose, right? We have

65:35

autonomy and and agency. We get to pick

65:38

how to

65:38

>> we do this. Okay, everybody. We'll see

65:40

you next time on the All-In interview.

65:43

Okay.

65:44

>> Well done, brother.

65:45

>> Thanks, man.

65:46

>> Good job.

65:47

>> Thank you, sir. That was awesome.

65:48

>> Good. Good.

65:49

>> Appreciate you. You guys are awesome.

65:51

>> Look at this. Look at this big crowd

65:52

behind you guys,

65:53

>> man. I think they're here for you.

Interactive Summary

Loading summary...