HomeVideos

GPUs, TPUs, & The Economics of AI Explained

Now Playing

GPUs, TPUs, & The Economics of AI Explained

Transcript

2548 segments

0:00

I will never forget when I first met

0:02

Gavin Baker. It was early days of the

0:04

podcast and he was one of the first

0:06

people I talked to about markets outside

0:07

of my area of expertise which at the

0:09

time was quantitative investing about

0:11

the incredible passionate experience

0:14

that he's had investing in technology

0:16

across his career. I find his interest

0:18

in markets, his curiosity about the

0:20

world to be about as infectious as any

0:22

investor that I've ever come across. He

0:24

is encyclopedic on what is going on in

0:27

the world of technology today. and I've

0:28

had the good fortune to host him every

0:31

year or two since that first meeting on

0:33

this podcast. In this latest

0:34

conversation, we talk about everything

0:36

that interests Gavin. We talk about

0:38

Nvidia, Google and its TPUs, the

0:40

changing AI landscape, the changing math

0:43

and business models around AI companies.

0:45

This is a life ordeath decision that

0:48

essentially everyone except Microsoft is

0:50

failing it. We even discussed the crazy

0:53

idea of data centers in space which he

0:55

communicates with his usual passion and

0:57

logic.

0:58

>> In every way, data centers in space from

1:02

a first principles perspective are

1:04

superior to data centers on earth.

1:06

Because Gavin is one of the most

1:08

passionate thinkers and investors that I

1:10

know, these conversations are always

1:12

amongst my most favorite. I hope you

1:13

enjoy this latest in a series of

1:15

discussions with Gavin Baker.

1:19

I would love to talk about how you like

1:22

in the nitty-gritty process new things

1:24

that come out in this whole like AI

1:27

world because it's happening so

1:28

constantly. I'm extremely interested in

1:30

it and I find it very hard to keep up

1:31

and I you know I have a couple blogs

1:33

that I go read and friends that I call

1:35

but like maybe let's take Gemini 3 as

1:37

like a recent example when that comes

1:39

out. What like literally like take me

1:41

into your office like what are you

1:42

doing? How do you and your team process

1:45

an update like that given how often

1:47

these things are happening?

1:48

>> I mean, I think the first thing is you

1:49

have to use it yourself.

1:50

>> And I would just say I'm amazed at how

1:53

many famous and August investors are

1:56

reaching really definitive conclusions

1:58

about AI. Well, no, based on the free

2:00

tier.

2:01

>> The free tier is like you're dealing

2:03

with a 10-year-old

2:04

>> and you're making conclusions about the

2:06

10-year-old's capabilities as an adult.

2:09

And you could just pay and I do think

2:11

actually you do need to pay for the

2:12

highest tier whether it's Gemini Ultra,

2:14

you know, um, Super Grock, whatever it

2:16

is, you have to pay the $200 per per

2:18

month ti whereas those are like a

2:20

fullyfledged 30 35y old. It's really

2:23

hard to extrapolate from an eight or a

2:25

10-year-old to the 35y old and yet a lot

2:27

of people are doing that. And the second

2:28

thing is there was a insider post about

2:30

open AI and they said to a large degree

2:33

open AI runs on Twitter vibes

2:36

>> and I just think AI happens on X and you

2:43

know there have been some really

2:44

memorable moments like there was a giant

2:46

fight between the PyTorch team at Meta

2:49

and the Jax team at Google on X and the

2:52

leaders of each lab had to step in

2:55

publicly say

2:57

>> no one from my lab is allowed to say bad

2:59

things about the other lab and I respect

3:01

them and that is the end of that.

3:02

>> Yeah,

3:03

>> the companies are all commenting on each

3:05

other's posts. You know, the research

3:06

papers come out. You know, if on planet

3:08

Earth there's 500 to a,000 people who

3:11

really really understand this and are at

3:13

the cutting edge of it and a good number

3:16

of them live in live in China. Um I just

3:19

think you have to follow those people

3:20

closely

3:21

>> and I think there is incredible signal

3:24

to me. Everything in AI is just

3:26

downstream

3:28

>> of those people.

3:29

>> Yeah. Everything Andre Carpathy writes,

3:31

you have to read it three times.

3:32

>> Yeah.

3:33

>> Minimum.

3:33

>> Yeah. He's incredible.

3:34

>> And then I would say anytime at one of

3:36

those labs, the four labs that matter,

3:39

you know, being uh OpenAI, Gemini,

3:42

Anthropic, and XAI, which are clearly

3:43

the four leading labs. Like anytime

3:45

somebody from one of those labs goes on

3:47

a podcast, I just think it's so

3:49

important to listen. And then for me,

3:52

for me, one of the best use cases of AI

3:55

is to keep up with all of this. You

3:58

know, just like listen to a podcast and

4:00

then if there are parts that I thought

4:01

were interesting, just talk about it

4:04

with AI. And I think it's really

4:05

important to like have as little

4:07

friction as possible, I'll bring it up.

4:08

You know, I have it like um you know, I

4:10

have I can either press this button and

4:14

pull up Gro or I have this.

4:15

>> Oh, wow. I don't touch that. That just

4:17

get brings it right up.

4:18

>> Yeah, it brings it right up. What do you

4:19

think of Patrick Oanaughy?

4:21

>> Oh man, Patrick Oshanaugh is one of my

4:24

favorite voices in investing. His Invest

4:26

Like the Best podcast is straight fire.

4:29

Does deep dives with folks like Bill

4:31

Gurly or

4:32

>> Girly? Yes.

4:34

>> It's so Can you believe we have this?

4:37

>> I know. It's like we have Yeah. I think

4:40

somebody said on on on X, you know, like

4:43

we imbued these rocks with crazy spells

4:48

and now we can summon super intelligent

4:51

genies

4:53

on our phones over the air. You know,

4:54

it's crazy

4:55

>> crazy.

4:55

>> So something like Gemini come 3 comes

4:57

out, you know, the public interpretation

4:59

was, oh, this is interesting. It seems

5:01

to say something about scaling laws and

5:03

the pre-training stuff. What is your

5:05

frame on like the state of prog general

5:08

progress in frontier models in general?

5:10

Like what are you watching most closely?

5:12

>> Yeah. Well, I do think Gemini 3 was very

5:14

important because it showed us that

5:16

scaling laws for pre-training are

5:17

intact. They, you know, stated that

5:19

unequivocally and that's important

5:21

because no one on planet Earth knows how

5:24

or why scaling laws for pre-training

5:26

work. It there it's actually not a law.

5:28

It's an empirical observation and it's

5:30

an empirical observation that we've

5:31

measured extremely precisely and has

5:33

held for a long time. But our

5:34

understanding of scaling laws for

5:36

pre-training and maybe this is a little

5:38

bit controversial with 20% of

5:39

researchers but probably not more than

5:41

that is kind of like the ancient British

5:43

people's understanding of the sun are

5:45

the ancient Egyptians understanding of

5:46

the sun. They can measure it so

5:48

precisely that the east west axis of the

5:50

great pyramids are perfectly aligned

5:52

with the equinoxes and so are the east

5:54

axises of Stonehenge. Perfect

5:57

measurement.

5:58

>> But they had they didn't understand

6:00

orbital mechanics. They had no idea how

6:03

or why, you know, it, you know, rose in

6:05

the east, set in the west, and, you

6:07

know, kind of moved across the horizon.

6:09

>> The aliens.

6:10

>> Yeah. Our god in a chariot. And so it's

6:13

really important every time we get a

6:15

confirmation of that.

6:16

>> Um, so Gemini 3 was very important in

6:18

that way. But I'd say I think there's

6:20

been a big misunderstanding maybe in the

6:22

public equity investing community or the

6:24

broader more generalist community based

6:26

on the scaling laws of pre-training.

6:28

There really should have been no

6:30

progress in 24 and 25.

6:32

>> And the reason for that is, you know,

6:34

after XAI figured out how to get um

6:37

200,000 hoppers coherent,

6:40

>> you had to wait for the next generation

6:42

of chips.

6:43

>> Um because you really can't get more

6:44

than 200,000 hoppers coherent. And

6:46

coherent just means you could just think

6:47

of it as each GPU knows what every other

6:50

GPU is thinking. They kind of are

6:51

sharing memory. You know, they're

6:52

connected. They scale up networks and

6:54

scale out. and um and they have to be

6:56

coherent um for during the pre-training

6:59

process. And I think there's a lot of

7:01

misunderstanding about Gemini 3 that I

7:02

think is really important. So everything

7:04

in AI has a struggle between Google and

7:06

Nvidia and Google has a TPU and Nvidia

7:10

has their GPUs and each of I mean Google

7:12

only has a TPU and they use a bunch of

7:13

other chips for networking. You know

7:15

Nvidia has the full stack and Blackwell

7:18

was delayed. Blackwell was Nvidia's next

7:20

generation chip and

7:23

the first iteration of that was the

7:25

Blackwell 200. A lot of different SKs

7:27

were cancelled and the reason for that

7:29

is it was by far the most complex

7:32

product transition we've ever gone

7:33

through in technology. Going from Hopper

7:36

to Blackwell, first you go from air

7:38

cooled to liquid cooled. Um the rack

7:40

goes from weighing round numbers 1,000

7:42

lb to 3,000 lb. goes from round numbers

7:46

30 kilowatts which is 30 American homes

7:48

to 130 kilowatts which is 130 American

7:50

homes you know. So I I analogize it to

7:53

imagine if to get a new iPhone you had

7:56

to change all the outlets in your house

7:58

to you know 220 volt put in a Tesla

8:01

power wall put in a generator put in

8:04

solar panels that's the power you know

8:06

put in a whole home humidification

8:08

system and then reinforce the floor

8:11

because you know the floor can't handle

8:14

this. So it was a huge product

8:15

transition and then just the rack was so

8:18

dense it was really hard for them to get

8:20

get the heat out. So Blackwells have

8:22

only really started to be deployed and

8:26

really scaled deployments over the last

8:29

3 or 4 months. Had reasoning not come

8:32

along, there would have been no AI

8:35

progress

8:37

from mid 2024

8:40

through essentially Gemini 3. there

8:43

would have been none. Everything would

8:45

have stalled and can you imagine what

8:47

that would have meant to the markets

8:48

like for sure we would have lived in a

8:51

very different environment. So reasoning

8:53

kind of bridged this like 18month gap.

8:56

Reasoning kind of saved AI because it

8:59

let AI make progress without Blackwell

9:03

or the next generation of TPU which were

9:05

necessary for the scaling laws for

9:07

pre-training to continue. The reason

9:09

we've had all this progress, maybe we

9:11

could show like the ARC AGI slide where

9:15

you had, you know, you went from 0 to 0

9:17

to 8 over four years, 0 to 8%

9:19

intelligence

9:20

>> and then you went from 8% to 95% in 3

9:23

months when the first reasoning model

9:25

came out from OpenAI is, you know, we

9:27

have these two new scaling laws of post-

9:28

training, which is just reinforcement

9:30

learning with verified rewards. Verified

9:32

is such an important concept in AI. Um,

9:35

like one of Karpathy's great things was

9:36

with software, anything you can specify,

9:39

you can automate. With AI, anything you

9:42

can verify, you can automate. It's such

9:44

an important concept and I think an

9:46

important distinction. And then test

9:47

time compute. And so all the progress

9:49

we've had, and we've had immense

9:51

progress um, since October 24th through

9:54

today was based entirely on these two

9:56

new scaling laws. And Gemini 3 was

9:59

arguably the first test since Hopper

10:02

came out of the scaling law for

10:05

pre-training and it held. And that's

10:07

great because all these scaling laws are

10:10

multiplicative. So now we're going to

10:11

apply these two new um reinforcement

10:15

learning with verified rewards and test

10:17

time compute um to much better base

10:20

models. Google came out with the TPU v6

10:23

in 2024 and the TPU v7 in 2025.

10:27

And in semiconductor time, it's like

10:31

almost like imagine like Hopper is like,

10:33

you know, it's like a World War II era

10:36

airplane. And it was by far the best

10:38

World War II era airplane. It's P-51

10:40

Mustang with the Merlin engine. And two

10:43

years later in semiconductor time,

10:44

that's like,

10:46

>> you know, you're an F4 Phantom. Okay.

10:49

Because Blackwell was such a complicated

10:51

product and so hard to ramp, Google was

10:54

training Gemini 3 on 24 and 25 era TPUs,

10:57

which are like F4 Phantoms. Like

10:59

Blackwell, it's like an F-35.

11:02

>> It just took a really long time to get

11:05

it going.

11:06

>> So, I think, you know, Google for sure

11:08

has this temporary advantage right now.

11:10

Um, from a pre-training perspective, I

11:13

think it's also important that they've

11:15

been the lowest cost producer of tokens.

11:18

Okay. And this is really important

11:20

because AI is the first time in my

11:23

career as a tech investor that being the

11:25

lowcost producer has ever mattered.

11:26

Apple is not worth trillions because

11:28

they're the lowcost producer of phones.

11:30

Microsoft is not worth trillions because

11:32

they're the low lowcost producer of

11:33

software. Nvidia is not worth trillions

11:35

cuz they're the lowcost producer of AI

11:37

accelerators. It's never mattered. And

11:39

this is really important because what

11:40

Google has been doing has the lowcost

11:43

producer is they have been I would say

11:46

sucking the economic oxygen out of the

11:48

AI ecosystem which is an extremely

11:50

rational strategy for them and for

11:53

anyone who's a lowcost producer you know

11:55

let's just let's make life really hard

11:59

for our competitors. Um and so what

12:03

happens now I think this has pretty

12:05

profound implications. One, we will see

12:07

the first models trained on Blackwell in

12:09

early 2026.

12:11

>> Y

12:11

>> I think the first Blackwell model will

12:13

come from XAI. And the reason for that

12:15

is just it's a according to Jensen, no

12:18

one builds data centers faster than

12:20

Elon. Yes, Jensen has said this on the

12:22

record. Even once you have the

12:24

Blackwells, it it takes 6 to9 months to

12:27

get them performing at the level of

12:29

Hopper

12:30

>> cuz the Hopper is finally tuned.

12:32

Everybody knows how to use it. The

12:33

software is perfect for it. engineers

12:35

know all its quirks. You know, everybody

12:37

knows how to architect a Hopper data

12:39

center at this point. And by the way,

12:40

when Hopper came out, it took 6 to 12

12:43

months for it to really outperform AER,

12:45

which was generation before. So, if

12:47

you're Jensen or Nvidia, you need to get

12:50

as many GPUs deployed in one data center

12:53

as fast as possible in a coherent

12:55

cluster so you can work out the bugs.

12:57

And so this is what XAI effectively does

12:59

for Nvidia because they build the data

13:01

centers the fastest. They can deploy,

13:03

you know, black wells that scale the

13:05

fastest and they can help work with

13:07

Nvidia to work out the bugs for everyone

13:09

else. So because they're the fastest,

13:12

they will they'll have the first

13:13

Blackwell model. We know that scaling

13:16

laws for pre-training are intact and

13:18

this means the Blackwell models are

13:20

going to be amazing. Blackwell is um I

13:23

mean it's not an F35 versus an F4

13:25

Phantom, but from my perspective it is a

13:27

better chip, you know, maybe it's like

13:29

an F-35 versus a Raphael. And so now

13:32

that we know pre-scaling holding, we

13:34

know that these Blackwell models are

13:35

going to be really good.

13:36

>> And you know, kind of based on the raw

13:38

specs, they should probably be better.

13:40

>> Then something even more important

13:41

happens.

13:42

>> So the GB200 was really really it was

13:46

really hard to get a coin. Um,

13:49

the GB300

13:51

is a great chip. It is drop in

13:55

compatible in every way with those GB200

13:57

racks. Now, you're not going to replace

13:58

the GB200s. No new power walls. Yeah.

14:01

>> Yeah. Just any data center that can

14:02

handle those. You can slot in the

14:04

GB300s. And now everybody's good at

14:06

making those racks and you know how to

14:07

get the heat out. You know how to cool

14:08

them.

14:09

>> You're going to put those GB300s in and

14:11

then the companies that use the GB300's,

14:14

they are going to be the lowcost

14:16

producer of tokens.

14:18

particularly if you're vertically

14:19

integrated. If you're paying a margin to

14:20

someone else to make those tokens,

14:22

you're probably not going to be. I think

14:23

this has pretty profound implications

14:26

because it ch I think it has to change

14:28

Google's strategic calculus. If you have

14:30

a decisive cost advantage and you're

14:33

Google and you have search and all these

14:35

other businesses, why not run AI at a

14:39

negative 30% margin?

14:41

>> It is by far the rational decision. You

14:44

take the economic oxygen out of the

14:46

environment. You eventually make it hard

14:48

for your competitors who need funding

14:51

unlike you to raise the capital they

14:53

need. And then on the other side of

14:55

that, maybe have an extremely dominant

14:57

share position. Well, all that calculus

14:59

changes once Google is no longer

15:02

>> the lowcost producer, which I think will

15:04

be the case. The black wells are now

15:06

being used for training. And then when

15:08

the that model is trained then you shift

15:11

you start shifting blackwell clusters

15:12

over to inference and then all these

15:14

cost calculations and these dynamics

15:16

change

15:17

>> and I do think it's this it's very

15:19

interesting like during the strategic

15:21

and economic calculations between the

15:23

players. I've never seen anything like

15:25

it. You know everyone understands their

15:29

position on the board, what the prize

15:31

is, you know what play their opponents

15:34

are running. Um, and it's really

15:35

interesting to watch. So, I just think

15:39

if Google changes its behavior, cuz it's

15:42

going to be really painful for them as a

15:44

higher cost producer to run that

15:45

negative 30% margin, it might start to

15:47

impact, you know, their stock. That has

15:49

pretty profound implications for the

15:51

economics of AI. And then when Reuben

15:53

comes out, we'll know the gap the gap is

15:57

going to expand significantly

15:58

>> versus TPUs.

15:59

>> Versus TPUs and and all other AS6. Now,

16:02

I think tranium 3 is probably going to

16:03

be pretty good. train for are going to

16:04

be good.

16:04

>> Why is that the case? Why why won't TPU

16:06

v8 V9 be every bit as good?

16:09

>> A couple of things. So one um for

16:12

whatever reason um Google made more

16:14

conservative design decisions.

16:17

I think part of that is so Google round

16:20

numbers Google like let's say the TPU

16:23

Google is so there's front end and back

16:25

end of semiconductor design and then

16:28

there's you know uh dealing with Taiwan

16:30

semi. You can make an ASIC in a lot of

16:32

ways. What Google does is they do mostly

16:35

the front end for the TPU and then

16:37

Broadcom does the back end and manages

16:39

Taiwan mean everything. It's a crude

16:42

analogy but like the front end is like

16:44

the architect of a house.

16:45

>> Yep.

16:46

>> They design the house. The back end is

16:48

the person who builds the house and the

16:50

managing Taiwan Simmyi is like stamping

16:52

out that house like LAR or you know Dr.

16:54

Horton and for doing those two ladder

16:57

parts broadcoms a 50 to 55% gross

17:00

margin. We don't know what on TPUs.

17:02

Okay, let's say in 2027

17:05

TPU I think it sits estimates maybe

17:06

somewhere around 30 billion again who

17:08

knows I mean

17:10

>> yeah yeah yeah but I 30 billion I think

17:12

is a reasonable estimate 50 to 55% gross

17:16

margins so Google is paying Broadcom $15

17:19

billion

17:20

>> okay that's a lot of money

17:22

>> and at a certain point it makes sense to

17:25

bring a semiconductor program entirely

17:28

in house so in other words Apple does

17:30

not have an ASIC partner for their chips

17:33

>> they do they do the front end themselves

17:34

the back end and they manage Taiwan semi

17:36

and the reason is they don't want to pay

17:38

that 50% margin so at a certain point it

17:40

becomes rational to renegotiate this and

17:43

just as perspective the entire opex of

17:46

Broadcom's semiconductor division is

17:48

round numbers $5 billion so it would be

17:50

economically rational now that Google's

17:53

paying if it's 30 billion we're paying

17:54

them 15 Google can go to every person

17:57

who works in Broadcom Smi double their

17:59

comp

18:00

>> and make an extra extra 5 billion. You

18:03

know, in 2028, let's just say it does 50

18:05

billion. Now it's 25 billion. You could

18:08

triple their comp. And by the way, you

18:09

don't need them all.

18:10

>> Yeah.

18:10

>> And and of course, they're not going to

18:11

do that because of competitive concerns.

18:14

>> But with TPUv8,

18:16

all of this and V9, all of this is

18:18

beginning to have an impact because

18:20

Google is bringing in MediaTek. This is

18:22

maybe the first way you send a warning

18:25

shot to Broadcom. were really not happy

18:27

about

18:28

>> all this money we're paying

18:29

>> but they did bring MediaTek in and the

18:30

Taiwanese ASIC companies have much lower

18:32

gross margins so this is kind of the

18:34

first shot against the bow and then

18:36

there's all this stuff people say oh but

18:39

has the best certiscom

18:41

has really good certis and certis is

18:43

like an extremely foundational

18:44

technology because it's how the chips

18:46

communicate with each other you have to

18:48

serialize and do serialize but there are

18:50

other good certis providers in the world

18:52

a really good certis is not at a certain

18:56

Maybe it's worth 10 or 15 billion a

18:57

year, but it's probably worth about

18:58

worth 25 billion a year. So because of

19:02

that friction, um, and I think

19:04

conservative design choices on the part

19:06

of Google and maybe the reason they made

19:08

those conservative design choices is

19:10

because they were going to a bifurcated

19:12

supply. You know, TPU is slowing down. I

19:16

would say has kind of the GPUs are

19:19

accelerating. This is the first, you

19:21

know, the com the competitive response

19:23

of Lisa and Jensen to everybody saying

19:26

we're gonna have our own ASIC is, hey,

19:27

we're just going to accelerate. We're

19:29

going to do do a GPU every year and you

19:31

cannot keep up with us. And then I think

19:33

what everybody is learning is like, oh

19:35

wow, that's so cool. You made your own

19:38

accelerator has an ASIC. Wow, what's the

19:40

nick going to be? What's the CPU going

19:42

to be? You know, what's the scaleup

19:44

switch going to be? What's the scaleup

19:46

protocol? What's the scale out switch?

19:49

what kind of optics are you going to

19:50

use? What's the software that's going to

19:52

make all this work together? And then

19:54

it's like, oh I made this tiny

19:57

little chip and you know, like whether

20:00

it's admitted or not, like you know, I'm

20:03

sure the GPUs don't GPU makers don't

20:05

love it when their customers make AS6

20:07

and try and compete with them

20:08

>> and like whoops

20:10

what what did I do? I thought this was

20:12

easy. How do you know? And it also it

20:14

takes at least three generations to make

20:16

a good chip like the TP TPU V1. I mean

20:19

it was an achievement and that they made

20:21

it.

20:21

>> Yeah.

20:22

>> Um it was really not till TPU V3 or V4

20:25

that the TPU started to become like even

20:28

vaguely competitive.

20:29

>> Is that just a classic like learning by

20:31

doing thing

20:32

>> 100%.

20:32

>> Yeah. And even if you've made like the

20:36

first from my perspective, the best ASIC

20:39

team at any semiconductor company is

20:41

actually the Amazon ASIC team.

20:43

>> You know, they were the first one to

20:44

make the gravitron CPU. They have this

20:46

nitro. Um it was the first, it's called

20:49

Supernick. They've been extremely

20:50

innovative, really clever. And like

20:53

Tranium and Infantry One, you know, they

20:57

maybe they're a little better than the

20:58

TPUV1, but only a little. Trannium 2,

21:01

you get a little better. Trium 3 it's I

21:04

think the first time it's like okay and

21:05

then you know I think tradeium 4 will

21:07

probably be good. I will be surprised if

21:10

there are a lot of AS6 other than

21:14

tranium and TPU

21:16

>> and by the way and tranium and TPU will

21:18

both run on customerowned tooling at

21:21

some point. We can debate when that will

21:22

happen but the economics of success that

21:26

I just described mean it's inevitable.

21:28

Like no matter what the companies say,

21:31

just the economics make it and reasoning

21:33

from first principles make it absolutely

21:35

inevitable.

21:36

>> If I were to zoom all the way out on

21:37

this stuff, because sometimes I just

21:39

it's I I find these details unbelievably

21:41

interesting and it's like the grandest

21:43

game that's ever been.

21:44

>> That's what I mean. It's crazy.

21:45

>> It's so crazy and so fun to follow.

21:47

Sometimes I forget to zoom out and say,

21:48

"Well, well, so what?" Like, okay, so

21:50

project this forward three generations

21:52

past Reuben or whatever. What what is

21:55

like the global human dividend of all

21:58

this crazy development? Like we keep

22:01

making the loss lower on these, you

22:02

know, pre uh pre pre-training scaling

22:04

models like who cares? Like it's been a

22:07

while since I've asked this thing

22:08

something that I wasn't kind of blown

22:10

away by the answer for me personally.

22:12

What are the next couple of things that

22:14

all this crazy infrastructure war allows

22:17

us to unlock because they're so

22:19

successful? If I were to posit like an

22:21

event path, I think the Blackwell models

22:23

are going to be amazing. The dramatic

22:25

reduction in per token cost enabled by

22:27

the GB300 and probably more the MI450

22:30

than the MI355, you know, will lead to

22:33

these models being allowed to think for

22:35

much longer, which means they're going

22:38

to be able to, you know, do new things.

22:40

Like I was very impressed Gemini 3 made

22:41

me a restaurant reservation.

22:43

>> It's the first time it's done something

22:45

for me. And I mean, other than like go

22:48

research something and teach me stuff,

22:50

>> but you know, if you can make a

22:51

restaurant reservation, you're not that

22:52

far from being able to make a hotel

22:54

reservation and an airplane reservation

22:57

and order me an Uber and

22:59

>> all of a sudden you got an assistant.

23:00

>> Yeah. And you can just imagine,

23:01

everybody talks about that, but you can

23:03

just imagine it's on your phone. I think

23:04

that's that's pretty near-term, but you

23:07

know, it's you know, it's some big

23:08

companies that are very tech forward.

23:11

you know 50% plus of customer support is

23:14

already done by AI and that's a $400

23:16

billion dollar industry and then if you

23:18

know what AI is great about is

23:19

persuasion that's sales and customer

23:21

support

23:22

>> and so of the functions of a company if

23:24

you think about them them they're to

23:26

make stuff sell stuff and then support

23:28

the customers so right now maybe you're

23:31

in late 26 you're going to be pretty

23:33

good at two of them um I do think it's

23:35

going to have a big impact on media like

23:37

I think robotics you know we talked

23:38

about the last time are going to finally

23:40

start to be real. You know, there's an

23:41

explosion in kind of exciting robotic

23:43

robotic startups. I do still think that

23:45

the main battle is going to be between

23:47

uh Tesla's Optimus and the Chinese

23:49

because, you know, it's easy to make

23:50

prototypes. It's hard to massproduce

23:52

them. But then it goes back to that what

23:54

Andre Karpathy said about AI can

23:56

automate anything that can be verified.

23:59

So any function where there's a right or

24:01

wrong answer, a right or wrong outcome,

24:03

you can apply reinforcement learning and

24:06

make the AI really good at that. Yeah.

24:08

>> What are your favorite examples of that

24:10

so far or theoretically?

24:11

>> Does the model balance? They'll be

24:13

really good at making models. Does you

24:15

know do all the books globally

24:16

reconcile? They'll be really good at

24:18

accounting because it you know was you

24:21

know double entry bookkeeping. It has to

24:22

balance. There's a verifiable you got it

24:24

right or wrong supporter sale. Did you

24:26

make the sale or not? That's very clear.

24:28

I mean that that's that's just like

24:30

AlphaGo.

24:31

>> You know, did you win or you lose? Did

24:33

the guy convert or not? Did the customer

24:35

ask for an escalation during customer

24:37

support or not?

24:39

>> It's like it's most important functions

24:42

are important because they can be

24:43

verified,

24:45

>> right? So I think if all of this starts

24:49

to happen and starts to happen in in in

24:52

26

24:54

like there'll be an ROI on Blackwell and

24:57

then all this will continue

24:59

>> and then we'll have Reuben and then

25:01

that'll be another big quantum of spin

25:03

Reuben and the MI450 and the TPU V9 and

25:06

then I do think just the most

25:07

interesting question is what are the

25:10

economic returns to artificial super

25:11

intelligence because all of these

25:13

companies in this great game they've

25:14

been in a prison prisoners dilemma.

25:16

They're terrified that if they slow

25:17

down,

25:18

>> they're just gone forever.

25:19

>> And their competitors don't, it's an

25:21

existential risk. And you know,

25:23

Microsoft blinked for like six weeks

25:26

earlier this year.

25:27

>> Yeah.

25:28

>> And like I I think they would say they

25:30

regret that.

25:31

>> Yeah.

25:31

>> But with Blackwell and for sure with

25:34

Reuben, economics are going to dominate

25:37

the prisoners dilemma from a decision-m

25:39

and spending perspective just because

25:41

the numbers are so big. And this goes to

25:44

kind of the ROI on AI question. And the

25:46

ROI on AI has

25:49

empirically, factually, unambiguously

25:52

been positive.

25:54

>> Like I just always find it strange that

25:56

there's any debate about this because

25:59

the largest biders on GPUs are public

26:01

companies. They report something called

26:03

audited quarterly financials. And you

26:05

can use those things to calculate

26:06

something called a return on invested

26:08

capital. And if you do that calculation,

26:10

the ROIC of the big public spenders on

26:12

GPUs is higher than it was before they

26:14

ramp spinning. And you could say, well,

26:15

part of that is, you know, opex savings.

26:18

Well, at some level that is part of what

26:20

you expect the ROI to be from AI. And

26:23

then you say, well, a lot of is actually

26:25

just applying GPUs, moving the big

26:26

recommener systems that power the

26:28

advertising and the recommendation

26:30

systems from CPUs to GPUs, and you've

26:32

had massive efficiency gains. And that's

26:34

why all the revenue growth at these

26:35

companies has accelerated. But like, so

26:37

what? the ROI has been there. Um, and it

26:39

is interesting like every big internet

26:42

company,

26:42

>> the people who are responsible for the

26:45

revenue

26:46

>> are intensely annoyed at the amount of

26:49

GPUs that are being given to the

26:51

researchers.

26:52

>> It's a very linear equation. If you give

26:54

me more GPUs, I will drive more revenue.

26:56

Give me those GPUs, we'll have more

26:58

revenue, more gross profit, and then we

26:59

can spend money. So, it's this constant

27:01

fight at every company. One of the

27:03

factors in the prisoners dilemma is

27:04

everybody has this like religious belief

27:07

that we're going to get to ASI and at

27:09

the end of the day what do they all

27:10

want? Almost all of them want to live

27:12

forever. Okay. And they think that ASI

27:14

is going to help them that

27:15

>> right good return.

27:16

>> That's a good return. But we don't know.

27:18

And if as humans we have pushed the

27:21

boundaries

27:22

of physics, biology and chemistry, the

27:24

natural laws that govern the universe.

27:26

I'm very curious about your favorite

27:28

sort of throw cold water on this stuff

27:30

type takes that you think about

27:31

sometimes. One would be like the things

27:33

that would cause I'm curious what you

27:35

think the things that would cause this

27:36

demand for compute to change or even the

27:39

trajectory of it to change. There's

27:40

there's one really obvious bare case and

27:44

it is just edge AI and it's connected to

27:46

the economic returns to ASI. in three

27:49

years on a bigger and bulkier phone to

27:52

fit the amount of DRAM necessary, you

27:54

know, and the battery won't probably

27:55

last as long, you will be able to

27:57

probably run a pruned down version of

28:00

something like Gemini 5 or Gro 4, Gro

28:04

4.1 or you know, Chat GPT at um

28:09

I don't know 30 60 tokens per second

28:13

>> and then that's free. And this is

28:14

clearly Apple's strategy. It's just

28:16

we're going to be a distributor of AI

28:18

>> and we're going to make it privacy safe

28:20

and run on the phone and then you can

28:22

call one of the big models, you know,

28:24

the the god models in the cloud whenever

28:26

you whatever you have a question. And if

28:28

that happens, if like 30 to 60 tokens at

28:31

like a one whatever it is a 115 30 60

28:35

tokens a second at a 115 IQ is good

28:38

enough.

28:40

I think that's

28:41

>> a bare case

28:42

>> other than just the scaling laws break,

28:45

you know. But in terms of if we assume

28:47

scaling laws continue and we now know

28:49

they're going to continue for

28:50

pre-training for at least one more

28:52

generation and we're very early in the

28:54

two new scaling laws you know for post-

28:56

training mid-training RLVR whatever

28:58

people want to call it and then test

28:59

time computed inference we're so early

29:01

in those and we're getting so much

29:03

better at helping the models hold more

29:05

and more context in the in their minds

29:08

as they do you know this test time

29:10

compute and that's really powerful

29:12

because you know everybody's like well

29:14

you know the how's the model going to

29:16

know this? Well, eventually if you can

29:17

hold enough context, you can just hold

29:19

every Slack message and Outlook message

29:22

and company manual in in a company in

29:26

your context.

29:26

>> Yeah.

29:28

>> And then you can compute the new task

29:31

>> and compare it with your knowledge of

29:32

the world, what you think, what the

29:34

model thinks, all this context. And you

29:36

know, it may be that like, you know, in

29:38

just really really long context windows

29:40

are the solution to a lot of the current

29:42

limitations. Um, and that's enabled by

29:44

some all these cool tricks like KV cash

29:46

offload and stuff. But I do think like

29:49

other than scaling laws slowing down,

29:51

other than there being low economic

29:53

returns to ASI, edge AI is to me by far

29:58

the most plausible and scariest bare

30:00

case.

30:01

>> I like to visualize like different

30:02

S-curves. invested through the iPhone

30:04

and I love to like see the visual of the

30:06

iPhone models as it as it sort of went

30:08

from this clunky bricky thing up to the

30:11

what we have now where like each one's

30:12

like a little bit, you know, obviously

30:13

we we've sort of petered out on its form

30:15

factor. If you if you picture something

30:17

similar for the Frontier models

30:19

themselves, does it feel like a like

30:21

it's at a certain part of that of that

30:24

natural technology paradigm progression

30:26

to you? If you're paying for Gemini

30:28

Ultra or Super Grock and you're getting,

30:31

you know, the good AI, it's hard to see

30:33

differences. Like, I have to go really

30:36

deep on something like, do you think PCI

30:39

Express or Ethernet is a better protocol

30:42

for scale up networking and why? Show me

30:45

the scientific papers. And if you shift

30:48

between models and you ask a question

30:50

like that where you know it really

30:52

deeply,

30:53

>> know that then

30:54

>> you know the answers. Yeah. then you see

30:56

differences. I do play fantasy football.

30:59

Um, winnings are donated to charity,

31:01

>> but it is like, you know, these new

31:03

models are quite a bit better at helping

31:07

like who should I play?

31:08

>> Yeah.

31:09

>> You know, and and they they think in

31:11

much more sophisticated ways.

31:13

>> Um, and by the way, if you're if you're

31:15

a historically good fantasy football

31:17

player and you're having a bad season,

31:19

it's why this is why because you're not

31:22

using it, you know. And I think we'll

31:24

see that in more and more domains. But I

31:26

do think they are already at a level

31:29

where unless you are a true expert or

31:32

just have an intellect that is beyond

31:34

mind

31:35

>> um it's hard to see um the progress and

31:39

that's why I do think we need to shift

31:42

from getting more intelligent to more

31:44

useful

31:45

>> unless more intelligence starts leading

31:47

to these massive scientific

31:49

breakthroughs and we're curing cancer in

31:51

26 and 27. Yeah,

31:52

>> I don't know that we're going to be

31:53

curing cancer, but I do think from a ROI

31:57

almost an ROIS curve, we need to kind of

32:00

hand off from intelligence to usefulness

32:03

>> and then usefulness will then have to

32:06

hand off to scientific breakthrough, you

32:09

know, just that creates whole new

32:11

industries.

32:11

>> What are the building blocks of

32:13

usefulness in your mind?

32:14

>> Just being able to do things

32:15

consistently and reliably. And a lot of

32:18

that is keeping all the context. Like

32:22

there's a lot of context if someone

32:23

wants to plan a trip for me. Like you

32:25

know I've I've acquired these strange

32:27

preferences. Like I follow that guy

32:29

Andrew Huberman. So I like to have an

32:30

east facing balcony so I can get morning

32:32

sun. You know the AI has to remember,

32:36

you know, being on a plane with Starlink

32:38

is important to me. Okay. Here are the

32:40

resorts I've historically liked. Here

32:41

are the kinds of areas I've liked. Here

32:43

are the rooms that I would really like

32:45

at each. That's a lot of context and to

32:48

keep all of that and kind of weight

32:50

those it's a hard problem. So I think

32:52

context windows are a big part of it.

32:54

You know, there's this meter task

32:55

evaluation thing like

32:58

>> how long it can work,

32:58

>> how long it can work for. And you could

33:01

think of that being related and in some

33:03

way to to context. Um although not

33:06

precisely, but that just task length

33:10

needs to keep expanding because you know

33:13

booking a restaurant and booking is

33:16

economically useful but you know it's

33:18

not that economically useful. But

33:20

booking me an entire vacation and

33:23

knowing the preferences of my parents,

33:26

my sister, my niece, and my nephew, and

33:28

what it means that like that's a much

33:31

harder problem. And that's something

33:32

that like a human might spend three or

33:35

four hours on optimizing that. And then

33:37

if you can do that, that's that's

33:39

amazing. But then again, I just think it

33:42

it has to be good at sales and customer

33:45

support relatively soon. I do think

33:47

we're going to see an kind of an

33:48

acceleration in the awesomeness of

33:50

various products just because engineers

33:53

are using AI to make products better and

33:55

faster.

33:56

>> We both invested in Forell, the hearing

33:58

aid company, which is just absolutely

34:00

remarkable. I think something I never

34:01

would have thought of.

34:02

>> And we're going to see I think something

34:03

like that in every vertical and that's

34:05

AI being used for the most core function

34:08

>> Yeah. of any company which is designing

34:11

the product and then it will be you know

34:13

there's already lots of examples of AI

34:15

being used to help manufacture the

34:17

product and distribute it more

34:19

efficiently you know whether it's

34:20

optimizing a supply chain you know

34:22

whether it's you know having a vision

34:23

system watch a production line you know

34:26

so I I think a lot of stuff is happening

34:29

the other thing I think is really

34:30

interesting in this whole ROI part is

34:32

Fortune 500 companies are always the

34:34

last to adopt a new technology they're

34:35

conservative they have lots of

34:36

regulations lots of lawyers Startups are

34:39

always the first. So let's think about

34:40

the cloud which was the first which was

34:43

the last like really truly kind of

34:45

transformative new technology for

34:47

enterprises. Being able to you know do

34:50

have all of your um compute and the

34:52

cloud and use SAS. So it's always

34:54

upgraded. It's always great etc. etc.

34:56

You can get it on every device. I mean

34:58

it's those were dark days before the

34:59

cloud. You know

35:00

>> the first AWS reinvent I think it was in

35:02

2013. Every startup on planet Earth ran

35:05

on the cloud.

35:05

>> Yeah. The idea that you would buy your

35:07

own server and storage box and router

35:09

was ridiculous. And that probably

35:11

happened like even earlier that that had

35:13

probably already happened before the

35:14

first reinvent. And then like you know

35:16

the first big Fortune 500 companies

35:17

started to standardize on it like maybe

35:19

5 years later. You see that with AI I'm

35:21

sure you've seen this in your startups

35:23

and I think one reason VCs are more

35:25

broadly bullish on AI than public market

35:26

investors is VCs see very real

35:29

productivity gains. There's all these

35:31

charts that for a given level of

35:33

revenue, a company today has

35:36

significantly lower employees than a

35:38

company of two years ago.

35:40

>> And the reason is AI is doing a lot of

35:43

the sales, the support and helping to

35:45

make the product. And I mean there is,

35:47

you know, Iconic has some charts. A6Z,

35:49

by the way, David George is a good

35:51

friend, great guy. You know, he has his

35:53

model busters thing. So there's very

35:55

clear data that this is happening. So

35:57

people who have a lens into the world of

35:59

venture see this. And I do think it was

36:00

very important in the third quarter,

36:01

this is the first quarter where we had

36:03

Fortune 500 companies outside of the

36:06

tech industry give specific quantitative

36:09

examples of AIdriven uplift. So C

36:12

Robinson went up something like 20% on

36:14

earnings. And should I tell people what

36:16

C Robinson does?

36:18

>> Let's just say a truck goes from, you

36:20

know, Chicago to Denver. And then, you

36:23

know, the trucker lives in Chicago. So

36:25

it's going to go back from Denver to

36:26

Chicago. There's an empty load. And CH

36:29

Robinson has all these relationships

36:31

with these truckers and trucking

36:32

companies and they match shippers demand

36:37

with that empty load supply to make the

36:39

trucking more efficient. You know,

36:40

they're a freight forwarder. You know,

36:41

there's there's there's actually lots of

36:43

companies like this, but kind of they're

36:44

the biggest and most dominant. So, one

36:46

of the most important things they do is

36:49

they quote price and availability. So,

36:51

somebody a customer calls them up and

36:52

says, "Hey, I urgently need three

36:55

18-wheelers from Chicago to Denver." You

36:57

know, in the past they said it would

36:59

take them, you know, 15 to 45 minutes

37:02

and they only quoted 60%

37:06

of inbound requests. With AI, they're

37:09

quoting 100% and doing it in seconds.

37:12

>> And so they printed a great quarter and

37:14

the stock went up 20% and it was because

37:16

of AIdriven productivity that's

37:18

impacting the revenue line, the cost

37:20

line, everything. And so I actually

37:22

think that's pretty important because I

37:23

was I was actually very worried about

37:25

like the idea that we might have this

37:27

Blackwell ROI air gap because we're

37:29

spending so much money on Blackwell.

37:31

Those Blackwells are being used for

37:32

training and there's no ROI on training.

37:35

Training is you're making the model. The

37:36

ROI comes from inference. So I was

37:38

really worried that you know we're going

37:39

to have

37:40

>> you know maybe this threequarter period

37:42

where the capex is unimaginably high.

37:45

>> Those black wheels are only being used

37:47

for training bars staying flat eyes

37:49

going up.

37:50

>> Yeah. Yeah. Exactly. And so ROIC goes

37:52

down and you can see like Meta Meta they

37:54

printed you know because Meta has not

37:56

been able to make a frontier model. Meta

37:58

printed you know a quarter where ROIC

38:00

declined and that was not good for the

38:02

stock. So I was wor really worried about

38:03

this. I do think that those data points

38:05

are important in terms of suggesting

38:07

that maybe we'll be able to navigate

38:09

this potential air gap and ROIC.

38:12

>> Yeah it makes me wonder about in this

38:14

market I'm like everybody else. It's the

38:16

10 companies at the top that are all the

38:17

market cap more than all of the

38:19

attention. There's 490 other companies

38:22

500. You studied those too. Like what

38:24

what do you think about that group? Like

38:26

what what is interesting to you about

38:27

the group that now nobody seems to talk

38:29

about and no one really seems to care

38:31

about because they don't they haven't

38:32

driven returns and they're a smaller

38:34

percent of the overall.

38:35

>> Well, I think that people are going to

38:37

start to care if you have more and more

38:38

companies print these CH Robinson like

38:40

quarters that companies that have

38:42

historically been really wellrun.

38:46

The reason like they have a long track

38:48

record of success is they have a long

38:51

you cannot succeed without using

38:53

technology well and so if you have a

38:55

kind of internal culture of

38:56

experimentation and innovation I think

38:58

you will do well with AI

39:00

>> you know so like I would bet on the best

39:02

investment banks to be early

39:05

you know earlier and better adopters of

39:07

AI than maybe some of the trailing banks

39:11

you know just sometimes

39:13

past prologue one thing that I strong

39:15

opinion I

39:17

you know, all these VCs are setting up

39:18

these holding companies and, you know,

39:20

we're going to use AI to make

39:21

traditional businesses better and, you

39:24

know, they're really smart VCs and

39:25

they're great track records. But that's

39:27

what private equity has been doing for

39:28

50 years.

39:30

>> You're just not going to beat private

39:32

equity at their game.

39:34

>> What Vista did in the early days, right?

39:35

>> Yeah. Private equity's maybe had a

39:37

little bit of a tough run. You know,

39:38

just multiples have gone up. Now,

39:40

private assets are more expensive. The

39:41

cost of financing has gone up. It's

39:43

tough to take a company public because

39:45

the public valuation is 30% lower than

39:47

the private valuation. So PE's had a

39:49

tough run. I actually think these

39:51

private equity firms are going to be

39:52

pretty good at systematically applying

39:56

AI. We haven't spent much time talking

39:58

about meta, anthropic or open AI. And

40:00

I'd love just like your impression on

40:02

everything that's going on in this

40:03

infrastructure side that we talked

40:04

about. These are three really important

40:06

players in this in this grand battle,

40:08

this grand this grand game. How does all

40:10

of this development that we've discussed

40:12

so far impact those players specifically

40:14

do you think? The first thing let me

40:16

just say about frontier models broadly.

40:17

>> Yeah.

40:18

>> You know in in 2023 and 24 I was fond of

40:22

quoting Eric Visria and Eric Fishria's

40:25

statement our friend um brilliant man

40:27

and Eric would always say foundation

40:29

models are the fastest appreciating

40:30

assets in history.

40:32

>> And I would say he was 90% right. I

40:34

modified the statement. I said

40:36

foundation models without unique data

40:38

and internet scale distribution are the

40:40

fastest appreciating assets in history.

40:42

But reasoning fundamentally changed that

40:44

in a really profound way. There was a

40:46

loop, a flywheel to quote Jeff Bezos

40:48

that it was at the heart of every great

40:51

internet company and it was you made a

40:53

good product, you got users, those users

40:56

using the product generated data that

40:57

could be fed back into the product to

40:59

make it better. And that flywheel has

41:01

been spinning at Netflix, at Amazon, at

41:03

Meta, at Google, you know, for over a

41:06

decade. And that's an incredibly

41:08

powerful flywheel. And it's why those

41:11

internet businesses were so tough to

41:12

compete with. It's why they were

41:14

increasing returns to scale. You

41:16

everybody talks about network effects

41:17

much more and you know network effects

41:19

are they were important for social

41:21

networks. I I don't know to what extent

41:23

meta is a social network anymore. It's

41:24

more like a content distribution

41:26

>> but they just had increasing returns to

41:28

scale because of that

41:30

>> flywheel. And that dynamic was not

41:33

present in the pre-reasoning world of

41:35

AI. You pre-trained a model, you let it

41:38

out in the world, and it was what it

41:40

was. And it was actually pretty hard.

41:43

They would do RLHF, reinforcement

41:45

learning with human feedback, and you

41:46

try and make the bot model better, and

41:48

maybe you'd get a sense from Twitter

41:49

vibes that people didn't like this, and

41:51

so you tweak it, and you know, there

41:53

were the little up and down arrows, but

41:55

it was actually pretty hard to feed that

41:57

back into the model. with reasoning.

42:00

It's early but that flywheel has started

42:02

to spin and that is really profound for

42:07

these frontier labs. So one reasoning

42:09

fundamentally changed the industry

42:11

dynamics of Frontier Labs. just explain

42:13

why specifically that is like what what

42:14

what is going on

42:15

>> because if a lot of people are asking a

42:18

similar question and

42:22

they're consistently either liking or

42:24

not liking the answer, then you can kind

42:28

of use that like that as a verifiable

42:31

reward. That's a good outcome. And then

42:33

you can kind of use feed those good

42:36

answers back into the model. and we're

42:39

very early at this flywheel spinning

42:42

>> like it's hard to do now,

42:44

>> but you can see it beginning to spin.

42:47

>> So, this is important fact number one

42:49

for all of those dynamics. Second,

42:52

>> I think it's really important that Meta,

42:54

you know, Mark Zuckerberg at the

42:55

beginning of this year in January said,

42:58

you know, I anticipate, you know, I'm

42:59

highly confident, I'm going to get the

43:00

quote wrong, that at some point in 2025,

43:03

we're going to have the best and most

43:04

performant AI. I don't know if he's in

43:06

the top hundred. Okay,

43:08

>> so he was as wrong as it was possible to

43:11

be. And I think that is a really

43:13

important fact because it suggests that

43:16

what these four companies have done is

43:18

really hard to do because Meta threw a

43:21

lot of money at it and they failed.

43:23

Yamakun had to leave. They had to have

43:26

the famous billion dollar for AI

43:28

researchers. And by the way, Microsoft

43:30

also failed. They did not make such an

43:33

unequivocal prediction but they hire but

43:36

they bought um inflection AI and you

43:39

know there were a lot of comments from

43:40

them that we anticipate our internal

43:41

models quickly getting better and we're

43:43

going to run more and more of cop you

43:45

know our AI on our internal models

43:48

>> Amazon they bought a company called

43:49

Adept AI

43:51

>> they have their models called Nova

43:53

>> no I don't think they're in the top 20

43:55

>> so clearly it's much harder to do than

43:58

people thought a year ago and there's

44:00

many many reasons for that like it's

44:01

actually really hard to keep a big

44:03

cluster of GPUs coherent. A lot of these

44:06

companies were used to running their

44:08

infrastructure to optimize for cost

44:10

>> right

44:11

>> instead of performance

44:13

>> complexity and performance

44:14

>> complexity and keeping the GPUs

44:17

running at high utilization rate in a

44:20

big cluster. It's actually really hard

44:22

and there are wild variations in how

44:26

well companies run GPUs.

44:27

>> Yeah. And if you're running if the most

44:30

anybody because the laws of physics, you

44:31

know, maybe you can get two or 30

44:32

hundred,000 black wells coherent, we'll

44:34

see. But if you have 30% uptime on that

44:37

cluster and you're competing with

44:38

somebody who has 90% uptime,

44:40

>> you're not even competing. So one,

44:42

there's a huge spectrum in how well

44:44

people run GPUs. Two, then I think there

44:47

is, you know, these AI researchers, they

44:48

like to talk about taste. I find it very

44:51

funny. You know, oh why do you make so

44:52

much money? I have very good taste. You

44:54

know what taste means is you have a good

44:57

intuitive sense for the experiments to

44:59

perform. And this is a this is is why

45:02

you pay people a lot of money because it

45:04

actually turns out that as these models

45:05

get bigger, you can no longer run an

45:07

experiment on a thousand GPU cluster and

45:09

replicate it on 100,000 GPUs. You need

45:12

to run that experiment on 50,000 GPUs

45:15

and maybe it takes, you know, days.

45:17

>> And so there's a very high opportunity

45:19

cost. And so you have to have a really

45:21

good team that can make the right

45:24

decisions about which experiments to run

45:26

on this. And then you need to do, you

45:28

know, all the reinforcement learning

45:30

during post- training well and the test

45:31

time compute. Well, complicated.

45:33

>> It's really hard to do. And everybody

45:35

thinks it's easy, but all those things,

45:37

you know, I used to have this saying

45:38

like, hey, I was a retail analyst long

45:40

ago. Pick any vertical in America. If

45:43

you can just run a thousand stores and

45:46

have them clean, well lit, stocked with

45:50

relevant goods at good prices and

45:54

staffed by friendly employees who are

45:57

not stealing from you, you're going to

45:59

be a $20 billion company, a $30 billion

46:01

company. Like 15 companies have been

46:03

able to do that. It's really hard. And

46:05

it's the same thing. Doing all of these

46:07

things well is really hard.

46:10

and then reasoning with this flywheel.

46:13

This is beginning to create more

46:15

separation.

46:16

>> And what's even more important, every

46:18

one of those labs, XAI, Gemini, OpenAI,

46:22

and Enthropic, they have a more advanced

46:25

checkpoint

46:27

internally of the model. Checkpoint is

46:29

just um you're kind of continuously

46:31

working on these models and then you

46:32

release kind of a checkpoint and then

46:34

the reason these models get fast

46:36

>> the one they're using internally is for

46:37

>> better and they're using that model to

46:39

train the next model

46:41

>> and if you do not have

46:43

>> that latest checkpoint it's

46:45

>> you're behind

46:46

>> you're it's getting really hard to catch

46:47

up. Chinese open source is a gift from

46:51

God to meta

46:52

>> because you can use Chinese open source

46:55

>> to try and that can be your checkpoint

46:58

and you can use that

46:59

>> as a way to kind of bootstrap this and

47:02

that's what I'm sure they're trying to

47:03

do and everybody else. Um the big

47:05

problem and the big a giant swing factor

47:08

I think China's made a terrible mistake

47:09

with this rarest thing you know I think

47:11

China because you know they have the

47:13

Huawei Asin and it's a decent chip and

47:16

verse something you know like you know

47:18

the the deprecated hop preserving

47:19

something it looks okay so they're

47:21

trying to force Chinese open source to

47:22

use their Chinese chips uh their

47:25

domestically designed chips. The problem

47:26

is Blackwell is going to come out now

47:29

and the gap between these American

47:31

frontier labs and Chinese open source is

47:33

going to blow out because of Blackwell

47:35

and actually DeepSeek in their most

47:37

recent technical paper v3.2 said like

47:40

one of the reasons we struggle to

47:41

compete with the American Frontier Labs

47:42

is we don't have enough compute. That

47:45

was their very politically correct,

47:47

still a little bit risky way of saying,

47:49

you know, cuz China said, "We don't want

47:51

the black wells, right?" And they're

47:52

saying, "Guys, that might be a big

47:54

mistake. That might be a big mistake."

47:57

And so, if you just kind of play this

47:58

out, these four American labs are going

48:01

to start to widen their gap versus

48:03

Chinese open source, which then makes it

48:05

harder for anyone else to catch up

48:07

because that gap is growing. So, you

48:09

can't use Chinese open source to

48:10

bootstrap. And then geopolitically,

48:13

China thought they had the leverage.

48:15

They're going to realize, oh, whoopsy

48:16

daisy. We do need the black wells. And

48:18

unfortunately, they'll probably for them

48:21

um they'll probably realize that in late

48:22

26. And at that point, there's an

48:25

enormous effort underway. DARPA has

48:27

there's all sorts of really cool DARPA

48:28

and DoD programs to incentivize really

48:31

clever technological solutions for rare

48:33

earths, you know, like using enzymes to

48:36

refine them or there's all sorts of

48:38

really cool things happening, you know,

48:40

and then, you know, there's a lot of

48:41

rare earth deposits in countries that

48:43

are very friendly to America that, you

48:45

know, don't mind actually refining it in

48:47

the, you know, traditional way. So, I

48:49

think rare earths are going to be solved

48:51

way faster than anyone thinks. You know,

48:53

they're obviously not that rare. They're

48:54

just misnamed. they're rare because, you

48:56

know, they're really messy to refine.

48:58

And so geopolitically, I actually think

49:00

Blackwell is pretty significant. Um, and

49:03

it's going to give America a lot of

49:04

leverage as this gap widens. And then in

49:08

the context of all of that, going back

49:09

to the dynamics between these companies,

49:11

XAI will be out with the first Blackwell

49:13

model and then they'll be the first ones

49:15

probably using Blackwell for inference

49:16

at scale. And I think that's an

49:17

important moment for them. And by the

49:19

way, it is funny like um you know if you

49:21

go on open router you can just look they

49:23

have dominant share now open router is

49:25

whatever it is it's 1% of of API tokens

49:28

but it's an indication

49:30

>> they process 1.35 trillion tokens Google

49:33

did like eight or 900 billion this is

49:34

like whatever it is last 7 days or last

49:36

month you know anthropic was at 700

49:39

billion like XAI is doing really really

49:41

well and the model is fantastic I highly

49:43

recommend it but you'll see XAI you know

49:46

come out with this open AAI will come

49:48

want faster. OpenAI's

49:50

issue that they're trying to solve with

49:52

Stargate is because they pay a margin to

49:54

people for compute

49:55

>> and maybe the people who run their

49:57

compute are not the best at running

49:58

GPUs. They are a high-cost producer of

50:01

tokens. Um, and I think this kind of

50:03

explains a lot of their

50:05

>> code red recently.

50:06

>> Yeah. Well, just the 1.4 $4 trillion in

50:08

spending commitments. And I think that

50:11

was just like, hey, they know they're

50:12

going to need to raise a lot of money.

50:14

Um, particularly if Google keeps its

50:17

current strategy of sucking the economic

50:19

oxygen out of the room and, you know,

50:21

you go from 1.4 trillion rough vibes

50:23

code red like pretty fast, you know, and

50:25

the reason they have a code red is

50:27

because of all these dynamics. So then

50:29

they'll come out with a model but they

50:31

will not have fixed their per token cost

50:33

disadvantage yet relative to both XAI

50:36

and Google and almost and anthropic at

50:38

that point. Anthropic is a good company.

50:40

You know they're burning dramatically

50:41

less cash than openai and growing

50:44

faster. So I think you have to give

50:45

anthropic a lot of credit and and a lot

50:47

of that is their relationship with

50:49

Google and Amazon for the TPUs and the

50:51

trainiums. And so Anthropic has been

50:52

able to benefit from the same dynamics

50:54

that Google has. I think is very

50:56

indicative in this great game of chess.

50:58

You know, you can look at Daario Jensen

51:00

maybe have taken a few there have been a

51:02

few public comments, you know, that

51:04

were, you know, made between them.

51:06

>> Jousting,

51:07

>> a little bit of jousting. Well,

51:08

Anthropic just signed the $5 billion

51:10

deal with Nvidia.

51:11

>> That is because Daario is a smart man

51:14

and he understands these dynamics about

51:16

Blackwell and Rubid relative to TPU. And

51:19

so Nvidia now goes from having two of

51:21

the fighters,

51:23

two fighters, XAI and OpenAI to three

51:26

fighters. So that that helps in this

51:29

Nvidia vers Google battle. And then if

51:33

Meta can catch up, that's really

51:36

important. And so I'm I am sure Nvidia

51:39

is doing whatever they can to help Meta,

51:41

you know, whatever. Like let us you're

51:44

running those GPUs this way. May maybe

51:46

we should maybe we should twist the

51:48

screw this way or turn the dial that way

51:50

and then it will be also if Blackwell

51:53

comes back to China which it seems like

51:55

it probably will happen that will also

51:57

be very good because then Chinese open

51:59

source will be back. What other I'm I'm

52:00

always so curious about the polls of

52:02

things like one poll would be the other

52:04

breakthroughs that you have your your

52:05

mind on things in the data center that

52:07

aren't chips that we've talked about

52:08

before as as one example. I think the

52:10

most important thing that's going to

52:11

happen in the world in this world in the

52:14

next 3 to four years is data centers in

52:17

space

52:18

>> and this has really profound

52:19

implications for everyone building a

52:22

power plant or a data center on planet

52:25

earth. Okay. And there is a giant gold

52:28

rush into this.

52:29

>> I haven't heard anything about this so

52:30

please.

52:30

>> Yeah. You know it's like everybody

52:31

thinks like hey AI is risky you know uh

52:34

but you know what I'm going to build a

52:36

data center. I'm going to build a power

52:37

plant that's going to do a data center.

52:38

We will need that. But if you think

52:40

about it from first principles, data

52:41

centers should be in space. Okay.

52:45

What are the fundamental inputs to

52:47

running a data center? There are power

52:49

and there are cooling

52:50

>> and then there are the chips.

52:52

>> That's like the total if you think about

52:53

it from a total cost perspective.

52:55

>> Yeah. And just the the inputs to making

52:57

the tokens come out of the magic

52:59

machines.

52:59

>> Yeah.

53:00

So in space you can keep a satellite in

53:04

the sun 24 hours a day

53:06

>> and the sun is 30% more intense. You

53:09

know you can keep it in the sun just

53:10

because like if the sun's here's this

53:13

you know you can have the satellite you

53:15

know always kind of catching

53:16

>> catching the light

53:17

>> catching the light. The sun is 30% more

53:19

intense and this results in six times

53:21

more irradiance in outer space than the

53:24

high than on planet earth. So you're

53:26

getting a lot of solar energy. Point

53:27

number one. Point number two, because

53:30

you're in the sun 24 hours a day, you

53:31

don't need a battery. And this is a

53:33

giant percentage of the cost. So the

53:36

lowest cost energy um available in our

53:40

solar system is solar energy and space.

53:43

Okay. Second, for cooling in one of

53:46

these racks, a majority of the mass and

53:49

the weight is cooling.

53:51

>> And the cooling in these data centers is

53:55

incredibly complicated. You know, I

53:56

mean, the HVAC, the CDUs, the liquid

53:59

cooling.

54:01

In space, cooling is free. You just put

54:03

a radiator on the dark side of the

54:05

satellite.

54:06

>> It's gold.

54:07

>> And it's as close to absolute zero as

54:09

you can get.

54:11

>> So, all that goes away and that is a

54:13

vast amount of cost. Okay, let's think

54:16

about um how this these, you know, maybe

54:19

each satellite is kind of a rack. It's

54:21

one way to think of it. Maybe some

54:22

people make bigger satellites that are

54:24

three racks. Well, how are you going to

54:26

collect connect those racks? Well, it's

54:28

funny. In the data center, the racks are

54:31

over a certain distance um connected

54:33

with fiber optics. And that just means a

54:35

laser going through a cable. The only

54:37

thing faster than a laser going through

54:39

a fiber optic cable is a laser going

54:41

through absolute vacuum. So, if you can

54:44

link these satellites in space together

54:47

using lasers, you actually have a faster

54:50

and more coherent network than in a data

54:53

center on Earth. Okay, for training

54:56

that's going to take a long time

54:58

>> just because it's so big.

54:59

>> Yeah, just because it's so big. But for

55:01

inference, but I think even training

55:03

will eventually happen. But then for

55:04

inference, let's think about the user

55:06

experience when I when we asked when you

55:08

know when I asked Gro about you and it

55:11

gave the nice answer. A radio wave

55:13

traveled from my cell phone to a cell

55:15

tower. Then it hit the base station,

55:17

went into a fiber optic cable, went to

55:20

some sort of metro aggregation facility

55:22

in New York, probably within like, you

55:24

know, 10 blocks of here. There's a small

55:26

little metro router that's routed those

55:30

packets to a big XAI data center

55:34

somewhere. Okay? And then the

55:36

computation was done and it came back

55:38

over the same path.

55:41

If the satellites can communicate

55:43

directly with the phone and Starlink has

55:46

demonstrated directto cell capability,

55:48

you just go boom boom. It's a much

55:51

better lowerc cost user experience. So

55:54

in every way data centers in space from

55:59

a first principles perspective are

56:01

superior to data centers on earth.

56:03

>> So if we could teleport that into

56:05

existence, I understand that that

56:06

portion. What are the frictions to that?

56:09

H like why will that not happen? And is

56:10

it launch cost? Is it launch

56:11

availability?

56:13

>> I mean, we need a lot of the space

56:14

starships. Like the Starships are the

56:16

only ones that can eomically make that

56:18

happen.

56:19

>> We need a lot of those Starships. Um,

56:21

you know, maybe China or Russia will be

56:23

able to land a rocket. Blue Origin just

56:25

landed a booster. It's an entirely new

56:28

and different way to think about SpaceX.

56:30

And it is interesting that you know Elon

56:33

posted yesterday or said in an interview

56:35

>> that Tesla, SpaceX and XAI kind of

56:39

converging

56:39

>> were converging and they really are. So

56:41

XAI will be the intelligence module for

56:44

Optimus made by Tesla with Tesla vision

56:47

has its you know perception system and

56:50

then you know SpaceX will have the data

56:52

centers in space that will will you know

56:55

power a lot of the AI presumably for XAI

56:58

and Tesla and the Octopuses and a lot of

57:00

other companies and it's just it is just

57:03

interesting the way that they're

57:04

converging and each one is kind of

57:07

creating competitive advantage for the

57:09

other you know so it's if If you're XAI,

57:12

it's really nice that you have this

57:13

built-in relationship with Optimus and

57:16

now, you know, Tesla's a public company,

57:17

so there's going to be like I cannot

57:19

imagine the level of vetting that will

57:21

go into that intercomp agreement, you

57:23

know, and then you have a big advantage

57:25

with these data centers in space. Um,

57:28

and then it's also nice if you're XAI

57:31

that you have two companies with a lot

57:33

of customers who you can use to help

57:36

build your customer support agents, your

57:39

customer sales agents with kind of

57:42

built-in customers. So, they really are

57:44

all kind of converging um in a neat way.

57:47

And I do think like it's going to be a

57:49

big moment when that first Blackwell

57:51

model comes out from XAI next year. Hm.

57:53

If I go to the other end of the spectrum

57:55

and I think about something that seems

57:57

to have been historically endemic to the

57:59

human economic experience that uh

58:01

shortages are always followed by gluts

58:03

in capital cycles. What if in this case

58:07

um the shortage is compute like Mark

58:09

Chen now is on the record as saying they

58:11

would consume 10x as much compute if you

58:13

gave it to them in like a couple weeks.

58:15

So so like there seems to still be a

58:16

massive shortage of compute which is all

58:18

the stuff we've talked about today. But

58:20

there also just seems to be this like

58:21

iron law of history that gluts follow

58:23

shortages. What do you think about like

58:25

that concept as it relates to this

58:27

>> technology be a glut?

58:28

>> Yeah.

58:29

>> You know, and AI is fundamentally

58:31

different than the software just in that

58:33

every time you use AI takes compute in a

58:37

way that traditional software just did

58:39

not. I mean it is true like I think

58:41

every one of these companies could

58:42

consume 10x more compute. Like what

58:44

would happen is just the $200 tier would

58:47

get a lot better. the free tier would

58:49

get like the $200 tier. Google has

58:51

started to monetize AI mode with ads

58:53

>> and I think that will give everyone else

58:54

permission to introduce ads into the

58:57

free mode and then that is going to be

58:58

an important source of ROI you know like

59:01

>> seems like OpenAI is tailor made to

59:02

>> Yeah. Absolutely. All of them and

59:04

actions like you know hey

59:06

>> you know here are your three vacations

59:07

would you like me to book one and then

59:09

they're for sure going to collect a

59:10

commission. Yeah.

59:11

>> You know here's you know there there

59:13

there's many ways you can make money. I

59:14

think we went into great detail on

59:17

maybe a prior podcast about how just

59:19

inventory dynamics made these inventory

59:21

cycles inevitable in semis. Um, and the

59:24

iron law of semis is just that customer

59:26

buffer buffer inventories have to equal

59:28

lead times. And that's why you got these

59:30

inventory cycles historically. We

59:32

haven't seen a true capacity cycle in

59:35

semis maybe arguably since the late 90s.

59:38

And that's because Taiwan Smi has been

59:40

so good at aggregating

59:42

and smoothing supply.

59:44

And a big problem in the world right now

59:47

is that Taiwan semi is not expanding

59:49

capacity as fast as their customers

59:51

want. And I think this is actually a

59:52

pretty big this they're they're in the

59:54

process of making a mistake just because

59:56

you know you do have Intel and with

59:58

these fabs and they're not as good and

60:00

it's really hard to work with their PDK

60:02

but now you have this guy Leapoo who's

60:04

who's a really good executive um and

60:07

really understands that business. I mean

60:09

by the way Patrick Elsinger I think was

60:10

was also a good executive and he put

60:12

Intel on the only strategy that could

60:14

result in su success and I actually

60:17

think it's shameful that the Intel board

60:18

fired him when they did it. But Leapoo

60:20

is a good executive and now he's reaping

60:21

the benefits of Patrick's strategy and

60:23

Intel has all these empty fabs and

60:25

eventually

60:28

given the shortages we have of compute

60:30

those fabs are going to be filled.

60:31

>> So I think Taiwan Sim is in the process

60:33

of making a mistake but they're just so

60:35

paranoid about an overbuilt. Yeah.

60:37

>> And they're so skeptical. You know

60:38

they're the guys who met with Sam Alman

60:40

and laughed and said he's a podcast bro.

60:43

He has no idea what he's talking about.

60:44

You know they're terrified of an

60:47

overbuild. So it may be that Taiwan

60:50

Simei

60:51

singlehandedly that they're cautious

60:53

>> the breaks on the bubble

60:54

>> is is is is the governor um and you know

60:58

and we do like I think you know I think

61:00

governors are good it's good that you

61:01

know it's good that power is a governor

61:03

it's good that Taiwan sim is a governor

61:05

if Taiwan semi opens up at the same time

61:09

when you know data centers in space

61:11

relieve all power constraints but that's

61:13

like I don't know five six years away

61:15

that data centers in space or majority

61:17

of deployed megawatt like yeah I think

61:19

you get it overbuild really fast but

61:21

just we have these two really powerful

61:23

natural governors

61:25

>> and I think that's good you know like

61:26

smoother and longer is good

61:28

>> we haven't talked about the power other

61:30

than alluding to it through the space

61:31

thing haven't talked about power very

61:33

much power was like the most

61:35

uninteresting topic because there's the

61:36

de demand and nothing really changed for

61:38

like a really really long time all of a

61:40

sudden we're trying to figure out how to

61:41

get like gigawatts here there and

61:42

everywhere how do you think about are

61:44

you interested in powers

61:45

>> I'm very interested I do feel lucky in a

61:47

prior life I was the sector leader for

61:49

the telecom and utilities team.

61:50

>> Okay,

61:51

>> I I do have some base level of

61:53

knowledge. So one, you know, having um

61:56

having watts as a constraint is like

61:58

really good for the most advanced

61:59

compute players because if watts are the

62:02

constraint,

62:03

>> the price you pay for compute is

62:04

irrelevant. The TCO of your compute is

62:06

absolutely irrelevant because if you

62:09

could get 3x or 4x or 5x more tokens per

62:12

watt, that is literally three or 4x or

62:14

5x more revenue.

62:17

And so, you know, it's just like if

62:18

you're going to build a like, okay, like

62:21

an advanced data center costs 50

62:23

billion. A data center with your ASIC

62:24

maybe costs 35 billion, but if that $50

62:27

billion revenue, if that $50 billion

62:30

data center pumps out 25 billion of

62:33

revenue and your ASIC data center at 35

62:35

billion is only pumping out 8 billion,

62:37

well, like you're, you know, you're

62:39

pretty bummed. It's good for like all of

62:41

the most advanced

62:43

technologies in the data center which is

62:45

exciting to me as an investor. So as

62:47

long as power is a governor the best

62:50

products are going to win irrespective

62:52

of price and have crazy pricing power.

62:54

Okay, I think that's that's the first

62:56

implication that's really important to

62:58

me. Second, it is in the only solutions

63:01

to this. We just can't build nuclear

63:03

fast enough in America. Like as much as

63:06

we would love to build nuclear quickly,

63:08

we just can't. We just can't. Yeah,

63:10

>> it's just too hard, you know. Um, NEPA,

63:13

all these rules, like it's just it's too

63:15

hard. Like a a rare ant that we could

63:17

move and it could be in a better

63:19

environment can totally delay the

63:22

construction of a nuclear power plant.

63:24

You know, one ant. It's crazy actually.

63:27

Um, like humans need to come first. We

63:29

need to have a humanentric view of the

63:31

world. But like the solutions are

63:32

natural gas and solar. And the great

63:35

thing is the great thing about these AI

63:37

data centers is apart from the ones that

63:39

you're going to do inference on, you can

63:41

locate them anywhere. So I think you

63:43

were going to see and you're this is why

63:45

you're seeing all this activity in

63:46

Abalene, you know, because it's in the

63:48

middle of a big natural gas basin and we

63:50

have a lot of natural gas in America

63:52

because of fracking. You I think we

63:54

we're going to have a lot of natural gas

63:55

for a long time. We ramp production

63:56

really fast. So I think this is going to

63:59

be solved. You know, you're going to

64:00

have power plants fed by gas or solar. I

64:03

think that's the solution. And you know,

64:05

you're already, you know, all these

64:06

turbine manufacturers were reluctant to

64:08

expand capacity. Caterpillar just said,

64:10

"We're going to increase capacity by 75%

64:13

over the next few years." So like the

64:15

system on the power side is beginning to

64:17

respond. One of the reasons that I

64:19

always so love talking to you is that

64:20

you do every like you do as much in the

64:22

top 10 companies in the world as you do

64:24

looking at brand new companies with, you

64:26

know, entrepreneurs that are 25 years

64:28

old trying to do something amazing. And

64:30

so you have this very broad sense of

64:32

what's going on. If I think about that

64:34

second category of young enterprising

64:36

technologists who now are like AI,

64:39

they're like kind of the first

64:40

generation of AI native entrepreneurs.

64:43

What are you seeing in that group that's

64:45

notable or surprising or interesting?

64:47

>> These young CEOs, they're just so

64:49

impressive in all ways and they get more

64:52

polished faster. And I think the reason

64:54

is is they're talking to the AI.

64:56

>> How should I deal with pitching this

64:57

investor? I'm meeting with Patrick

65:00

Oanaughy. What What do you think the

65:02

best ways I should pitch him are?

65:03

>> Yeah. And it works.

65:04

>> Do a deep research. And it's good. You

65:06

know, hey, I have this difficult HR

65:09

situation.

65:10

>> How would you handle it?

65:11

>> That's correct.

65:12

>> And it's good at that. How would you,

65:15

you know, we're struggling to sell our

65:17

product. What changes would you make?

65:20

And it's really good at all of that

65:22

today.

65:23

And so, and that goes to these, you

65:26

know, VCs are seeing massive AI

65:29

productivity in all their companies.

65:30

It's because their companies are full of

65:32

these, you know, 23, 24 or, you know,

65:35

even younger AI natives. I've been so

65:38

impressed with like young investment

65:41

talent

65:42

>> and it's just part of it. Like your

65:44

podcast is part of that. There's just,

65:47

you know, knowledge and very very

65:49

specific knowledge has became so

65:52

accessible, you know, through podcasts

65:54

and the internet. Impressive young

65:55

people

65:56

>> come in and they're just I feel like

65:58

they're where I was as an investor like

66:00

in my, you know, early 30s and they're

66:02

22 and I'm like, "Oh my god,

66:04

>> like I have to run so fast to keep up."

66:06

these kids who are growing up native in

66:09

AI, they are just proficient with it um

66:12

in a way that I am trying really hard to

66:14

become.

66:15

>> Can we talk about semi VC specifically

66:18

and like what is interesting in that

66:19

universe?

66:20

>> Oh, just the one thing I just think that

66:21

I just think is so cool about it and so

66:23

underappreciated is your average

66:25

semiconductor venture founder is like 50

66:28

years old.

66:29

>> Okay. and Jensen and what's happened

66:32

with Nvidia and the market cap of Nvidia

66:34

has like singlehandedly

66:36

ignited semiconductor venture but the

66:38

way it's ignited it's ignited in an

66:39

awesome way that's like really good for

66:42

actually Nvidia and Google and everyone

66:44

>> is like let's just say you were the best

66:46

DSP architect in the world you had made

66:49

for the last 20 years every two years

66:52

because that's what you have to do

66:53

semiconductors it's like every two years

66:55

you have to win run a race

66:57

>> and if you won the last race you start

66:59

like a foot ahead

67:01

>> and over time those compound um and make

67:04

each race easier to win but like maybe

67:06

that person and his team maybe he's the

67:08

head of networking at a big public

67:10

company he's making a lot of money and

67:12

he has a good life and then because he

67:14

sees these outcomes and the size of the

67:17

markets in the data center he's like wow

67:18

why don't I just go start my own company

67:20

but the reason that's important is that

67:22

you know I forget the number but I mean

67:23

there are thousands of parts in a

67:24

blackwell rack and you know and there's

67:26

thousands of parts in a TPU rack And in

67:29

the Blackwell rack, you know, maybe

67:30

Nvidia makes,

67:32

I don't know, two two or 30 hundred of

67:34

those parts. And, you know, same thing

67:37

in an AMD rack. And they need all of

67:40

those other parts to accelerate with

67:42

them.

67:44

>> So, they couldn't go to this one-year

67:46

cadence if the rest everything was not

67:51

>> keeping up with them. The fact that

67:52

semiconductor venture venture has come

67:54

back with a vengeance, you know, Silicon

67:55

Valley stopped being Silicon Valley long

67:57

ago. My little firm maybe has done more

67:59

semiconductor deals in the last seven

68:01

years than the top 10 VCs combined, you

68:04

know, but that's really really important

68:06

because now you have an ecosystem of

68:09

companies who can keep up and then that

68:12

ecosystem of these venture companies is

68:14

putting pressure on the public companies

68:17

that are also need to part of part of

68:19

this if we're going to go to this annual

68:21

cadence which is just so hard. Um, and

68:24

it's one reason I'm really skeptical of

68:25

these AS6 that don't already have some

68:29

degree of success. So, I do think that's

68:31

a super super important dynamic and one

68:34

that's

68:35

absolutely foundational and necessary

68:37

for all of this to happen

68:39

>> because not even Nvidia can do it alone.

68:41

Not AMD can't do it alone. Google can't

68:43

do it alone. You need, you know, the

68:45

people who make the transceivers. You

68:47

need the people who make the wires, who

68:49

make the back blades, you know, who make

68:51

every who make the lasers. They all have

68:54

to accelerate with you. And one thing

68:55

that I think is very cool about AI as an

68:57

investor is it's just it's the first

68:58

time where every level of the stack

69:02

>> that I look at at least the most

69:04

important competitors are public and

69:05

private,

69:06

>> you know. So Nvidia they're they're very

69:09

important you know private competitors

69:11

you know Broadcom important private

69:12

competitors Marll important private

69:14

competitors you know luminum coherent

69:16

all these companies um you know there's

69:18

even like a wave of innovation in memory

69:20

which is really exciting to see because

69:22

memory and is such a gating factor by

69:25

the way something that could slow all

69:26

this down and be a natural governor is

69:27

if we get our first true DRAM cycle

69:30

since the late

69:31

>> 90s say more what that means

69:32

>> you know if like a DRAM wafer is like

69:35

valued at like a 5 karat a diamond in

69:37

the '90s when you had these true

69:39

capacity cycles before Taiwan semi kind

69:41

of smoothed everything out and DRAM

69:43

became more of an oligopoly. You know,

69:45

you would have these crazy shortages

69:46

where the price would just go 10x things

69:49

that are unimaginable

69:52

>> relative to the last 25 years where like

69:54

a giant DRAM cycle, a good DRM cycle is

69:56

the price start stops going down. An

69:59

epic cycle is maybe it goes up, you

70:00

know, whatever it is 30 40 50%. But I

70:03

mean, if it starts to go up by X's

70:05

instead of percentages, that's a whole

70:07

different game. By the way, we should

70:09

talk about SAS.

70:10

>> Yeah, let's talk about it. What do you

70:11

think's going to happen?

70:12

>> Application SAS companies are making the

70:14

exact same mistake that brick-andmortar

70:16

retailers did with e-commerce.

70:18

>> So, brick and mortar retailers um you

70:20

know, particularly after the um you

70:23

know, the the telecom bubble crashed,

70:25

you know, they looked at Amazon and they

70:26

said, "Oh, it's losing money. You know,

70:28

e-commerce is going to be a low margin

70:30

business." you know, how how can just,

70:32

you know, from first principles, how can

70:35

it ever be more efficient as a business?

70:37

Right now, our customers pay to

70:40

transport themselves to the store and

70:42

then they pay to transport the goods

70:43

home. How could it ever be more

70:45

efficient if we're, you know, sending

70:47

shipments out, you know, to individual

70:49

customers, you know, and Amazon's

70:51

vision, of course, well, eventually

70:52

we're just going to go down a street and

70:53

drop off a package at every house. And

70:55

so, they did not invest in e-commerce.

70:57

They they clearly saw customer demand

70:59

for it, but they did not like the margin

71:02

structure of e-commerce. That is the

71:04

fundamental reason that essentially

71:06

every brick brick-and-mortar retailer

71:08

was really slow to invest in e-commerce.

71:10

And now here we are and you know Amazon

71:12

has higher margins in their North

71:14

American retail business than a lot of

71:16

retailers that are mass market retailers

71:18

you know so margins can change and if

71:20

there's a fundamental transformative

71:23

kind of um new new technology that

71:26

customers are demanding it's always a

71:28

mistake not to embrace it

71:30

>> and that's exactly what the SAS

71:32

companies are doing they have their 70

71:34

80 90% gross margins and they are

71:37

reluctant to accept AI gross margins you

71:39

know the very nature of AI is you know

71:41

software you write it once and it's

71:43

written very efficiently and then you

71:45

can distribute it broadly at very low

71:47

cost and that's why it was a great

71:48

business AI is the exact opposite where

71:52

you have to recomputee the answer every

71:53

time and so you know a good AI company

71:57

might have gross margins of 40%.

72:00

Now, the crazy thing is because of those

72:02

efficiency gains, they're generating

72:04

cash way earlier than other people, you

72:07

know, than other than SAS companies did

72:09

historically, but they're generating

72:10

cash earlier, not because they have high

72:12

gross margins, but because they have

72:13

very few human employees. And it's just

72:16

tragic to watch all of these companies

72:18

like you want to have an agent, it's

72:21

never going to succeed if you're not

72:23

willing to run it at a sub 35% gross

72:26

margin

72:27

>> because that's what the AI natives are

72:29

running it at. Yeah,

72:30

>> maybe they're running it at 40. So if

72:31

you are trying to preserve an 80% gross

72:34

margin structure, you are guaranteeing

72:37

that you will not succeed at AI.

72:40

>> Absolute guarantee. And this is so crazy

72:43

to me because one, we have an existence

72:46

proof for software investors being

72:49

willing to tolerate gross margin

72:51

pressure as long as gross profit dollars

72:53

are okay. And it's called the cloud.

72:55

People don't remember but you know when

72:58

Adobe converted from on premise to uh

73:02

the CL you know to a SAS model not only

73:04

did their margins implode their actually

73:06

revenues imploded too because you went

73:08

from charging up front you know to

73:10

charging over a period of years.

73:12

Microsoft, it was less dramatic, but you

73:14

know, Microsoft was a tough stock in the

73:17

early, you know, in the early days of

73:19

the cloud transition because investors

73:21

were like, "Oh my god, you're an 80%

73:23

gross margin business." And the cloud is

73:25

the 50s and they're like, "Well, it's

73:27

going to be gross profit dollar

73:29

creative. It probably will improve those

73:31

margins over time." Microsoft, they

73:33

bought GitHub and they use GitHub has a

73:36

distribution channel for, you know, uh,

73:40

or Copilot. co-pilot for coding that's

73:42

become a giant business a giant business

73:45

now for sure it runs at much lower gross

73:48

margins but there are so many SAS

73:51

companies like I can't think of a single

73:53

application SAS company that could not

73:56

be running a successful agent strategy

73:59

they have a giant advantage over these

74:00

AI natives and that they have a cash

74:03

generative business

74:05

>> like and I think there is room for

74:07

someone to be a new kind of activist or

74:10

constructive ist and just go to SAS

74:13

companies and say stop being so dumb.

74:18

>> All you have to do is say here are my AI

74:20

revenues

74:22

>> and here are my AI gross margins and you

74:23

know it's real AI because it's low gross

74:25

margins. I'm going to show you that and

74:28

here's a venture competitor over here

74:30

that's losing a lot of money. So maybe

74:31

I'll actually take my gross margins to

74:33

zero for a while but I have this

74:35

business that the venturef funed company

74:37

doesn't have. And this is just such a

74:40

like obvious playbook that you can run

74:43

Salesforce, Service Now, HubSpot,

74:46

GitLab, Atlassian,

74:48

all of them could run this. And the way

74:51

that those companies could or should

74:53

think about the way to use agents is

74:55

just to ask the question, okay, what are

74:57

the core functions we do for the

74:59

customer now? Like how can we further

75:00

automate that with agents effectively?

75:02

Or is it some other

75:03

>> 100% just like if you're in CRM? Well,

75:05

what our customers do, they talk to talk

75:07

to their customers. Yeah,

75:09

>> we're customer relationship management

75:10

software and we do some customer

75:12

support, too.

75:13

>> So, make an agent that can do that,

75:14

right?

75:15

>> And sell that,

75:16

>> right,

75:16

>> at 10 to 20% and let that agent access

75:19

all the data you have,

75:20

>> right?

75:20

>> Cuz what's happening right now is

75:23

another agent,

75:24

>> another agent

75:25

>> made by someone else is accessing your

75:27

systems

75:28

>> to do this job,

75:29

>> pulling the data into their system,

75:31

>> and then you will eventually be turned

75:32

off. And it's just crazy. And it's just

75:35

because, oh wow, but we want to preserve

75:37

our 80% gross margins. This is a life

75:41

ordeath decision. And essentially

75:43

everyone except Microsoft

75:46

is failing it. To quote that memo from

75:49

that um Noia guy long ago, like their

75:51

their platforms are burning.

75:52

>> Burning platform. Yeah.

75:53

>> Yeah. There's a really nice platform

75:56

right over there and you can just hop to

75:58

it and then you can put out the fire in

76:00

your platform that's on fire. And now

76:03

you GOT TWO PLATFORMS AND IT'S GREAT.

76:05

You know,

76:05

>> your data centers and space thing makes

76:07

me wonder if there are other kind of

76:09

like less discussed off-the-wall things

76:13

that you're thinking about in in the

76:14

markets in general that we haven't

76:16

talked about. It does feel like since

76:20

2020 kicked off and you know 2022

76:22

punctured this kind of a series of

76:24

rolling bubbles you know so in 2020 you

76:27

know there was a bubble in like EV

76:29

startup EVs company startup EVs that

76:31

were not Tesla and that's for sure a

76:33

bubble and they all went down you know

76:35

99%. And there was kind of a bubble in,

76:37

you know, more speculative stocks. Uh,

76:40

you know, then we had the meme stocks,

76:42

you know, GameStop. And now it feels

76:44

like the rolling bubble is in nuclear

76:47

and quantum.

76:48

>> And these are, you know, fusion and SMR.

76:51

Like it's it would be a, you know, it's

76:54

it would be a transformative technology.

76:55

It's amazing. But sadly from my

76:59

perspective, none of the public ways you

77:01

can invest in this are really good

77:04

expressions of this theme are likely to

77:05

succeed or have any real fundamental

77:07

support. And same thing with quantum

77:10

like we I've I've been looking at

77:11

quantum for 10 years. We have a really

77:13

good understanding of quantum and the

77:16

public quantum companies again are not

77:18

the leaders. You know, from my

77:20

perspective, the leaders in quantum

77:21

would be Google, IBM, and then the

77:24

Honeywell Quantum, you know. So the

77:26

public ways you can invest in this theme

77:28

which probably is exciting are not the

77:30

best. So you have two really clear

77:31

bubbles. I also think quantum supremacy

77:34

is very misunderstood. People hear it

77:36

and I think that mean it means that

77:38

quantum computers are going to be better

77:39

than classical computers at everything.

77:41

With quantum you you can do you can do

77:44

some calculations that classical

77:46

computers cannot do.

77:49

>> That's it. That's going to be really

77:50

useful and exciting and awesome. But it

77:54

doesn't mean that quantum takes over the

77:56

world. The thought that I have had, this

77:59

is maybe less related

78:02

to markets than just AI. I have just

78:05

been fascinated that for the last two

78:08

years, whatever AI needs

78:12

>> to keep growing and advancing, it gets.

78:17

Have you ever seen public opinion change

78:21

so fast in the United States on any

78:24

issue has nuclear power?

78:27

>> Just happened like that.

78:28

>> Like that.

78:30

And like why did that happen like right

78:34

when AI needed it to happen? Now we're

78:36

running up on boundaries of power on

78:38

earth. you know, all of a sudden data

78:42

centers in space,

78:43

>> you know, just it's just a little

78:46

strange to me that whenever there is

78:50

something

78:51

>> a bottleneck

78:51

>> that a bottleneck that might slow it

78:53

down,

78:55

everything accelerates, you know, like

78:58

Reuben is going to be such an easy,

79:00

seamless transition relative to

79:02

Blackwell and Reuben's a great chip and

79:04

then you you know, you have MI, you

79:05

know, AMD getting into the game with the

79:07

MI450. Like it's just whatever AI needs,

79:10

it gets.

79:11

>> You're a deep reader of sci-fi, so uh

79:13

Yeah, exactly. You're making me think of

79:15

of Kevin Kelly's great, uh book, What

79:17

Technology Wants. He calls it the

79:19

technium, like the like the overall mass

79:21

of technology that just like is supplied

79:23

by humans.

79:25

>> Absolutely.

79:25

>> To grow more powerful.

79:26

>> Yes. It just wants to grow more and more

79:28

powerful. And now we're going into an

79:29

instate.

79:30

>> I have a selfish closing question.

79:31

Speaking of speaking of uh young people,

79:34

so my kids who are 12 and 10, but

79:36

especially my son who's older is

79:38

developing an interest in what I do,

79:40

which I think is quite natural. And I'm

79:42

going to try to start asking my friends

79:44

who are the most passionate about

79:46

entrepreneurship and investing why they

79:48

are so passionate about it and what

79:50

about it is so interesting and

79:52

life-giving to them. How would you pitch

79:55

what you've done, the career you built,

79:57

the this part of the world to a young

80:00

person that's interested in this?

80:01

>> I do believe at some level kind of

80:03

investing is the search for truth. And

80:05

if you find truth first,

80:08

and you're right about it being a truth,

80:10

that's how you generate alpha. And it

80:12

has to be a truth that other people

80:13

don't have have not yet seen. You're

80:16

searching for hidden truths. Earliest

80:18

thing I can remember is being interested

80:21

in history. You know, looking at books

80:24

with pictures of the Phoenetians and the

80:25

Egyptians and the Greeks and the Romans

80:28

and pyramids. I loved history.

80:31

>> I vividly remember like in the I think

80:33

in the second grade as my dad drove me

80:36

to school every day, he would we'd go we

80:38

went through the whole history of World

80:39

War II in one year and I loved that. And

80:43

then that translated into a real

80:46

interest in current events very early.

80:49

So, like as a pretty young person, you

80:52

know, I don't know if it was eighth

80:53

grade or seventh grade or ninth grade,

80:55

like I was reading the New York Times

80:58

and the Washington Post and I would get

81:00

so excited when the mail came because it

81:03

meant that maybe there was an economist

81:05

or a Newsweek or a Time or US News and I

81:08

was really into current events, you

81:10

know, because current events is kind of

81:11

like applied history and watching

81:12

history happen and like thinking about

81:15

what might happen next.

81:17

And you know, I didn't know anything

81:20

about investing. My parents were both

81:22

attorneys. Like I was anytime I won an

81:24

argument, I was super rewarded.

81:26

>> Like, you know, if I could make a

81:28

reasonable argument why I should stay up

81:29

late, my parents would be so proud and

81:31

they'd let me stay up late, but I had to

81:33

beat them, you know, like I was just

81:34

kind of going through life and, you

81:36

know, I really love to ski and I love

81:37

rock climbing and I go to college and

81:40

rock climbing is, you know, by far the

81:42

most important thing in my life. I

81:43

dedicated myself to it completely. I did

81:46

all my homework at the gym. I got to the

81:48

rock climbing gym like at 7 am would,

81:51

you know, skip a lot of classes to stay

81:53

in the gym. I'd do my homework on like a

81:55

big bouldering mat.

81:57

>> Like every weekend I went and climbed

81:59

somewhere with the Dartmouth

81:59

Mountaineering Club. And as part of

82:01

that, like on climbing trips,

82:03

>> you know, maybe we'd play poker. The

82:05

movie came out while I was in college.

82:07

We started playing poker. I like to play

82:09

chess. Um, and I was never that good at

82:11

chess or poker. You never really

82:13

dedicated myself to either. And my plan

82:16

like you know after two or three years

82:17

of college was I was going to leave. I

82:20

was going to work as a I I was a ski

82:23

bomb at Alta in college. I I was a

82:25

housekeeper. I've cleaned a lot of

82:27

toilets. Um it is it was shocking to me

82:30

how people treated me and it is like

82:32

permanently impacted how I treat other

82:34

people. You know

82:35

>> like I want you know like you'd be

82:37

cleaning somebody's room and they'd be

82:38

in it and they'd be reading the same

82:40

book as you and like you know you'd say

82:42

oh that's a great book. you know, I'm

82:43

about where you are and a they look at

82:46

you like you're a space alien, like you

82:49

speak

82:50

>> and then they get even more shocked. You

82:52

read, you know, so it like had a big

82:55

impact on how I've like just treated

82:57

everyone since then. But anyways, I was

82:58

going to be a ski bum in the winters.

83:00

Um, I was going to work on a river in

83:02

the in the summers and that was how I

83:04

was going to support myself. And then I

83:06

was going to climb in the shoulder

83:07

seasons, going to try and be a wildlife

83:09

photographer and write the next great

83:11

American novel. I can't believe I never

83:12

knew this.

83:13

>> That was my plan. This was like my plan

83:14

of record. I was really lucky. My

83:16

parents very supportive of everything I

83:18

wanted to do. My parents had very strict

83:20

parents, so of course they're extremely

83:21

permissive with me. So, you know, I'll

83:23

probably end up being a strict parent.

83:25

Just the cycle continues.

83:27

>> My parents were lawyers. You know, they

83:28

they they had done reasonably well. Um

83:30

they both grew up in in um I would say

83:33

very economically disadvantaged

83:34

circumstances. You know, like my dad

83:36

talks about like he remembers every

83:38

person who bought him a beer

83:40

>> just because he could not he couldn't

83:42

afford a beer. You know, he worked the

83:43

whole way through college. He was there

83:45

on a scholarship. You know, he had one

83:47

pair of shoes all through high school.

83:49

But anyways, and so they were super on

83:50

board with this plan and I'd been very

83:52

lucky. They sent me to college and I

83:54

didn't have to pay pay for college. They

83:55

paid for my college education. They

83:57

said, "You know, Gavin, we think this

83:58

plan of being, you know, ski bum, river

84:01

rafting guide, wildlife photographer,

84:04

climbing the shoulder seasons, tried to

84:05

write a novel. We think it sounds like a

84:07

great plan, but you know, we've never

84:08

asked you for anything. We've haven't

84:10

encouraged you to study anything. We've

84:11

supported you in everything you've

84:13

wanted to do. Will you please get one

84:15

professional internship, just one, and

84:18

we don't care what it is."

84:19

>> The only internship I could get, this

84:21

was at the end of my sophomore summer at

84:22

Dartmouth, was an internship with

84:24

Donaldson Lufkin Engineer. Um, DJ, my

84:27

job was to every time DJ published a

84:30

research report, it was in like the

84:32

private wealth management division and I

84:34

worked for the guy who ran the office

84:36

and my job was whenever they produced a

84:39

piece of research, I would go through

84:40

and look at which of his clients owned

84:42

that stock.

84:44

Then I would put the research I would

84:46

mail it to the clients, you know. So

84:48

this day we wrote on General Electric.

84:51

So, I need to mail the GE report to

84:52

these 30 people

84:54

>> and then I need to, you know, email the

84:56

Cisco report to these 20 people. And

84:58

then I started like reading the reports

84:59

and I was like, "Oh my god, this is like

85:02

the most interesting thing imaginable."

85:04

Investing. I kind of conceptualized it.

85:06

It's a game of skill and chance, kind of

85:08

like something like poker. Um, and you

85:10

know, there's obviously chance in

85:12

investing. you know, like if you're an

85:14

investor in a company and a meteor hits

85:15

their headquarters, like that's that's

85:17

bad luck, but like you own that outcome.

85:20

Um, so there is chance um that is

85:23

irreducible, but there's skill, too. So

85:25

that really appealed to me. And the way

85:27

you got an edge in this the greatest

85:30

game of skill and chance imaginable was

85:33

you had the most thorough knowledge

85:36

possible of history. And you intersected

85:39

that with the most accurate

85:41

understanding of current events in the

85:43

world

85:44

>> to form a differential opinion on what

85:46

was going to happen next in this game of

85:48

skill and chance. Which stock is

85:50

mispriced in the Perry Mutual system?

85:52

>> That is the stock market. And that was

85:55

like day three. I went to the bookstore

85:59

and I bought like the books that they

86:00

had which were Peter Lynch's books. I

86:02

read those books in like two days. I'm

86:04

I'm a very fast reader. And then I read

86:06

all these Warren books, books about

86:08

Warren Buffett. Then I read Market

86:10

Wizards. Then I read Warren Buffett's

86:12

letters to his shareholders. This is

86:14

like during my internship. Then I read

86:16

Warren Buffett's letters to his

86:17

shareholders again. Then I taught myself

86:19

accounting. There's this great book, Why

86:21

Stocks Go Up and Down. Then I went back

86:22

to school. I changed my majors from

86:24

English and history to history and

86:25

economics. And I never looked back. And

86:28

it consumed like I continued to really

86:32

focus on climbing. I would be in the gym

86:34

and I would print out everything that

86:37

the um people on the mly fool wrote.

86:39

they had these fools and and you know

86:42

they they were they were early to

86:43

talking about return on invested capital

86:45

and in incremental ROIC is like a really

86:48

important indicator and I would just

86:50

read it and I would underline it and I'd

86:52

read books and then I'd read the Wall

86:54

Street Journal and then eventually there

86:56

was a computer terminal finally set up

86:58

near the gym and I'd go to that gym and

86:59

just you know read news about stocks and

87:01

it was the most important thing in my

87:03

life and like I barely kept my grades up

87:05

and yeah that's how I got into it man

87:06

history current events skill and dance

87:09

and I am a competitive person and I've

87:12

actually never been good at anything

87:14

else. Okay, I got picked last for every

87:17

sports team. Like I love to ski. I've

87:19

literally spent a small fortune on

87:22

private skiing lessons. I'm not that

87:23

good of a skier. I like to play

87:25

pingpong. All my friends could beat me.

87:28

Um I tried to get really good at chess

87:30

and this was before the you know when

87:33

you you actually had to play the games.

87:34

It was before it was easy to do it on

87:36

the phone. And my goal was to beat one

87:38

of the people. I'm sure there's a park

87:40

somewhere.

87:40

>> It's literally right there. Famous one

87:42

is right there.

87:42

>> Okay. Well, there's one in Cambridge.

87:44

And I wanted to beat one of them. Never

87:45

beat one of them. Never been good at

87:47

anything. I thought I would be good at

87:49

this.

87:50

>> And the idea of being good at something

87:52

other than taking a test

87:55

that was competitive was very appealing

87:58

to me.

87:59

>> And so I think that's been a really

88:00

important thing, too. And to this day,

88:02

this is the only thing I've been vaguely

88:04

competitive at. I'd love to be good at

88:06

something else. I'm just not, you know.

88:09

>> I think I'm going to start asking this

88:10

question of everybody. Uh, the ongoing

88:12

education of Pearson May, amazing place

88:14

to close. I love talking about

88:16

everything so much.

88:17

>> This is great, man. Thank you. Thank

88:18

you. Thank you.

Interactive Summary

This conversation features an in-depth discussion on the rapid advancements and strategic implications of Artificial Intelligence, particularly focusing on the hardware race between Nvidia and Google's TPUs. It highlights the critical role of chip development, such as Nvidia's Blackwell, and the emergence of new scaling laws in AI. The discussion also touches upon the economic factors driving AI development, the potential for AI to transform various industries, and the future of computing with concepts like data centers in space. Furthermore, it explores the challenges faced by AI companies, the evolving landscape of AI infrastructure, and the impact of AI on business models and investment strategies.

Suggested questions

11 ready-made prompts