HomeVideos

Energy! Chips! ...and INSURANCE? (WTF)

Now Playing

Energy! Chips! ...and INSURANCE? (WTF)

Transcript

687 segments

0:00

So, I made this video uh or I did this

0:02

research kind of out of my own curiosity

0:04

and I realized that it would make a good

0:05

video. And basically, the question is

0:07

why isn't AI going faster and how can we

0:09

accelerate it faster? But really, this

0:12

is just a deep dive into what is the

0:14

holdup with artificial intelligence as

0:16

of the beginning of 2026. So, let's get

0:18

into it. The TLDDR upfront is that

0:23

everything uh the the biggest kind of

0:26

barriers is the thermodynamic wall. of

0:28

power, grid interconnection, physics,

0:30

that sort of thing. This comes down to

0:33

the grid capacity. This comes down to

0:35

transformers on the grid. This comes

0:37

down to actual generation capacity. So

0:39

basically everything to do with energy

0:42

is there. Then there's the supply which

0:44

is also structural. Um mostly the the

0:47

co-as so the chip on wafer on on

0:51

substrate. So if you see the COS, this

0:54

is basically the whole package getting

0:56

delivered. that is now the the actual

0:59

bottleneck and so this is why like

1:01

memory is sold out and and that sort of

1:03

thing. So then you have operational

1:05

friction which is further down the

1:07

stack. So this is enterprise execution,

1:09

data quality and ROI and then finally uh

1:14

things like AI safety and security and

1:16

public opinion that's actually just

1:18

noise. So when you have people saying oh

1:20

well you have to prove that AI is safe.

1:22

It's like actually and I mentioned this

1:24

in in other videos when and I said that

1:26

like serious researchers, policy people,

1:29

nobody takes extra seriously anymore. Um

1:32

it's basically just been

1:33

operationalized. The actual conversation

1:36

sphere has not caught up with reality.

1:38

Um which is really interesting and it's

1:39

kind of frustrating and so it's it's

1:41

basically kind of become a dog and pony

1:43

show of like you know but Ellie Azer

1:44

Yukowski said this and it's like nobody

1:47

who actually has decision power actually

1:49

cares. Um the and one thing I will say

1:51

is that in in layer three here is most

1:54

interestingly is insurance and we'll

1:56

talk about that but insurance doesn't

1:57

know how to price AI risk and that's

1:59

actually become more of a source of

2:02

friction than people being like AI is

2:04

just slop nobody cares that AI is slop

2:05

from a from an acceleration standpoint

2:07

so next infinite incentives abundant

2:10

capital what I have always said is that

2:13

acceleration is the default policy and

2:15

there's a lot of reasons that

2:16

acceleration is the default policy first

2:19

and foremost most being geopolitical

2:21

competition between America and China.

2:24

Uh all the great powers know that

2:26

artificial intelligence is the defining

2:28

technology of the next century or at

2:30

least the next few decades. And so

2:33

looking at the past uh as as guidance

2:36

for the future,

2:38

America and China and to a lesser extent

2:41

Russia, although they're not really a

2:43

player in this, but they all know that

2:45

artificial intelligence is the way to

2:47

go. Then you have market competition

2:48

which is basically the gold rush of all

2:51

the big tech companies wanting to get in

2:53

on artificial intelligence. So you say

2:55

okay well how much is being spent? So

2:57

$350 billion annual uh spend on on

3:01

hyperscalers. So that's um basically

3:03

data centers. $22 billion projected um

3:07

2025 VC funding that'll probably be

3:10

higher in 2026 which amounts to about

3:13

1.9% of GDP rivaling the Manhattan

3:16

project. So we are basically living

3:18

through a decentralized Manhattan

3:20

project. Um so the interstate highway

3:22

system, the Apollo program, the

3:24

Manhattan project, the 2025 and 20 and

3:27

beyond AI buildout is similar in scale.

3:31

We have achieved that. So the idea

3:33

however has been if we need it we can

3:35

buy it. The assumption that money

3:36

instantly dissolves physical

3:38

bottlenecks. It's just the bottlenecks

3:40

are not where we thought that they would

3:41

be. So that's the that's the level of

3:43

spend that we're seeing and the economy

3:45

is already restructuring around these

3:47

new problems. So how does that look?

3:50

Energy is the hardest bottleneck. This

3:52

is the hardest stop. So the average US

3:54

interconnection weight is 7 years. So if

3:56

you want to build a new data center, you

3:58

have to wait on average 7 years to

4:00

connect that new data center to the

4:02

grid, which is insane because that would

4:05

put us in 2032 or 2033 by that by which

4:09

point we should be at like super

4:10

intelligence and beyond. So the demand

4:13

is is expected to grow from 4 gawatt

4:16

total of data centers. Um, and when you

4:18

when you say 4 gawatts total in 2024,

4:21

consider that Microsoft and Meta and XAI

4:24

and OpenAI are all trying to spin up one

4:27

to two gawatt data centers, multiple

4:30

each of them. So just in the next few

4:32

years, they're they're wanting to more

4:34

than double what the current demand was.

4:37

Uh, and so the aggregate demand is

4:39

expected to hit 134 gawatts of data

4:41

center over the next 5 or 6 years. uh or

4:45

I guess the six years between 2024 and

4:47

2030. It's 2026 now. So we are on track

4:52

to have 134 gawatts of demand. For

4:55

comparison, your average run-of-the-mill

4:57

nuclear reactor produces about 1 gawatt.

5:00

Uh so that's a lot of juice. And I was

5:04

skeptical about these numbers. I'm like,

5:05

it can't possibly be that much. But when

5:07

you look at the rate of buildout and the

5:09

amount of backlog, like yes, that this

5:12

is not hypothetical future demand. This

5:14

is basically what is on deck and what

5:16

people are trying to get permitted and

5:17

approved right now. Uh we're going to

5:20

hit that and we might exceed it. That's

5:22

of course speculation on my part. We

5:24

might not hit that especially if we find

5:26

new efficiencies. Uh but energy is one

5:29

of the biggest bottlenecks. Now there

5:30

are solutions that didn't make it into

5:32

this presentation. So like um micro

5:35

grids where you have a combination of

5:36

solar batteries um and natural gas

5:39

turbines on site. Uh that's one answer.

5:42

uh directly connecting nuclear power

5:44

plants to data centers so that they're

5:45

not attached to the grid. Um basically

5:48

the grid is the bottleneck. So a lot of

5:50

well a lot of what these companies are

5:52

doing is becoming their own utility

5:53

companies as well. And we've seen this

5:55

like um there's a there's a site in the

5:58

far east of Russia where there's a

6:00

natural there's a natural dam or not a

6:02

natural dam there they use hydroelectric

6:05

to power aluminum smel sme sme sme sme

6:06

sme sme sme sme sme sme smelting and the

6:07

energy is just used primarily for the

6:10

aluminum because it's really expensive

6:12

so that's the only reason it exists now

6:13

obviously we don't have Siberia in

6:15

America where we just have thousands of

6:17

miles of unused space to spread out like

6:19

that but we do have lots and lots of

6:20

desert so we can build solar in that

6:22

anyways riffing on a little side tangent

6:25

on where things are at in terms of

6:26

energy. So, next is physics. So, this is

6:30

this is the highest latency part. So,

6:32

the the idea of getting the grid, you

6:36

know, because we've got we've got the

6:37

whole United States to power. So,

6:39

getting the grid up to speed is going to

6:42

take a lot of time and a lot of work. Um

6:44

because that means new transmission

6:45

lines, new transformers, and that sort

6:47

of thing. the transformers themselves,

6:50

the delay, the lead time is up to 210

6:52

weeks, which I thought was nuts. 210

6:55

weeks. Um, so for com for for

6:58

comparison, uh, a year is what, 56

7:00

weeks, 54 weeks? I don't remember off

7:02

the top of my head, but that's basically

7:04

four years almost. Um, and these are not

7:06

the little transformers that you see at

7:08

like the power substation. These are the

7:10

gigantic transformers that you don't

7:11

see. These these are the house-sized

7:13

transformers. Um, and so we're we've got

7:16

a backlog of those in demand. And then

7:18

the nuclear delusion. So the idea is

7:21

that nuclear the earliest that nuclear

7:23

could get spun back up is like 2030s. So

7:26

what really is emerging from the data

7:28

and the research is that the crisis

7:29

window is going to be over the next

7:31

couple years. So 2026 through 2028 is

7:34

really when it's going to be the hardest

7:35

because uh as the infrastructure

7:38

projects ramp up and as the orders for

7:41

transformers ramp up, those will

7:43

eventually get solved over the next 2 to

7:45

3 years. But nuclear is not going to get

7:47

solved that quickly. Uh so small modular

7:50

reactors, they're going to be too late.

7:52

So what we really need is stuff that you

7:54

can spin up very quickly. So that's

7:57

natural gas turbines, that's solar,

7:59

that's uh grid scale batteries. And we

8:01

do have cheap grid scale batteries now,

8:03

which is the um uh iron air batteries.

8:05

So if you haven't heard of those,

8:06

they're not very efficient, but they're

8:08

dirt cheap and they last forever. Um and

8:11

when you don't when they don't need to

8:12

be portable, they can be really heavy.

8:14

So iron air batteries are probably the

8:16

way to go if I had to guess um as the as

8:18

the best grid scale uh battery

8:20

technology. Cuz a lot of people like,

8:22

"Yeah, but you need to cover base load."

8:23

It's like you don't really need to cover

8:25

base load if you have enough batteries.

8:27

Um, and that's, you know, solar, right?

8:29

A lot of people say, "Well, solar can't

8:30

do everything." Solar can do everything

8:32

when it's combined with grids scale

8:33

batteries. So, moving on, the supply

8:36

chain. So, this is this is this is

8:39

interesting. GPUs are no longer the

8:41

bottleneck. Um, so the logic dies. So,

8:44

that's that's the actual wafer. That's

8:46

no longer the bottleneck. High bandwidth

8:47

memory is now the bottleneck. So, this

8:49

is sold out through the end of the year,

8:50

which is why people are making memes

8:52

about how expensive memory is. Um, and

8:54

part of the reason for this is that um

8:57

all the all the companies that can make

8:59

memory, instead of making memory for

9:00

your laptop and your desktop, they're

9:02

saying, "Well, we can make a lot more

9:04

money on memory for uh for the for the

9:07

uh accelerator. So, we're going to make

9:08

memory for that instead." Um, so this is

9:11

basically a price signal saying, "Hey,

9:13

we need a lot more memory, so the people

9:15

who can make it are abandoning it." So

9:17

the legacy DDR3 and DDR4 they're

9:19

abandoned. Um and that means that

9:21

robots, autos, laptops and that sort of

9:24

thing are all facing memory crunches. AI

9:27

nodes uh so that the hyperscalers are

9:30

consuming all of the available capacity

9:33

to build memory and then Nvidia books

9:36

greater than 50% of all cos capacity. So

9:39

that's the chip on wafer on substrate.

9:42

So what we're seeing here is the entire

9:43

package. So if you hear packaging, this

9:47

is the the the GPU, the onboard memory,

9:50

the high bandwidth memory, and then the

9:52

rest of the substrate that you put it

9:54

on. Uh and Nvidia is booking most of it.

9:56

So Nvidia is dominating this space as

9:58

well. It will take time. You know,

10:00

remember a year or two ago, we said,

10:02

"Oh, we're going to have a GPU

10:04

shortage." Well, it took a year and a

10:06

half to two years to figure out the GPU

10:08

shortage shortage, sorry. So now it'll

10:11

take another year and a half to two

10:12

years to sort out the bandwidth and

10:14

substrate shortage. There we go. So

10:18

those are again the the market will fix

10:20

these problems. That's kind of what I

10:21

want to drive home here is uh energy.

10:24

This is this requires state

10:26

intervention. This requires lots of uh

10:28

cutting through red tape. This is not

10:31

something that the market on its own

10:32

will sort out uh post haste. This needs

10:36

regulatory help. This is purely market.

10:38

So the the the the processors and the

10:41

memory the market will sort that out.

10:43

It'll just take time. Now the other

10:46

bottleneck is data. So people keep

10:48

talking about data. Uh and you know the

10:52

the thing is I'm not convinced by this

10:55

argument because we are seeing a very

10:57

very sharp rise of synthetic data. Um

11:00

however it is still a concern because AI

11:03

trained on its own data does tend to

11:05

result in model collapse. But the latest

11:07

generation of AI models are trained on

11:10

more and more synthetic data. So I'm not

11:13

convinced that this is actually going to

11:14

be a bottleneck. Now at the same time,

11:17

it is true that the amount of raw human

11:20

data we're going to run out of it. But

11:22

as many of you have pointed out over the

11:24

past, humans do a lot more with a lot

11:26

less data. Like a human brain is trained

11:29

on like less than 1% of the of the

11:31

amount of data that a uh that an AI

11:33

model is trained on. And we're still

11:35

better. We're still smarter. So, we are

11:37

clearly doing something different. We're

11:39

not getting the most out of the data

11:41

that we have. So, and I always say like

11:43

necessity is the mother of invention,

11:45

but constraints are the father of

11:46

creativity. So, if we run out of data,

11:49

I'm convinced that we will find better

11:50

algorithms to make better use of the

11:52

data that we do have because we clearly

11:55

have enough data to generate, you know,

11:57

super intelligent human level

11:59

performance. Again, if a human can

12:01

spend, you know, just 20 years or 30

12:04

years learning less than 1% of the data

12:06

that we do have available and advance

12:08

physics, then AI should be able to do

12:10

the same. So AI is just not that

12:12

efficient at learning. So we will get

12:14

better at making AI learn. So again, I

12:17

don't I don't really see this as a

12:19

barrier, but this is one of the things

12:20

that came up in the research that people

12:22

are concerned about. So I wanted to at

12:24

least include it. Now, the $600 billion

12:27

question. So this is the infrastructure

12:29

spend and the ROI gap the market

12:31

rotation. So investors are fleeing

12:33

infrastructure for software

12:34

productivity. This is the most

12:36

interesting thing where um investors are

12:39

getting more and more wary of you know

12:41

hey write me a check for hundred billion

12:44

to build data centers. When are the data

12:45

centers going to turn a profit? Oh maybe

12:47

5 to 8 years. Investors don't like that.

12:49

They want to see returns on investment

12:51

very quickly. Uh so the hyperscalers

12:54

that raised hundred billion in debt in

12:56

2025 the investors are starting to look

12:58

a scance at them which is why you see

13:00

Sam Alman going to places like Saudi

13:02

Arabia saying hey can you can you

13:04

finance our data centers [snorts]

13:06

the question can spending 20% of GDP

13:08

sustain on belief alone? This is this

13:12

goes back to the bubble fears. Um, so

13:14

you know, yes, some of the most

13:16

profitable companies on the planet

13:17

already are using AI. Although you might

13:20

say, well, where's the evidence? We're

13:22

in Solo's paradox or the J curve of

13:24

productivity where AI is is uh the

13:27

profusion is happening. AI is getting in

13:29

into the market and it's saturating.

13:31

But, you know, while most companies are

13:34

using AI, it's not saturated to the

13:36

degree that you'd want it to be

13:37

saturated because it's simply still too

13:40

expensive. In the same way that in the

13:42

80s the personal computer was not as

13:43

mature as it is today. Um you know it's

13:46

like okay you have a PC on your desk and

13:48

it you know has a a monochrome screen

13:51

that can connect to a central database.

13:53

Great. That's definitely more productive

13:55

but it's not nearly as productive as

13:56

even just a cell phone is today. We are

13:59

at that era where yes it's it's there.

14:02

It's useful but it's not saturated. So

14:03

it'll take another few years to really

14:05

find all the uses to saturate the AI,

14:07

but in the meantime scaling up the

14:10

production of the AI. So we've got like,

14:11

you know, this log jam between it's not

14:14

as mature as it could be. It's still

14:15

expensive to run. And this is also very

14:17

similar to early internet. Um where, you

14:20

know, when you still had like dialup and

14:21

ISDN and that sort of thing. Um then you

14:25

had to or I no ISB in his book. um the

14:30

ISDN I'm remembering incorrectly DSL

14:32

basically digital subscriber lines um so

14:36

when the internet was new it was useful

14:38

but it was still too too expensive

14:40

because the the cost per packet was too

14:43

high now of course the internet is

14:44

basically free you just have a flat

14:46

monthly subscription and you get as much

14:48

internet as you need and it's at gigabit

14:50

plus speeds um so it'll take a it'll

14:53

take a little bit of time for us to

14:55

build up that much infrastructure until

14:57

the point that AI is just kind of a

14:58

background expense and you're not

14:59

thinking about the cost per token. So,

15:02

and in the meantime, the the enterprise

15:04

ROI is not necessarily manifesting as

15:07

quickly as they would like, although

15:08

most business leaders understand that AI

15:11

is very competitive and compelling and

15:13

and that they need to not be asleep at

15:14

the wheel. So, then of course the ROI is

15:18

about, you know, why 88% of pilots fail.

15:20

And of course, this is different from

15:21

the 95% of pilots that fail um from the

15:24

MIT study. I've talked about this plenty

15:26

of times, but I'll I'll go over it

15:27

briefly, which is basically uh most

15:29

pilots fail. Uh AI is no different.

15:32

That's the point of a pilot is to see

15:33

can we realize value. Um so you start

15:36

with 100% of AI pilots that you initiate

15:40

integration complexity and data mess and

15:42

then 12% reach production. So that's

15:44

still better than the 95% from the MIT

15:46

study. Uh but it's still the vast

15:48

majority fail. So what are the barriers?

15:51

One is number one is data quality. Um

15:53

and getting good data is always the

15:55

problem. I mean getting good data

15:57

integration and data governance is

15:59

always step one. Uh number two is

16:01

integration with legacy systems. This is

16:03

another thing is big companies are

16:05

running stuff that is 20 30 plus years

16:07

old. Um and it it many of them don't

16:09

even have modern APIs. Some of them

16:11

don't even have modern operating

16:12

systems. Uh some of them are running on

16:15

old operating systems like SCO Unix. Um

16:17

and then they also have a lack of

16:19

talent. So the lack of talent, the lack

16:22

of integration, the lack of being AI

16:24

ready, it has nothing to do with safety

16:27

or ethics. The barriers are very mundane

16:30

which is cost and ROI and then also

16:32

insurance oddly enough. Um so then when

16:35

we say what what actually gets talked

16:37

about. So the the vast majority of

16:40

public opinion about AI is safety and

16:44

you know uh AI art is crap and but what

16:48

about copyright and what about the what

16:49

about the actors and general regulation

16:52

about consumer protections? That's where

16:55

most people are talking. That's where

16:56

the conversation is. But the reality is

16:59

that it's like less than 5% like of what

17:02

of of what actually gets discussed. So,

17:06

Frontier Labs accelerate despite the

17:07

resignations that happen. Why? Because

17:09

there's a lot of money to be made if

17:10

you're if you're a talented AI engineer,

17:12

and there will be for a while. Um, one

17:15

of the numbers that I saw is that

17:16

there's globally there's about 22,000

17:20

high-end AI engineers out there, which

17:22

is just not enough. So, if you want to

17:24

make bank in the next few years, become

17:26

a crack AI engineer. Um, and and help

17:29

help with the acceleration. That's

17:30

that's one of the biggest bottlenecks

17:32

that any individual can help overcome.

17:34

B2B adoption ignores public sentiment.

17:37

So this is basically like, yeah, you

17:39

guys can whine about AI art and all this

17:41

other stuff and copyright infringement.

17:43

The businesses don't care. It doesn't

17:45

impact them whatsoever. And then US

17:47

federal regulation is largely noise. So

17:49

the EU is slowing things down, which is

17:52

like, okay, Europe, you're you're

17:54

hurting yourself to, I don't know, spite

17:56

America. I don't know. But um but the

17:58

regulation is not really slowing down

18:01

artificial intelligence. Neither is

18:02

safety discussions.

18:04

So however where regulation is the real

18:07

friction. So number one is the USChina

18:10

um uh rivalry. So there's export

18:13

controls and import controls and this is

18:16

if you take a step back and you say

18:18

globally what is the barrier to friction

18:20

and it is the compute gap. Already

18:23

America has like two to three to four

18:25

times or five times the amount of

18:26

compute that China has. But our compute

18:29

is pulling ahead of Chinese compute not

18:31

only in terms of quality like per unit

18:33

quality but overall volume which means

18:36

by 2027 America is expected to have 17

18:39

times the compute of China and that is

18:42

due to uh you know export controls which

18:44

from a geopolitical perspective that

18:46

means America is winning that's our

18:48

moat. Um now you might say globally

18:50

that's not a good thing because that

18:51

slows down Chinese research although as

18:53

I said constraints are the father of

18:55

creativity and the Chinese are very

18:57

creative. they're able to do a lot of

19:00

research with a lot less resources. Now,

19:02

when you look at the center for AI

19:04

security and innovation, which is uh is

19:06

that the one that is the subset of NIST?

19:09

Um I think I think that's the department

19:11

inside of NIST. Their studies basically

19:13

show that the Chinese models are

19:14

inferior across the board. Some of them

19:17

are cheaper, some of them are more

19:18

efficient, but in terms of uh in terms

19:21

of cyber security risk, they're much

19:23

higher. In terms of intelligence,

19:24

they're not as good. Um, so they don't

19:27

really represent a threat uh on a

19:29

military level. Then you've got the EU

19:31

AI Act. So this is the break. This is

19:33

the compliance wall. So the licensing

19:36

cost for any high-risk systems is 52,000

19:39

a year, which means that you're

19:41

basically going to stifle any startup.

19:43

No startups dealing with higher risk AI

19:46

use cases will be in Europe. And this is

19:48

already true. Um, and so where do they

19:50

go? They go elsewhere. They go to

19:52

America. Some of them go to China. Some

19:54

of them go to Saudi Arabia. Basically,

19:57

Europe is very good at ensuring that

19:59

frontier business does not happen in

20:01

Europe, which is kind of silly in my

20:03

opinion. It's like, you know, they're

20:05

they're very very proactive and there's

20:07

a lot of regulatory capture and and and

20:09

what I call a vetocracy. So, basically

20:12

the department of no, that's what Europe

20:14

is. And then finally, liability. This

20:16

was the most interesting finding from my

20:18

my research is that uh many insurance

20:21

policies have absolute AI exclusions.

20:24

Meaning that if you use AI and there is

20:28

um if there's an OSHA violation, if a

20:30

patient dies, then the insurance company

20:33

completely washes their hands. They say

20:35

if AI touched that incident, we are not

20:38

taking any responsibility whatsoever.

20:40

And the reason is because it's new. It's

20:42

high risk. It's high variance. They

20:44

don't know how to price it. I used to

20:45

work at a at a workman's comp insurance

20:47

company. When you have a and this was at

20:50

the very beginning of my career, when

20:52

you have a domain that is well

20:53

understood and the risks can be

20:55

controlled um and and and understood and

20:58

measured, then you then they know how to

21:00

price it. And so like we had a we had a

21:02

division of our of our insurance company

21:05

that would handle like the one-off cases

21:07

where it's like oh hey you're a

21:08

physician who operates out of a yacht or

21:10

something like that like you know cuz

21:12

you have onboard onboard physicians and

21:14

that is not as well understood as a

21:16

physician operating in say a hospital or

21:19

a clinic. So then it's like, okay, well,

21:21

how do you ensure a physician if they're

21:23

working under suboptimal conditions on,

21:26

you know, a really expensive yacht and

21:28

with really expensive clients, how do

21:31

you price that insurance policy? And so

21:33

you can price those things, but it takes

21:35

a lot of work and a lot of effort to do

21:37

a one-off, and that's not how you

21:39

operationalize and how you scale. So

21:42

likewise, when you have you, you know,

21:45

your your average corner store or your

21:47

mom and pop shop or your average

21:49

enterprise that really just wants a a a

21:51

templated insurance policy to say, "Hey,

21:54

you know, we're using AI for this. Make

21:56

sure it covers our customers. Make sure

21:57

it covers our own butts." The insurance

22:00

company says, "We don't know how to

22:01

price that." And so we're not going to

22:03

cover it, and we're going to ensure that

22:05

our policy has an exclusion saying, "We

22:07

will not be on the hook." which and if

22:09

if there's no insurance then the

22:11

companies then the enterprises simply

22:12

won't do it. Um and so this is this is I

22:16

think the most ironic uh kind of barrier

22:19

uh uh legal friction for adopting AI. So

22:21

if I have any insurance people any

22:24

insurance pros in the audience who want

22:25

to like help explain how we could make

22:27

this better, I think that would be great

22:29

because the this this to me is the

22:31

dumbest reason to slow down [laughter]

22:32

AI. Um, I get it, but it's it's like

22:35

really like we're just gonna we're

22:37

because because we don't know how to

22:39

price it. Like moving on. So what this

22:42

is being called this this phase that

22:44

we're in 2026 through 2028 is the

22:46

digestion phase. So the hype cycle was

22:50

2023 to 2025. Now we're catching up with

22:54

reality. So reality says, you know, we

22:56

need grid interconnects, we need

22:57

transformers, we need high bandwidth

22:59

memory, we need the chips and wafers, we

23:02

need verifiable synthetic data, and we

23:04

need to figure out the insurance

23:05

policies. So it's now from bigger models

23:08

and scale is all you need to efficiency

23:11

and distillation and make do with what

23:14

you've got. Just do the best with what

23:15

you've got. And that's the that's the

23:17

paradigm that we're going to be in over

23:18

the next couple years. Now, that's not

23:19

to say that no new capacity is coming

23:22

online. And obviously there's new

23:23

capacity coming online all the time. Um

23:26

but we're not the the the the distance

23:29

between what we would like to install

23:30

and deploy and where we can actually

23:32

deploy is still growing. So then uh

23:35

after 2028 that's when acceleration and

23:38

the the economic pivot will uh continue.

23:42

So the friction map so this is basically

23:44

a recap is the critical and binding

23:47

constraints are power availability grid

23:49

interconnection and HBM supply moderate

23:52

and addressable so that's data quality

23:54

again I'm not concerned about that the

23:56

packaging um that seems like it's going

23:58

to be solved by the market deployment

23:59

friction um deployment friction again

24:02

that it just takes time for enterprises

24:03

to learn how to deploy these things and

24:05

to operationalize it and then liability

24:08

the insurance liability this to me is

24:09

the silliest thing um the things that

24:12

are not uh uh a friction at least for

24:16

America is number one safety and x-risk,

24:18

number two federal regulation, and

24:20

number three capital availability. So

24:22

that's it's a really interesting place

24:24

to be. Um you know there is there is no

24:26

bubble to speak of. Um there are

24:29

frictions, but as I've talked about in

24:31

previous videos, the difference between

24:33

a typical bubble and what we're seeing

24:35

is everything is sold out, meaning we

24:37

still have unmet demand. With a bubble,

24:39

that's pure speculation. uh with with

24:42

this it's like everyone wants more AI

24:44

and AI is not even like at its prime

24:47

time yet. It's not even fully matured

24:49

yet and it and people want more than

24:50

than we can give them. So this is very

24:52

very different from a bubble and that I

24:55

believe is it. So Adams eat arguments to

24:58

accelerate AI stop arguing about

25:00

philosophy and start pouring concrete.

25:02

The future belongs to those who master

25:03

the physical world. grid permits, fab

25:05

capacity, energy generation, focus on

25:08

atoms, and then we have the sources and

25:11

references. So, with all that being

25:12

said, thanks for watching to the end,

25:13

and I'll check you all later. You are

25:15

now aware of why AI is going to go

25:17

slower than we would prefer over the

25:18

next few years, and it has nothing to do

25:20

with the actual research being done. The

25:22

research is continuing a pace. Now, we

25:24

are entering where the the phase where

25:26

the rubber meets the road. So, the the

25:28

friction with reality is the barrier

25:30

now. Cheers.

Interactive Summary

The video discusses the primary bottlenecks hindering the acceleration of AI development, focusing on the period between 2026 and 2028, termed the 'digestion phase'. Key constraints identified are energy availability and grid interconnection, supply chain issues particularly with high-bandwidth memory (HBM) and chip packaging, and operational friction such as data quality, enterprise execution, and return on investment (ROI). While AI safety, security, and public opinion are mentioned, they are largely dismissed as 'noise' or operationalized concerns. Geopolitical competition, especially between the US and China, and market competition are identified as major drivers for AI acceleration. The video highlights the massive investment in AI, comparable to the Manhattan Project, but points out that physical and infrastructural limitations, rather than a lack of capital or research, are the main impediments. Specifically, energy-related issues like grid capacity, transformer lead times, and the long wait for data center interconnection (7 years on average) are critical. Supply chain bottlenecks have shifted from GPUs to HBM and chip packaging, with Nvidia dominating the latter. The video also touches upon the role of synthetic data, the inefficiencies of current AI learning compared to humans, and the reluctance of investors to fund long-term infrastructure projects due to ROI concerns. Insurance liability, with its AI exclusions, is presented as a surprisingly significant friction point. The conclusion emphasizes a shift from focusing on large models to efficiency and leveraging existing resources, with a call to action to focus on physical infrastructure ('pouring concrete') rather than philosophical debates.

Suggested questions

5 ready-made prompts