HomeVideos

The Early Days of Anthropic & How 21 of 22 VCs Rejected It | The Four Bottlenecks in AI | Anj Midha

Now Playing

The Early Days of Anthropic & How 21 of 22 VCs Rejected It | The Four Bottlenecks in AI | Anj Midha

Transcript

2084 segments

0:00

AI alignment, don't get me wrong, is

0:01

hard, but not the hardest problem. Human

0:03

alignment is really the problem right

0:04

now. Our guest today is the most

0:06

prominent AI investor in the ecosystem,

0:09

an Midhar. Why is he the most prominent?

0:11

Three reasons. Number one, he's one of

0:13

the founding investors of Anthropic.

0:15

Number two, he led AI investments for

0:17

Andre Horus where he made investments in

0:20

Black Forest Labs, Mistral, Sesame among

0:23

others. And then third and finally today

0:25

he's the founder of AMP where he

0:27

provides compute and invests in the

0:30

world's best AI companies. If we don't

0:33

secure frontier model inference or what

0:35

I call state-of-the-art inference behind

0:37

a coordinated Iron Dome, I don't think

0:38

we have a sustainable shot at staying at

0:40

the frontier over the next decade.

0:42

There's no saturation in superconductor

0:43

discovery at all.

0:46

>> Ready to go.

0:58

and I am so looking forward to this

1:00

dude. I have stalked the [ __ ] out of you

1:01

for the last three or four days. I spoke

1:03

to Bing Gordon. I had a catch up with

1:05

Bing before this. Very nice to speak to

1:08

him. Uh so thank you so much for joining

1:10

me today, dude.

1:11

>> Thanks for having me. It's too long. It

1:12

only took us what, eight years, nine

1:14

years? I forget when it was.

1:16

>> I was 12 when we last did it. Yeah.

1:19

Well, 12 in startup land is 25, right?

1:23

So,

1:23

>> dude, I'm confused. Help me out. I had

1:25

Damis on the show the other day from

1:27

DeepMind. He was like, "Yeah, I'm not

1:30

sure if we're seeing scaling laws, but

1:32

we are definitely seeing slightly

1:34

diminishing like return/performance as

1:36

we scale." So, potentially, are we

1:38

getting to a stage where increased

1:40

compute is no longer leading to

1:42

increased performance?

1:44

>> Oh, no. Absolutely not. No, that's

1:46

that's not true at all. in in certain

1:48

domains that are well explored like

1:50

coding for example yes there's an

1:52

increasing amount of compute required to

1:54

get an incremental gain in some eval

1:57

that's super saturated but if you said

2:00

what about material science you know I'm

2:02

sitting here at periodic labs office

2:03

this is my incubate like the my latest

2:05

in incubation is called periodic labs I

2:08

spend 3 days a week here in in Mando

2:10

Park we have a 30,000t facility where we

2:12

have LLMs that then predict new

2:15

materials new superc conductors We then

2:18

have robots synthesize those new

2:20

materials and then we have we have

2:23

physical machines like X-ray defraction

2:24

machines validate whether those

2:26

materials have the properties that were

2:28

predicted by the LLMs and then we pipe

2:31

that we we we pipe that verification

2:33

data back into our training run. You

2:35

know how many other times we need and I

2:38

can tell you throwing more compute at

2:39

the problem is probably having yeah

2:42

super exponential gains right now per

2:44

iteration. So it depends on which domain

2:46

you're talking about, which modality.

2:48

There's no saturation in superconductor

2:50

discovery for example at all. The bitter

2:52

lesson is holding is well and alive.

2:55

>> I I totally get that. Can I ask you when

2:58

we look at the bottlenecks around

3:00

performance and progression today? What

3:02

are the bottlenecks that really persist

3:05

most significantly to you? Is it is it

3:07

algorithms? Is it data? Is it compute?

3:09

Can you help me understand which is most

3:11

lagging?

3:12

>> So there's four or five. It's um context

3:16

feedback which I'm happy to talk about.

3:19

It's compute.

3:21

There's capital which you need to like

3:24

you know continuously sort of deploy the

3:26

compute and context feedback loops. And

3:28

then there's culture and I think that

3:29

culture actually might be the most

3:32

important bottleneck of all time. But

3:34

those are the four I would say. Now

3:36

look, algorithmic innovation I think is

3:38

a function of culture basically because

3:40

if you have the right culture, you get

3:42

to attract the best researchers. The

3:45

best researchers, the best research

3:46

talent then wants to work on pushing the

3:48

frontier. And algorithmic innovation

3:51

just falls out of having a really good

3:53

team that's very flexible on what kind

3:55

of architecture they want to use. If you

3:58

have the right culture, the algorithmic

4:00

innovation bottleneck solves itself

4:02

because then the the researchers are not

4:04

focused or like tied to one architecture

4:07

versus another. They're not going, I'm

4:08

all in on LLMs or transformers versus

4:10

diffusion models. The best scientists

4:13

and researchers just want to solve the

4:14

problem, the mission. And if you have a

4:16

very missiondriven culture where they're

4:18

like, we want to move the frontier of

4:19

coding or the frontier of material

4:22

science, the algorithmic

4:24

stuff takes care of itself.

4:27

But so so I'm not that cons that's

4:29

actually not the bottleneck anymore in

4:30

my view. 2 three years ago that was a

4:32

huge bottleneck where we were trying to

4:34

figure out which algorithms scale is

4:36

there are there some limits to the

4:37

transformer architecture versus

4:38

diffusion models. And what I've come to

4:40

realize is if you solve the culture

4:43

problem you can solve the research and

4:44

the algorithmic problem. Then the

4:46

bottlenecks of context feedback which is

4:48

what is the data you need to keep doing

4:51

frontier research over and over again is

4:54

is is step number one. because actually

4:56

I think that is also where you have the

4:57

most business and commercial advantage.

5:00

I think there's lots of alpha and uh

5:03

value to be gained in pre-training,

5:05

mid-training and so on. But you know

5:06

that last mile where you you deploy a

5:10

model or an agent in some new domain and

5:12

then you collect feedback on how it's

5:14

performing in real time and then you

5:16

like I was saying here we do physical

5:18

verification of material science at

5:19

periodic um where wherever there are

5:22

some unique context feedback loops that

5:25

are that are missing today that's where

5:27

you probably have the biggest

5:28

bottlenecks on capabilities. And so what

5:30

you should be doing if you're trying to

5:32

advance the frontiers is going okay you

5:35

know these models suck. For example

5:36

about a year ago as an example I

5:39

realized there was a lot of talk about

5:41

models being good at physical physics

5:44

and chemistry AI for science and I was a

5:47

visiting scientist at the applied

5:49

physics department at Stanford and we

5:50

started benchmarking these models you

5:53

know claude Gemini and so on and

5:56

surprise they sucked. they were so bad.

6:00

I was like, there's there's there's this

6:01

disconnect between the marketing hype of

6:04

AI for science and the reality where

6:06

these models are terrible at the time at

6:07

least. They were starting to get good at

6:09

code, but they were terrible at

6:10

scientific analysis.

6:12

And you know, the conclusion was pretty

6:14

simple. They were just missing a lot of

6:16

the the physics and chemistry data you

6:20

need to reason about the physical world.

6:22

But to do that, we don't have enough of

6:25

that data on the internet because the

6:26

internet is mostly pre-trained data

6:28

about things like blogs and blah blah

6:31

blah and coding. But if you need physics

6:33

and science, that's a real bottleneck

6:36

cuz that data is locked up in national

6:38

labs and academic labs. It's locked up

6:40

in physical uh you know semiconductor

6:42

manufacturing plants. How do you get

6:44

that data in? That was the bottleneck I

6:46

realized was really the the critical

6:48

part of getting these models to reason

6:49

about the physics and science frontier

6:51

which is something I care about deeply.

6:53

And so the way we solved that at

6:54

periodic was you know set up a physical

6:57

lab with robots doing all that. You

6:59

could you could apply that same recipe

7:01

to whatever domain where you want to see

7:03

more and more progress. Then you ask

7:05

okay how much comput and infrastructure

7:07

do you need to keep that RL loop or the

7:09

physical verification loop scaling at

7:11

bigger and bigger scale. And then you

7:13

need the capital to fund all this. You

7:14

need equity, debt, a whole bunch of

7:17

different structured finance vehicles to

7:18

get, you know, land, power, shell. So

7:21

that's the compute bottleneck. And then

7:23

lastly is the culture. Cuz if you have

7:24

all of those three things, but you don't

7:26

have the right team and the right

7:27

missiondriven culture, the whole thing

7:28

falls apart. And and so those in my mind

7:30

are the four bottlenecks I wake up, you

7:32

know, every day trying to figure out how

7:34

we we unblock for the best teams. If we

7:37

just go through them, when we look at

7:38

that context feedback on the data side,

7:40

will we see then a generation of

7:42

vertically integrated foundation model

7:44

companies like periodic for a ton of

7:46

different things? Yeah.

7:48

>> Yeah. You know, when I went to grad

7:50

school uh for machine learning, I I I

7:52

went to Stanford for bioinformatics,

7:54

which was machine learning applied to

7:55

healthcare. We were the space was not as

7:58

good as marketing as it is today. So

8:01

super intelligence, love it. You know,

8:04

at the end of the day, what are we

8:04

talking about? We're talking about very

8:06

powerful models within some domain and

8:08

and we are seeing though sort of within

8:10

distribution very very powerful

8:13

capabilities that are you could

8:15

definitely call them superhuman because

8:16

there's no way for example I as a an

8:19

individual scientist could analyze the

8:20

reams and reams of data coming out of

8:22

the lab here without AI models there's

8:24

just no chance and so the fact that you

8:27

can take all of the data from you know

8:29

training from from a a physical lab and

8:32

just throw it at a bunch of AI models

8:35

and ask it to analyze things is a

8:36

superhuman capability. We didn't have

8:38

that before. Okay, fine. So, let's call

8:40

that super intelligence. Within coding,

8:42

within material science, within each of

8:44

these domain distributions, we are

8:46

seeing capabilities that are super

8:48

human. We didn't have them before. And

8:50

and and in fact, I would say we're even

8:52

starting to see automation of those

8:54

tasks, especially where there's there's

8:57

coding involved to starting to be

9:00

somewhat recursive, right? where if you

9:02

have a good coding model then you can

9:04

say okay let me automate like data

9:06

analysis let me automate like data

9:09

cleaning and so on some people would

9:12

call that recursive self-improvement

9:13

totally happening but it's it's not like

9:17

I can just say to a coding model please

9:21

bootstrap a a physical R&D lab for me in

9:24

Menllo Park get all the permitting go

9:27

you know go find an to raise money from

9:30

go set up the physical infrastru

9:31

structure and just like bootstrap all

9:33

this data. That's just an entirely

9:34

different kind of frontier and execution

9:37

and sort of problem.

9:39

>> My question to you then is like how do I

9:41

determine what is not going to get

9:43

claudified in that vertical model

9:47

company buildout because you could look

9:48

at a cursor and say well they've built

9:51

their own vertical model end to end and

9:54

it's been claified if we're being blunt.

9:57

periodic won't be because of the

10:00

physical data that's being produced in

10:01

the labs. How do I know what will be

10:03

cladified versus won't in that model

10:05

there?

10:06

>> Yeah, this is a good question. Okay, if

10:09

we want to sort of unlock frontier

10:12

progress generally across a bunch of

10:14

domains, then where are the bottlenecks

10:17

and where will the value acrew? Context

10:20

is

10:22

is not necessarily the moat. I would not

10:24

say yet. I I think I think venture

10:27

capitalists are very quick to analyze

10:29

modes but I would say context feedback

10:31

loops where you have you have unique and

10:32

differentiated access is where progress

10:36

will be most legible to you and if there

10:40

are other teams who don't have access to

10:41

that context it'll also be where you

10:44

have a superior business model and so

10:45

here's an example I give in the class

10:47

right sovereign data are you familiar

10:50

with the cloud act

10:52

>> yeah okay so the you know the the US

10:54

cloud act says that hey if there's

10:57

mission if if there's any data workloads

11:00

infrastru cloud workloads running on

11:01

infrastructure that is managed by an

11:03

American company then the US government

11:05

has to be able to access that data now

11:08

if you happen to be running military

11:11

defense mission critical workloads in

11:13

Europe on AI infrastructure that is

11:16

managed by an American company well that

11:19

context which is super critical can't be

11:23

sent over across the border

11:25

That's an example of a unique and

11:27

sensitive context that needs to be run

11:29

locally. And so if your ASML, your um

11:34

CMACGM that's doing logistics at scale

11:37

and some of this logistics is with

11:39

missionritical supplies, you can't have

11:42

your supply chain data being processed

11:44

by an AI bot that's running on servers

11:47

that is subject to the cloud act. So

11:50

what do you do? You look for local

11:51

infrastructure partners. you start

11:53

going, hey, who are the providers, AI

11:55

infrastructure providers in Europe that

11:57

we trust? Well, it turns out there

11:59

aren't that many who can actually handle

12:01

mission critical infrastructure at scale

12:02

for AI. So you call up someone called

12:04

Arthur Mench who is a French scientist

12:08

from DeepMind turned entrepreneur and

12:11

starts a lab called Mistral who is

12:13

running massive workloads and you say

12:14

Arthur would you actually build

12:18

infrastructure that can be secure

12:20

locally and that's why suddenly in July

12:23

of 202

12:27

at the at Vivate in Paris you have

12:29

President Mccron and Jensen standing on

12:32

stage next to Arthur, a 33-year-old

12:34

scientist unveiling a gigawatt AI

12:37

infrastructure facility in Paris. Why?

12:40

Because the context, the mission

12:42

critical context of those workloads is

12:44

so important to be run locally that you

12:47

can't run them on Amazon AWS, GCP or

12:50

Azure. And it's the first time in 15

12:54

years that the that the sort of

12:56

hyperscaler dominance is um up for grabs

13:00

for startups. With the greatest of

13:02

respect, is that the core investment

13:03

thesis of Mistral for you?

13:06

>> For me, yeah. Independence at scale of

13:09

at every part of the AI infrastructure

13:10

stack like land PowerShell in Europe,

13:13

that's sovereign, it's local, compute

13:16

infrastructure, that's local. And models

13:18

that are trained locally, by the way,

13:20

fully open, so they can be deployed and

13:22

customized globally wherever needed. But

13:24

certainly in Europe, like the full

13:27

independent stack is is the is the bet.

13:30

Yeah. Do

13:31

>> Anthropic and Open AI just accept that

13:33

and roll over? I I don't understand

13:35

because government is a mega portion of

13:37

their efforts and workload today and

13:39

like both of them when I speak to them

13:41

are like, "Oh, we're absolutely coming

13:42

for Europe."

13:44

So, so how do they get around that?

13:46

>> Well, I can't speak for OpenAI too much

13:48

uh cuz I'm not involved there directly,

13:50

but Anthropic, I will say, you know, the

13:53

mission and vision has always been very

13:56

um I think it's always been very

13:58

American aligned, right? They've always

14:00

said, "Hey, America is the crown jewel

14:03

of the world in terms of innovation.

14:06

This is where we're located." Anthropic

14:07

is located in Silicon Valley. Um, and I

14:09

think the company really, really wants

14:11

to do what's best for the American

14:12

government and the American way of life,

14:14

which is democracy and freedom. It turns

14:16

out the world's largest enterprise

14:18

customers are governments and Fortune

14:20

500 companies. And many of those that

14:23

are overseas need these workloads to be

14:26

running locally. you said about

14:27

obviously being involved with anthropics

14:29

since the earliest of days. I'm just

14:30

fascinated.

14:32

I think people kind of forget about

14:34

their early days almost. People talk

14:37

about like, oh, SPF investing early and

14:39

what a visionary he was,

14:40

>> right?

14:41

>> What was what was Anthropic and Dario

14:43

like in the early days?

14:44

>> Well, so I've known Tom forever. Uh Tom,

14:47

you know, was one of the the lead

14:48

authors on GP3.

14:51

Um we've been friends for many we'd been

14:54

friends for many years. Tom gave me a

14:56

call and said, "An you know, we for

14:58

various reasons, we want to leave and

15:00

start this new lab called Enthropic.

15:01

We're going to need uh a lot of capital.

15:04

We're going to need compute." I I had

15:06

already sold Ubiquity 6 at that point.

15:07

So, I'd kind of gone through the founder

15:09

journey. Um and so Dario, Tom and I

15:11

started doing these weekly sessions in

15:12

early 2021 to try to figure out how to

15:15

turn what was really a research

15:17

hypothesis, right, which is scale the

15:18

scaling recipe into a business

15:21

hypothesis. Um, and look, I would say it

15:24

it took like really 12 to 24 months. Um,

15:27

and they did a lot of the hard work on

15:29

figuring out how how do we really sort

15:31

of operationalize this the idea of this

15:34

AI pair programmer, right? where you

15:36

take the context feedback loop of the

15:38

local repository, the files, the

15:41

directories of programming and kind of

15:46

sort of in a in a very methodical way

15:48

make predictable progress on the

15:51

capabilities of um of of software

15:55

engineering.

15:56

And I I thought it was a very you know

15:59

if if anything my biggest flaw is as an

16:02

investor as a founder is being too early

16:05

to things. That that was my lesson with

16:07

ubiquity 6. I was early to the whole

16:09

computer vision which is now you know

16:11

obviously blowing up the whole

16:12

multimodal sort of generative modeling

16:14

space. Um and since then I have I think

16:16

updated my strategy on how to get timing

16:18

right. But at the time you know our our

16:21

the recipe was pretty simple right?

16:23

raise some money, buy some compute, get

16:26

a little bit of context data on

16:28

programming, put out a basic version of

16:30

the model, deploy it with with teams

16:33

that we trust who are doing coding, and

16:35

then pipe that feedback loop back into

16:37

the training run over and over again.

16:40

And when you do that with inference,

16:41

inference gives you sort of two things,

16:43

right? It gives you revenue to buy more

16:45

compute, and it gives you the context

16:46

feedback to keep improving the

16:48

capabilities curve. And I was like,

16:50

great, this makes total sense, guys.

16:51

let's go raise money. I invested a bunch

16:53

of my money uh that was just life

16:55

savings which was not much given I was a

16:57

poor founder at the time which where

16:59

most of my net worth was tied up in

17:00

Discord stock and it and and it pains me

17:02

sometimes to to look back at the emails

17:06

of friends. So I introduced them to 22

17:08

you know friends up and down Sand Hill

17:10

Road and so there's some investors there

17:12

and we got 21 nos, right? And I was like

17:15

what what are you guys thinking? And

17:17

they said, "Well, an this this recipe

17:19

sounds good in theory, but like where's

17:21

the proof?" And I said, "Proof? The

17:23

these are the guys who invented GPT3.

17:25

How much more proof do you want?" And

17:27

they said, "What's GPT3?" I was like,

17:29

"Oh my god." Like, how do you go about

17:32

educating somebody who doesn't even

17:34

understand the technology and the

17:36

breakthroughs that are happening in the

17:38

machine learning community? Now, I was

17:39

lucky cuz I I had that training from

17:40

grad school. I'd started a computer

17:42

vision company. So, something that was

17:44

super legible to me just was a

17:46

completely different world. And then for

17:47

those investors, we were pitching,

17:49

remember we we originally tried to go

17:50

out and raise 500 million and then had

17:52

to reanchor to only raising a hund00

17:53

million seed round, which at the time

17:55

felt like a lot, but of course was tiny

17:57

compared to how much OpenAI had raised,

17:58

cuz by then I think Opened a billion

18:00

dollars. And so the whole idea of

18:02

compute multipliers where we could for

18:05

every dollar of venture capital raised

18:08

produce a unit of of intelligence for

18:11

six times less was not like the VCs did

18:15

not understand it which is why you know

18:17

over the next 24 months the people who

18:19

got it were either people like you know

18:21

some of the folks in the ML community

18:23

who also had an overlap with the

18:24

effective altruist community like SPF

18:27

but also Amazon right this was very

18:29

legible to Amazon on because they were

18:31

watching what was happening with Azure

18:32

and OpenAI and they were like, well,

18:35

this is super aligned. If you guys

18:36

actually can create a bunch of

18:38

state-of-the-art models that are hosted

18:40

on Amazon, that's super accretive to to

18:44

the AWS business. And that's why, you

18:46

know, it resulted in deep compute and

18:48

capital for equity partnership with

18:50

Amazon that was originally $4 billion.

18:52

You know, a lot of this is public now,

18:54

but at the time it was it it was a

18:57

really tough journey. And I would give

18:59

Daario, Tom, the other co-founders, you

19:03

know, Daniela,

19:04

Jack, Sam Mcandlish,

19:07

um

19:10

like it Jared, Jared Kaplan, they were

19:13

it was such a brutal time getting this

19:15

company going. like

19:18

people don't is there anything you would

19:20

have advised them differently knowing

19:22

all that you know now

19:23

>> I'm not sure I would because the world

19:26

is a very different place today you know

19:29

and at the time it really did feel like

19:33

there was no one they could trust

19:35

>> is it not impossible not to be hauled up

19:37

in front of Congress if you reach a

19:39

certain scale

19:40

>> whether whether you're Google or whether

19:42

you're Facebook or whether you're

19:44

anthropic fighting against the pent

19:45

Pentagon it you get to a scale where it

19:50

is impossible not to have that conflict.

19:53

>> Oh absolutely. No. What are you talking

19:54

about? Look, I started AMP as a public

19:56

benefit corporation cuz I I think it's

19:57

actually a very aligned model. Have you

19:59

heard of REI, right? REI is a public

20:01

benefit corporation. They make billions

20:03

of dollars in revenue and profit. Have

20:04

they ever been held up in front of

20:06

Congress? No. Like Ben and Jerry's

20:08

public benefit corporation. Have they

20:10

been, you know, hauled up in front of

20:11

Cong? No. It's because they

20:13

self-modderated

20:14

right at a time and they said here's our

20:16

mission but we have to make we have to

20:18

build a business and as long as you hold

20:21

those two things in sort of those things

20:23

are not in conflict long term. If your

20:25

goal in life is long-term to push

20:28

humanity forward in some stable reliable

20:30

way, then you all there are always

20:32

tensions where you have your mission and

20:35

then you have your profit motive. And

20:36

you've got to be able to to like

20:38

moderate between those two. And I think

20:41

public benefit governance allows you to

20:43

do that. And I think we need more public

20:45

benefit charters in Silicon Valley and

20:46

in technology. And I think we will get

20:48

there. If you look at the arc of

20:50

infrastructure businesses, for example,

20:52

right? I I actually I actually had a

20:54

chat with a mutual friend of ours who

20:56

asked not to be revealed.

20:58

>> Okay.

20:59

>> Um and they said, "For [ __ ] sake, all

21:02

these PBC's, public benefit

21:04

corporations, will these startup

21:06

founders not just [ __ ] win their

21:07

market first?"

21:11

I mean, how are they feeling? Are they

21:12

investors in anthropic?

21:15

>> No.

21:16

>> Okay. So tell them to give me a call

21:18

when they'd like to be investors in the

21:20

world's fastest growing business of all

21:22

time. And then they can lecture me about

21:24

public benefit governance and market

21:26

share adoptions. Public benefit

21:27

governance gives the leadership the

21:29

ability to make decisions that sometimes

21:31

are not legible to shareholders as best

21:34

for them. What decision?

21:37

>> What decision can you foresee with AMP

21:40

that is aligned to your mission but does

21:43

not put the profit motive incentive

21:46

first? There are many up and down the

21:48

stack because we see ourselves as a full

21:50

stack scaling partner to the best

21:52

frontier technology teams and we also

21:54

kind of see ourselves a little bit as

21:55

have our job is to propose independent

21:58

standards for AI and as an institution

22:01

try to uh evangelize the adoption of

22:03

those standards through you know profit

22:05

generating businesses. We have a venture

22:07

capital business. We also have an

22:08

infrastructure business and a good

22:09

example of this for now is we're

22:12

actually giving away most of our compute

22:13

at cost. Now, if you're a shareholder,

22:16

you'd go, "Wait an billions of dollars

22:19

of compute infrastructure you're giving

22:21

away at cost."

22:23

Yes, because we think that's the right

22:25

thing for humanity. And we think that's

22:27

the right way long-term to have a

22:29

healthy independent ecosystem, which is

22:31

what our mission is. Our mission says

22:33

AMP is a public benefit holdings

22:34

company. Our our vision is is to ensure

22:38

there's a healthy independent frontier

22:39

technology ecosystem. Our mission is to

22:42

maximize the world's frontier output. to

22:44

do that long term. We think the teams

22:46

that are truly doing innovation like

22:48

truly doing pushing the frontier of

22:50

science and engineering need act compute

22:53

access and many of those teams today

22:55

can't afford to pay price gouging the

22:58

the the in extraordinary prices for

23:01

comput infrastructure today and so you

23:03

know what yeah we're happy to provide

23:04

them access of that in a way that's

23:05

mission aligned

23:06

>> an how do you secure the compute supply

23:09

maybe I should know this but it's the

23:10

most starved resource today how do you

23:13

secure a resource that no one else can

23:15

seemingly secure.

23:17

>> Well, step one is you get there first

23:19

before people realize how how valuable

23:21

it is. And uh luck, you know, I've been

23:25

um beating the the drum beat on this for

23:28

4 years now, right? I when I got to E16Z

23:32

as a general partner, the first thing I

23:33

did is I sat down with Mark and Ben and

23:34

said, "We need more compute. We need

23:36

compute access for these incubations I'm

23:37

going to do." And they said, "No

23:39

problem, An. Let's set up a program.

23:40

What do you So we used you know our

23:42

balance sheet to start procuring compute

23:44

through the oxygen program. That gave me

23:46

the ability to build pretty deep

23:49

relationships with the industry and

23:52

build trust with compute partners who

23:55

now we have lots and lots of

23:57

relationships with that we're scaling um

24:00

in ways that would be very hard if I

24:02

didn't have that time and the uh sort of

24:07

flexibility to understand that what is

24:09

required to really get that

24:11

infrastructure right. You know, we've

24:12

talked a little bit publicly about what

24:13

we're building, which is the AMP grid,

24:15

which is essentially a a a what what the

24:18

electricity grid did for electricity,

24:20

we're trying to do for compute

24:21

infrastructure. We see ourselves as an

24:23

independent system operator of the grid.

24:24

We we're not a cloud provider. We don't

24:26

own our own data centers. Uh we're not a

24:28

traditional venture capital firm either.

24:30

We see ourselves as an independent

24:31

system operator, which means our job is

24:33

to coordinate capacity across the

24:34

ecosystem in a way that allows the best

24:36

teams, the best independent teams to

24:38

provision for their base load, not their

24:40

peak. So they don't have to

24:41

overprovision but when they want to be

24:43

able to spike up and down for training

24:45

runs for inference needs they they feel

24:48

secure that the capacity exists. We are

24:51

roughly in 1885 industrial you know

24:54

revolution England right now where you

24:56

have all you know these these frontier

24:57

labs are like factories that the steam

24:59

engine has been discovered. You can use

25:02

steam to produce all kinds of new

25:03

products and many of them are running

25:05

their own generators in their backyards

25:07

at half capacity. And I'm going, this

25:10

makes no sense. Let's all pull our

25:12

generators so that a shoe factory can

25:14

spike up during the day, a steel factory

25:16

can spike up during the night, and then

25:17

you maximize utilization um and

25:20

ultimately output. When you think about

25:22

allocating it, are you not using compute

25:25

and the cost of compute as a loss leader

25:27

for your venture fund business which

25:29

then comes in and says okay you name any

25:33

of your incredible businesses that you

25:34

own whether it's your anthropics or your

25:35

MLS or your Black Forest Labs and say

25:38

okay you'll get the compute at cost but

25:42

for that we need $300 million invested

25:46

and that's your way of winning. That's

25:48

that's not at all how we make the th

25:51

those are not that's not the deal. The

25:54

deal is

25:54

>> okay

25:55

>> we the deal is I incubate new companies

25:58

like periodic labs one at a time. That's

26:01

I can only do this one at a time because

26:02

I I like to team up with scientists or

26:04

engineers who at the forefront of their

26:06

field.

26:07

It takes a lot of work to create these

26:09

new companies from scratch. You know it

26:12

in many ways I had the privilege to to

26:15

realize that we are entering a back to

26:17

the future era of venture capital. If if

26:19

you think about the birth of modern

26:22

industries,

26:24

you know, let's talk about

26:25

semiconductors,

26:27

uh, gene editing, you know, the biotech

26:29

industry or, uh, self-driving cars,

26:33

Silicon Valley in the early days of the

26:35

founding of what I call these frontier

26:37

industries. The way you start the most

26:39

iconic companies is very different from

26:41

how fun companies were funded for the

26:43

last 10 years in the ZER era. Intel for

26:46

example, right, was a very close

26:48

partnership between a couple of

26:49

scientists and a investor called Arthur

26:51

Rock who was a founding investor and was

26:53

at the office every day. Arthur

26:55

literally used to Arthur wrote the stock

26:57

incentive plan. He used to run all hands

27:00

at the company every week. If you look

27:02

at Jenn which was incubated in the

27:04

basement of Kleiner right Bob um it was

27:07

co-founders were Herb Boyer professor at

27:10

UCSF and Bob Swanson who was an

27:11

associate at Kleiner and I I got to

27:13

apprentice in that mode of venture

27:15

capital because when I got to Kleiner

27:16

you know I was 20 I was wrapping up grad

27:19

school at at Stanford med school but I

27:21

was working nights and weekends um at

27:23

Kleiner on the investing team and Brooke

27:26

Buyers who was the KPCB&B had an office

27:28

next to me and he had some free time so

27:30

I would go to him and be like Brooke

27:32

you know, teach me your ways. And he

27:33

regailed me with all the stories of how

27:34

Genentech was being founded. And I was

27:35

like, wait. So, you're saying basically

27:38

Bob like co-founded Genentech here in

27:40

the basement at Kleiner. He's like,

27:41

yeah, we were that's what it meant to be

27:43

a partner. And I said, well, that's not

27:46

what happens here anymore. Like we write

27:47

a bunch of checks to SAS companies and

27:49

then they go off and do stuff. And he

27:51

was like different times. And if you

27:54

look at that,

27:55

>> are they mutually exclusive? And what I

27:58

mean by that is can you have a venture

28:00

ecosystem where you have a bunch of

28:02

people writing a bunch of checks as we

28:04

have done for the last 10 years and a

28:07

next generation or to your point a back

28:09

to the future era of venture capital

28:11

where you co-ound the business side by

28:12

side. Can they run side by side or are

28:15

we actually entering an era where we're

28:17

back to the future era as you say where

28:19

value acrruel is in the co-founding and

28:21

incubation side?

28:23

Um, I I think it's very hard for them to

28:25

coexist inside of one person.

28:29

And it's very hard to coexist sometimes

28:30

inside of even one firm because, you

28:32

know, there's a reason I'm sitting here

28:33

at Periodic Labs. I work here 3 days a

28:35

week. Every day from 8:00 a.m. to 8:30

28:38

a.m. for the last year, Liam Do and I

28:41

have had a standup every morning where

28:42

we go through the priorities of the

28:43

company and then we we make them, we

28:46

prioritize, we go and execute. I mean

28:49

the compute team at of AMP is sitting

28:51

upstairs procuring compute for for the

28:53

periodic guys. I my role models have

28:55

always been the Arthur rocks and the Bob

28:58

Swanson's and the Mike Mike Mara

29:00

personal computing effectively the first

29:02

CEO for the first year of Apple was Mike

29:04

Markel. He was an angel investor and he

29:06

was the one doing all the capex, you

29:08

know, supply chain and capital and all

29:10

of that stuff that allowed Steve and and

29:13

jobs and was to focus on the product and

29:15

the engineering and and that kind of

29:17

deep partnership is what I get really

29:19

excited about.

29:19

>> Can I go back to something you said

29:21

before which is like we're at the

29:22

industrial revolution stage and I was

29:25

like, okay, help me understand that. If

29:27

we're at the industrial revolution

29:29

stage, what does that mean for where

29:31

we're going and how I should be acting

29:33

as an investor today?

29:35

>> You have to hold two things in conflict

29:38

that can seem paradoxical. Um, and this

29:41

is this is the most important thing I

29:42

learned from Mark and Ben, which is when

29:45

the future the future is not uh is is

29:49

not determined. And so anyone who tells

29:52

you that they can predict the future

29:53

with certainty should be taken with a

29:54

healthy dose of

29:57

uh suspicion and and instead I try to

30:00

approach things like a scientist and go

30:01

what are the biggest bottlenecks let's

30:03

come up with a hypothesis on how these

30:04

bottlenecks will be solved and let let's

30:06

run multiple experiments in parallel and

30:07

then whichever one emerges you just have

30:09

to be very truth seeeking and and be

30:11

willing to claim like say you're wrong

30:14

right and and and I would say as an

30:16

investor your job is to come up with a

30:18

hypothesis for where the future is

30:20

and be willing to to to make multiple

30:23

different experiments that are aligned

30:25

with your mission in parallel and be

30:27

willing to be wrong and be honest with

30:28

your LPs that some of them may be wrong

30:30

honest. What do you what do you say to a

30:32

Brian Singerman of the world who always

30:33

said that I'm not smart enough to

30:35

predict the future but I my job is to

30:37

pick founders that are able to do so.

30:39

>> I think that the most the safest way to

30:41

predict the future is to invent it

30:44

right. So do the hard work. come up with

30:46

your point of view on if we're in

30:48

industrial revolution England, what

30:50

happened next and what were the emerging

30:53

properties of the businesses that became

30:55

valuable in institutions over the next

30:57

50 years after 1885 and then figure out

31:00

which part of that world which figure

31:03

from history of that era do you do you

31:05

look up to the most and what were you

31:08

know go read about their lives and the

31:10

businesses they ran and the and the

31:12

tensions that emerged in the practice of

31:13

their business later in life cuz then

31:15

they made mistakes when they were young

31:16

and try to learn from their mistakes and

31:18

then and then go and execute.

31:20

>> What's a parallel property direction

31:23

from 1885 onwards style time frame that

31:26

you think will play out in the next era?

31:28

>> Well, obviously in the world of

31:29

infrastructure, I think we need

31:31

something like the grid for in the

31:33

computer infrastructure. So that's what

31:35

I've spent most of my days on which is a

31:37

coordinating mechanism for uh that

31:40

allowed this the commod not the

31:42

commoditization necessarily but the

31:43

transition of uh coal and electricity

31:47

from being these resources that were

31:49

being hoarded to being stable reliable

31:52

uh commodities that that the best

31:55

engineering teams the best factories had

31:57

access to. Right? That so that's that's

31:59

what I think about a lot. I think if

32:01

you're since since you're so talented at

32:03

media and you're so talented at

32:06

storytelling um I think I would and and

32:09

your mission is to push the European

32:11

continent. I think one of the things if

32:13

I was you is I would be talk trying to

32:15

figure out how do we educate

32:18

the leading capital allocators and

32:19

infrastructure allocators in Europe

32:22

about the coming era whether that's

32:25

through media whether that's through

32:27

educational programs and get them to

32:29

understand their role in unblocking the

32:31

bottlenecks for the best scientists and

32:32

engineers in Europe

32:34

>> it's largely a lack of pension fund

32:36

reform in a lot of cases to be quite

32:38

honest

32:38

>> okay so spend your time on pension fund

32:40

reform

32:40

>> how much more

32:42

do we need in Europe for Frontier AI to

32:46

be what we think it can be? Is it like

32:49

2x? Is it 10x?

32:52

That's a good question. I I would try to

32:54

go about it from a top downs approach

32:56

and bottoms up sizing approach.

32:59

Um you know for us at AMP when I look at

33:03

the grid we are building out which is

33:05

sort of a reasoning by analogy. uh we

33:08

have started securing about 1.3 gawatts

33:10

of compute infrastructure that's roughly

33:12

$40 billion of cloud spend over the next

33:14

four years and that is financed roughly

33:17

you know between with about 20% of

33:20

equity the remaining is debt so 20%

33:23

that's about $10 billion of equity

33:25

capital the remaining is all debt

33:26

capital we have a bunch of partners that

33:28

help us put together these equity and

33:31

debt packages to secure computer

33:33

infrastructure for our companies I would

33:35

say in Europe

33:37

I would talk to Arthur and figure out

33:39

how much he thinks is required for the

33:41

independent ecosystem over there. But in

33:43

multiples of gigawatt like if if you're

33:45

doing sort of your atomic unit of math

33:46

in gigawatts I would from a from a top

33:50

down perspective

33:52

you know I think Google is roughly at 12

33:56

to 15 gawatt of that I'm aware aware of

33:59

of infrastructure for internal and

34:01

external deployed needs. Now they have a

34:03

huge land power shell pipeline coming

34:05

but you know I if Europe does not have

34:07

access to Google level infrastructure

34:09

then what are you guys even doing right

34:12

like that's roughly what the continent

34:13

needs for full sovereignty right to have

34:16

as at least as much infrastructure

34:17

locally as there is within the alphabet

34:20

holdings sort of pool

34:24

over the next four years

34:26

>> is the what's easier the equity raise or

34:28

the debt raise

34:29

>> I would say the biggest challenge has in

34:33

figuring out the right aligned financial

34:36

structure

34:38

across both in a way that's legible to

34:42

capital allocators at scale. Took me

34:44

about a year to really get all the

34:46

pieces right. But there are very large

34:48

equity pools.

34:50

Let me put this. a lot of balance

34:51

sheets, long-term missional aligned

34:53

balance sheets in the world who don't

34:54

who have um who are missional aligned at

34:58

wanting to help frontier scientists, re

35:00

researchers, university labs get access

35:02

to the comput they want, but they don't

35:04

have operex.

35:06

They don't have cash to spend on the

35:08

compute. So, if you can find a way to

35:11

align equity um debt, balance sheets in

35:14

a way that's risk sort of um derisked,

35:18

the fundraising is not a problem. It's

35:20

it's actually a systems design problem

35:22

which it took me again a year it

35:24

probably took me four years to get right

35:25

but now that we figured it out it's it

35:28

it's not been a problem.

35:30

>> Do you think we are underinvested still

35:32

today in data centers?

35:35

>> We are deeply underinvested in security

35:38

in secure compute. Okay let me put this.

35:41

We are not in an AI crisis. We are not

35:43

in an AI bubble for sure. I'll tell you

35:44

that which is the the the question I

35:47

keep getting asked. We are definitely in

35:49

a GPU wastage bubble where there are

35:52

stranded pockets of compute like

35:54

billions of dollars of compute that are

35:56

sitting unutilized and if we could pull

35:59

them together on a grid across the

36:00

independent ecosystem.

36:01

>> Why are they unutilized? Sorry.

36:04

>> For a couple of different reasons. Um

36:06

one is they're comput is not fungeible.

36:09

So unlike electricity which had to go

36:11

through a process of standardization you

36:14

know AC/DC where megawatts or megawatts

36:16

are megawatts computer is not funible

36:19

today. So for forget fungeibility of

36:22

compute across different manufacturers

36:24

like Nvidia and AMD within a

36:26

manufacturer

36:27

Nvidia chips for example the H100s the

36:30

GB200s the GB300s these are all

36:32

completely different chip types. So if

36:33

you have one cluster where you're doing

36:35

a training run on H100s and then you

36:37

want to sort of do continued post

36:40

training of that or or or have that do a

36:43

distributed training run of that um

36:45

training uh workload on GB200's

36:50

doesn't work. So they're just like

36:53

stranded pools of compute cuz flops are

36:55

the atomic unit of computation is flops.

36:57

I wish flops were fungeible but not all

37:00

flops are born equal today. And so if

37:02

you provisioned a cluster 2 3 years ago

37:04

with H100s and now you want to you

37:08

actually want to run some of those

37:09

workloads on for the newer generation

37:12

models, you're memory bound by H100

37:14

chips, you can't unlock, you know, the

37:16

the the benefits of the Blackwell chip

37:18

without basically just like buying a new

37:20

cluster. And so now suddenly you have

37:21

this H100 cluster

37:24

that you don't want to do training on

37:25

anymore because it's it's old school. it

37:28

doesn't like the chip doesn't have the

37:29

right memory memory properties to train

37:32

your frontier models and so and it's

37:34

very hard for any individual company to

37:36

h like see all of this stuff but when

37:38

you're on seven or eight boards like I

37:40

am and you've been doing this you know

37:42

15 years and you start to see patterns

37:44

emerge you're going wait a minute why is

37:46

there all this unutilized compute

37:47

sitting here and there

37:48

>> this is lof

37:50

are frontier models moving faster than

37:53

the pace of uh chips as you said that

37:56

with H100s where you you have newer and

37:59

newer models and then you're training

38:00

them on older and older chips because

38:02

that's what's free and it's not moving

38:04

in lock step. Is that is that the

38:06

problem that we're articulating?

38:08

>> No, no, no. The problem we're

38:09

articulating is that compute is not

38:11

funible. There are no standards for

38:12

fungibility and there are no

38:13

institutions enforcing standardization

38:16

of compute enough. So, we are in the

38:18

pre-standardization

38:19

era of compute today, which which was

38:22

the pre-standardization era of

38:23

electricity in 1885. And the next I I

38:26

hope we can we can self-regulate,

38:29

self-standardize

38:30

and self um enforce standardization so

38:34

that we can skip the boom and bust

38:36

cycles that happen with electricity over

38:37

the next 50 years. And this happens with

38:38

every infrastructure cycle in the

38:39

pre-standardization era. It happened

38:42

with electricity in 1885. It happened

38:44

with steel. It happened with railroads.

38:46

And every time you have this boom and

38:47

bust cycle, what happens is wars are

38:50

fought.

38:52

Companies backstab each other.

38:55

It's super painful. It's annoying. And

38:57

my view is that compute not being

38:59

funible is what's resulting in the all

39:02

this talk about AI, the AI bubble. But

39:04

what people forget is that we don't have

39:06

a AI capabilities bubble. The

39:09

capabilities are extraordinary in every

39:10

domain. We have an infra infrastructure

39:13

wastage crisis right now. And it's

39:14

because there are no open standards.

39:16

There's no open protocol for how flops

39:18

from one um data center can flow to

39:21

somebody else who needs it across chip

39:23

types across secure boundaries and uh

39:26

it's resulting in a lot of pain for the

39:27

ecosystem. People are just

39:30

>> if we have compute standardization in

39:32

the way that you said will we remove the

39:34

boom and bust cycle or is that just one

39:36

part of it? I think that will go a long

39:39

way in in preventing this and instead

39:43

just allowing this.

39:44

>> I'm sorry for asking. So, you're like,

39:46

"Jesus Christ, Harry, I'm a professor at

39:48

Stanford and you waste my time with

39:50

this." Which is a fair statement. Um,

39:53

British accent goes a long way though.

39:55

Um, what is the biggest bottleneck or

39:57

barrier to compute standardization that

39:59

you want to achieve?

40:01

>> Uh, it all goes back to alignment, man.

40:04

Misaligned incentives up and down the

40:06

stack. How is Silicon Valley and DC not

40:08

on the same alignment?

40:10

>> For one, I don't think we have

40:12

standardized on whether AI should be

40:16

regulated,

40:17

treated, procured as just as good

40:20

old-fashioned software or like a new

40:24

kind of system, you know, like I again I

40:26

went to grad school for machine learning

40:27

and what you learn in machine learning

40:29

101 is

40:31

models are statistical.

40:33

They're not deterministic, right? So

40:35

when you have a statistical system, it's

40:37

different from there are some properties

40:39

of a statistical system that are

40:41

different from a spreadsheet. A

40:43

spreadsheet is deterministic software

40:45

and a statistical model today is not.

40:47

And so should the procurement of a

40:50

spreadsheet be the same from an IT

40:53

perspective as a statistical model? Open

40:57

debate. That is the core debate. That's

40:59

the problem. Like AI alignment, don't

41:02

get me wrong, is hard but not the

41:04

hardest problem. Human misalignment,

41:06

human alignment is really is really the

41:07

problem right now we have in in the

41:09

world. We need technologists who are who

41:13

understand the difference between

41:15

deterministic software and statistical

41:17

systems to propose a set of standards

41:20

for how procurement for this should

41:22

work. And then we need standards people

41:24

in DC. We have this thing called NIST.

41:28

We have various bodies in the government

41:29

that should get together and say, "Thank

41:31

you guys for proposing this standard.

41:33

This is where it makes sense. This is

41:35

where it doesn't." This is called an RFC

41:37

process.

41:38

And we're going to standardize on this

41:40

definition of procurement. This is what

41:42

happened with TCP IP with the internet.

41:44

It happened with ACDC and electricity.

41:46

We have not done that yet for the model

41:48

era. And unfortunately, the difference

41:51

between st like these are called open

41:53

standards. The standardization process

41:55

is being confused with marketing. Now,

41:57

President Trump is actually, I think,

41:59

trying to do his best from what I can

42:00

tell in at least giving America enough

42:03

freedom to innovate that these standards

42:05

can even be discovered in our labs here

42:08

cuz first you need somebody to actually

42:09

pioneer and figure out what the

42:10

standards even should look like. I think

42:13

that

42:14

there's just a lot of noise. Do you

42:17

worry that basically the CCP is

42:19

subsidizing a generation of Chinese

42:21

models that are now being used by

42:23

American companies whereby they have

42:25

frontier models to essentially set where

42:28

model capabilities can be and then have

42:31

a real effort to make the open- source

42:33

Chinese models as close to those

42:35

benchmarks as possible much much

42:37

cheaper.

42:38

>> I mean the engineering execution right

42:40

now up and down the stack in China is

42:42

extreme. Here's what's happening right?

42:45

What they realized

42:47

is that the AI scaling race is not a

42:50

chip race. It's a full stack systems

42:53

code race where if you if you can't

42:56

compete head-to-head on chips for now,

42:59

what do you do? You compete on systems

43:01

design. You say, "Okay, we can't we

43:05

don't have leading edge chips here,

43:06

right, yet. So, let's try to compete on

43:09

systems." the you co-design the chip

43:14

that you have might be Huawei with the

43:16

computer infrastructure with the

43:17

training run and then you design that

43:20

okay to to have a bunch of performance

43:22

improvements at every layer of the stack

43:24

and then what you do is you do

43:26

adversarial distillation at scale where

43:27

you take western models and then you

43:29

from various different endpoints you

43:32

distill the the state-of-the-art and

43:35

then you try to get as many performance

43:37

gains as possible on that data and then

43:39

you release that back out to the world

43:41

as open models and then you see what

43:43

people react to and then you get

43:45

feedback and then you do the next run

43:47

and the next run and then you catch up

43:48

and at the point you catch up you say

43:50

wait a minute we're starting to be at

43:52

the frontier. Why do we need to open

43:54

source anymore? This is good enough for

43:55

our local domestic needs. It's

43:58

beautiful. It's actually and and and

44:00

that has actually by the way resulted in

44:01

innovation. They're they're innovating

44:02

at every step part of the cycle. And

44:06

that's why Huawei chips are able to

44:08

produce capabilities, improvements today

44:11

in China that rival some of the best

44:14

chips here when when integrated up and

44:17

down the stack. In a sense, it's the

44:18

Google strategy, right? Google is

44:20

integrated land power shell, TPUs, Borg,

44:23

Xborg, GQM, Gemini. Then the deployment

44:27

I mean the systems code design there up

44:29

and down results in efficiency that that

44:32

gives you huge performance gains at the

44:34

end of the day. China's replicated that

44:36

strategy using open source as sort of a

44:38

bootstrapping mechanism to catch up.

44:41

It's it's extraordinary.

44:43

>> Does that concern you?

44:45

>> Are you kidding? Absolutely. That's why

44:47

I think what we need is a western grid

44:49

that is where all inference frontier

44:51

inference is served through an iron

44:52

dome, right? where where if there's any

44:55

adversarial distillation attacks on any

44:57

one of our teams, we coordinate

45:00

together. So, because I'm on seven

45:01

boards, I I'm in group chats where I get

45:04

texted by one founder saying, "An is

45:07

anyone else noticing today that there's

45:08

a huge spike in distillation on from

45:10

this region and then I put them in a

45:12

group chat, we coordinate." It's very

45:13

informal right now, but what we need is

45:16

>> you said before that state sponsored

45:17

attacks on Frontier AI labs are getting

45:19

worse. What do we not know that we

45:22

should know?

45:25

Um, we should know that there are

45:27

insider threats.

45:29

We should know that there's distillation

45:31

happening across the US and Europe that

45:35

is taking advantage of our dist of of us

45:39

all not being united. They're that that

45:42

distillation is is taking advantage of

45:45

our political systems that our mission

45:48

critical infrastructure is is quite

45:51

vulnerable especially data centers that

45:53

are serving

45:55

uh workloads that are being used by

45:58

enterprises and I think that from a

46:00

business standpoint if we don't secure

46:03

frontier model inference or what I call

46:06

state-of-the-art inference behind a

46:07

coordinated Iron Dome we I don't think

46:10

we have a sustainable shot at at staying

46:12

at the frontier over the next decade.

46:15

>> I'm sorry. What does that mean? An iron

46:17

dome for inference in terms of

46:19

sustaining it.

46:21

>> It means that all inference is served,

46:24

no matter which company is serving it,

46:26

is served through a shared proxy that

46:28

can tell each other when there's an

46:31

attack happening on one part of the

46:32

frontier. Think of it as an iron dome

46:35

across the entire Western Front, right?

46:37

And just because you're here, you're in

46:40

one company,

46:41

you you you can't see that your model

46:44

being served through this other company

46:45

is being distilled. So it's it's a

46:48

deployment coordination protocol. It

46:51

it's it's basically my group chat that

46:52

I've got with like you know a bunch of

46:54

different founders but scaled where

46:57

people go we're seeing this attack today

47:00

and others go we are too. Let's

47:02

coordinate on defensive response.

47:04

>> I'm sorry for my lack of cohesion on

47:06

question. really I feel guilty and I

47:08

don't blame you for leaving this

47:09

interview thinking God he's got worse

47:10

over the 8 years not better but I was

47:12

watching this interview was speaking of

47:13

inference with someone I think from base

47:16

10 and they were saying that the demand

47:18

for inference has grown not linearly but

47:20

combinatorally and that is how we would

47:23

see it progress over the next 3 to 5

47:26

years do you agree with that

47:28

>> if we keep scaling capabilities that

47:32

will definitely happen the problem is

47:35

there are a couple bottlenecks on

47:36

scaling capabilities that are quite

47:37

existential. One of them we've talked

47:39

about is I mean the four core

47:41

bottlenecks on the capabilities progress

47:42

we've talked about right it's context

47:44

compute capital and culture and I think

47:47

capital allocation huge problem we got

47:49

to educate people on why this is why

47:51

these capabilities are extraordinary

47:53

like this this is like the biggest

47:55

financial bonanza of all time if you

47:56

know where to allocate I mean there's a

47:57

reason why I invest in anthropic in the

47:59

seed round and now as you've pointed out

48:02

like the returns of all the the body of

48:04

work I've done the last four years are

48:06

attracting LPS at the highest levels

48:08

But we're just getting started. And so

48:11

that that I I think some of these

48:12

projections you see are correct. If we

48:15

unblock the bottlenecks along the way in

48:17

computer infrastructure, secure compute

48:19

infrastructure that's funible, that's

48:20

standardized, that's the biggest

48:22

bottleneck. I think if there's any

48:23

reason why OpenAI, Enthropic, Gemini,

48:26

and so on don't hit their revenue

48:28

targets over the next few years, it's

48:29

because they won't have access to enough

48:30

compute. I will say there's there's like

48:33

a related bottleneck. When I was at

48:36

Stanford many years ago as a kid, I I

48:38

took this class that Peter taught called

48:40

uh I think it was turned into this book

48:42

called 0ero to1. This is Peter Teal. I

48:44

used to be I was an editor for the

48:45

Stanford Review and he had this um quote

48:50

right which is competition is for losers

48:53

and um

48:55

you know having done this now for 15

48:57

years I've kind of updated my theory of

48:59

business and I think he was he was not

49:01

wrong but he was insufficiently precise

49:03

which is that I think perfect

49:04

competition is for losers. I also think

49:07

monopolistic

49:08

>> what does that what does that mean

49:09

>> perfect competitions for these

49:10

>> it means that if you have 10 different

49:12

like 50 companies all doing LLM training

49:15

or doing coding models that's that's a

49:17

losing proposition it's it's like you

49:19

know perfect competition is like

49:20

restaurants there's no defensibility

49:22

that's why restaurants go out of

49:23

business all the time it's very hard for

49:24

them to differentiate on the other hand

49:26

in monopolistic comp monopolies are

49:28

mafias if once you have a monopoly at

49:31

one part of the stack they stop

49:32

innovating and instead they try to go up

49:34

or down by using the balance sheet to

49:36

acquire. They start hoarding resources.

49:38

They start saying, "You give me this and

49:40

I will force you to basically subsume

49:42

yourself to me." And I'm seeing that

49:43

kind of behavior up and down the stack.

49:45

And mafias are not good for innovation.

49:47

I I think we're in an era of op what we

49:49

need is optimal competition. The optimal

49:52

competition

49:54

setup is you have three or four teams in

49:57

every frontier that are making

49:59

extraordinary progress and so if you

50:01

invest in them you get extraordinary

50:02

returns but they're not so comfortable

50:06

as to be a monopoly such that they can

50:08

stop innovating and that's important

50:10

because when they stop innovating as

50:11

humanity we're [ __ ] And so I believe

50:14

that optimal competition we are living

50:17

we we need to transition to the optimal

50:19

competition in frontier technology and I

50:22

think we need leaders stewards venture

50:25

capitalists politicians educators to

50:28

remind the world that we have already

50:30

lived through this era of boom and bust

50:32

and so on and so these these companies

50:34

like what's going to happen right like

50:36

you said an banan and inference all

50:38

these companies inference is an

50:40

extraordinary growth curve ahead

50:44

But it's not going to be an

50:45

extraordinary growth curve if there are

50:46

50 inference companies all competing

50:48

with each other on a race to the bottom,

50:49

which is kind of what's happening right

50:51

now. Like it is not clear to me that we

50:54

need 50 inference companies. And it's

50:57

not clear to me that VCs are smart

50:58

enough to realize that they're just

51:00

lighting hundreds of millions of dollars

51:02

on fire in a category where having four

51:05

or five really good inference trusted

51:08

providers is net good.

51:10

But will the VC subsidization of 50 20

51:15

50 60 70 whatever companies it is not

51:18

make it impossible for the good

51:20

companies the four five to progress

51:23

through that cycle. It it's a bit of a

51:26

selfdestructive mechanism because if you

51:28

have 50 different companies all

51:30

competing for scarce compute resources

51:31

then the the folks who are actually

51:33

innovating don't have can't get it and

51:34

so they can't do their next round of

51:36

product innovation and so on. And that's

51:38

the problem when you have like this Is

51:40

that where we are now though?

51:41

>> That's where we are right now is the

51:43

best inference teams are calling me up.

51:45

Actually, all inference teams are

51:46

calling me up and saying, "And do you

51:47

have compute for us?" Cuz that's their

51:49

product is reselling compute. But it's

51:51

been hoarded. It's been hoarded by the

51:53

hyperscalers. It's been hoarded by

51:55

people who are not innovating but are

51:57

sitting on compute. And it's so obvious

51:59

to me now that I've left A6Z, I'm an

52:02

independent ecosystem public benefit

52:03

corporation that the that the

52:08

existential threat to innovation in this

52:10

category is lack of compute. Now that's

52:12

why AMP started procuring compute for

52:14

the independent ecosystem a while ago.

52:16

And so we are trying to find a way to

52:18

get these teams enough compute that they

52:19

need to keep innovating. But

52:21

>> we'll determine the four or five

52:22

inference companies that win versus the

52:25

others that don't.

52:26

>> Supply access to supply.

52:28

>> It's that simple.

52:31

>> Yep. Comput supply. If you don't have

52:33

compute, how do you do inference, man?

52:35

What are you selling? You need a product

52:37

to sell. So, if you're if you're making

52:40

a steam engine, you need coal. One of

52:42

your former partners tweeted last night

52:44

that we're going to enter a time where

52:47

only model I'm trying to remember it and

52:49

I wrote down parts of it, but only model

52:52

creators access the most powerful models

52:55

and that will power obviously the

52:57

services and the application layer or

52:59

the apps that they provide. Do you

53:02

believe that will be a world in which we

53:04

exist where model providers inherently

53:06

kind of safeguard the best models for

53:09

their provisioning of apps? Allah Claude

53:13

potentially or not? What Martine is

53:15

suggesting is that in competing cases

53:18

they will offer a worse model which

53:21

gives them an advantage. As an example,

53:23

11 Labs, which serves a huge amount of

53:25

application layer companies, will

53:27

reserve their latest models so they can

53:30

offer the best customer support and then

53:32

sell their older models to Sierra and

53:35

Decagon so they have a worse quality

53:37

model retaining the best for themselves.

53:41

The embedded assumption there right what

53:44

we have learned over the year like

53:45

empirically over the history of

53:46

technology is that you want if you have

53:50

a general purpose product like the

53:52

iPhone right that works for everybody

53:55

then the natural the natural incentive

53:58

is to amortize the cost of product

54:00

development of this over the largest

54:02

number of users. So if you have a

54:04

general model that's good for everybody

54:06

it will be available to everyone. If you

54:08

have specialized models that are good

54:10

for some people, there will be price

54:11

there will be product segmentation. And

54:13

I think what this is telling us is that

54:16

if there are many custom models, they

54:18

will some of them will be accessible,

54:19

others will not be. And so if anything,

54:21

I I think we should see the fact that

54:23

like there are Frontier Model Labs

54:25

saying, "Hey, here's a new model we

54:26

have. It only makes sense for some large

54:28

enterprises to access this as

54:30

vindication of the of the like ecosystem

54:34

truth that they're going to there's

54:36

going to be an ecosystem of different

54:37

models of different types. There's no

54:38

one large god model and uh if because if

54:42

there was I think there would be the

54:44

market desire to have you know prime

54:47

ministers, presidents and I and students

54:49

all use the same iPhone cuz inherently

54:51

you can raise the most money and invest

54:52

the most product budget dollars to for a

54:55

general product and amortize the cost of

54:57

that across everybody. But if you have

54:59

specialized models, yeah, I don't think

55:01

they're going to be accessible to

55:02

everybody and they don't need to be. I I

55:04

I I think this open and closed access

55:07

thing is somewhat overblown. I think

55:08

just empirically from a systems

55:10

perspective, if you look at the history

55:11

of technology, if you have general

55:14

products, they're they're they're

55:16

distributed to the masses. If you have

55:19

custom products, they have enterprise

55:21

segmentation. Some are accessible to the

55:22

enterprise, others are not.

55:24

>> Are there foundation model layer

55:26

companies that are yet to be built that

55:28

will be worth over hundred billion

55:30

dollars?

55:31

>> Oh, so many. I'm periodic is one I'm

55:34

sitting in one right here, right? But

55:36

they're not foundation model companies.

55:38

I would call them frontier systems

55:39

companies. This is the problem. Every

55:41

time I kept calling trying to educate

55:43

people, you know, four years ago where

55:44

they'd be like an but you know,

55:45

Anthropic is a foundation model company

55:47

and Mistral is a foundation model

55:48

company. No guys, that's just one part

55:51

of what they do. Maybe they're starting

55:53

there because that's very that's a core

55:54

competence

55:56

but there's a reason why you know

55:59

anthropic also has a thing called cloud

56:01

code and there's al there's a reason why

56:04

mistral has something called mistral

56:05

compute and there's something called

56:07

there's a there's a reason why you know

56:09

Microsoft who's a cloud also has a

56:11

co-pilot business you know these labels

56:14

or categories of foundation model when

56:17

need to be viewed I think with more

56:19

suspicion than they are like what

56:20

matters is the full the systems code

56:22

design the systems the the full stack

56:25

like like frontier research loop that

56:27

you need to run with customers and then

56:29

later when that happens when you say oh

56:31

my god anthropic

56:33

is now they have they have they were a

56:35

model company and now they're launching

56:37

a product called cloud code I was like

56:39

what do you mean that was part of the

56:40

plan all along of course you need to

56:42

have a a pair programmer interface for a

56:44

model like why why would you assume

56:45

otherwise oh cuz you just weren't paying

56:47

attention and you had your neat market

56:48

maps that your associates were giving

56:49

you and you thought that was That was

56:51

truth. The these

56:54

the commercial community has forgotten

56:57

how to build businesses and they've

56:59

forgotten the difference between first

57:01

principles and marketing.

57:04

That's the problem. That's one of the

57:05

other misalignment problems. The ground

57:07

truth of these businesses, machine

57:09

learning systems businesses, they've

57:10

always been frontier systems businesses.

57:12

They were never just foundation model

57:14

businesses. Now, okay, if you had to

57:16

package that up and tell your LPs that

57:18

because that was legible to them,

57:20

then I I can't blame you, I guess. But

57:23

the LPs I work with, I'm very upfront

57:25

with them. I say, "Look, these

57:27

categories are going through huge

57:29

reinventions and and and if you want

57:31

when you partner with me, what you get

57:32

is a full stack sort of partner." And I

57:35

will tell you the first principles of

57:36

what's going on and these first

57:37

principles insights will change over

57:39

time. But you got to be comfortable with

57:42

huge capex outlays in businesses that

57:44

end up winning the entire category.

57:46

That's what Frontier Technology is. So I

57:49

don't know I think foundation models

57:52

have been a deeply mis and and this is

57:53

part of why I started the class four

57:55

years ago. I just thought security at

57:56

scale was going through a bunch of

57:57

reinvention and then we reinvented the

57:59

class to be infrastructure at scale last

58:00

year and this year it's frontier systems

58:02

because not enough people realize that

58:06

to keep the the tech the capabilities

58:08

frontier moving you need to think about

58:11

these projects these companies as

58:13

frontier systems projects not foundation

58:16

model projects. Does that make sense? It

58:19

does. But when I hear about the capex

58:21

required, I I respectfully ask, do you

58:24

have enough money? I think the $1.3

58:26

billion was

58:29

>> Yeah. like how much money Yeah. How much

58:32

money do you need? An

58:34

>> well for the gigawatt 1.3 gawatt which

58:36

was kind of our our proof of concept

58:37

that that capital is not a problem. I

58:39

think the question is if we want to

58:41

scale beyond that,

58:43

>> yeah, we need way more capital to be

58:45

deployed in across the western front in

58:48

the United States and US allied

58:51

countries.

58:51

>> How much money do you think you need?

58:53

>> As long as the capabilities frontier

58:54

keep moving and we want a healthy

58:56

independent ecosystem, we'll just keep

58:58

raising more capital. There's no end to

59:00

that. I I don't I don't really The day

59:03

machine learning stops working as a

59:05

systematic way to give humanity more

59:07

capabilities, that's when I'll say we

59:09

have enough, Harry, but that's so far

59:12

out I don't even know how to reason

59:13

about that.

59:15

>> I could talk to you all day, but before

59:17

we do a quick fire, how will Vans be

59:19

fundamentally different in 5 years time

59:21

than it is today?

59:23

>> Well, again, go back to history, right?

59:25

I think there will be a few people like

59:27

Arthur Rock and

59:30

um

59:32

you know Bob Swanson and and Mike Mara

59:35

who turn their their practice into

59:39

institutions then there'll be others who

59:42

don't and I think if they don't evolve

59:43

themselves for what entrepreneurs of

59:45

this era need then I think they should

59:47

get out of the venture capital business

59:48

because we don't need more bankers like

59:50

you know one of the beautiful things I I

59:52

my friend Vlad who runs Robin Hood

59:54

floated did recently this like venture

59:56

fund thing on on Robin Hood.

59:58

>> Yeah. Venture Robin Hood Ventures I

60:00

think it is.

60:01

>> Yeah. Yeah. But like when you have

60:03

software that can play many of the

60:06

coordinating roles of venture capital

60:08

firms, why do you need somebody who's

60:10

just a pure to borrow a Marcism, a

60:13

rapper on LPs, right? The the look

60:16

here's here's what I'm most concerned

60:17

about with the capital ecosystem. Not

60:20

enough of the wealth creation

60:22

opportunity that's happening in Frontier

60:23

is being shared with the public and and

60:26

that's not good for anybody because

60:30

if you don't share this wealth creation

60:31

opportunity with the people who are

60:33

supposed to be welcoming this technology

60:34

into their lives which is ultimately the

60:36

public what are they going to do say I

60:38

don't want these

60:38

>> data with the with the greatest of

60:40

respect a lot of the money in venture

60:42

capital funds are from endowments

60:45

pension funds teachers funds and so that

60:49

wealth distribution should ultimately

60:51

trickle down if we believe in that.

60:53

>> But how many venture capital firms were

60:55

in the seed round of entropic?

60:58

>> Oh, none.

61:00

>> That's the answer for you. And that's

61:02

happening again and again and again.

61:04

There's a huge misallocation of public

61:06

capital into venture managers who did

61:08

are not capturing enough value in

61:10

Frontier AI. Instead, they're investing

61:11

a bunch of stuff that's not going to

61:14

exist and the public's going to be mad.

61:16

Did you put 300 million bucks into

61:18

Anthropic in one go?

61:19

>> I've had the privilege to invest many

61:21

hundreds of millions of dollars into

61:23

Anthropic across several rounds from the

61:25

first to the most recent one. So, I

61:27

consider that uh lucky. I I intend to

61:29

give most that away to public benefit um

61:32

causes, public benefit education

61:34

programs. And I I I I think we're at the

61:38

very beginning part of anthropics uh

61:41

journey on commercial progress. Dude,

61:43

I'm going to do a quick fire around with

61:45

you because otherwise I'm going to take

61:46

all day. You can advise, you can advise

61:48

an LP investing in venture funds. One

61:51

thing,

61:52

>> what do you advise them?

61:54

>> Educate yourself. Take the class. Do all

61:57

the readings. Do the readings. Do don't

61:59

skip the hard work. too too many LPs are

62:02

outsourcing their hard work, the the

62:04

work they're supposed to be doing as

62:05

capital allocators, which is like

62:06

understanding what's actually going on

62:09

and then decide which venture managers

62:11

and allocators you think have a unique

62:13

defensible advantage of the bottlenecks.

62:14

I I would be investing in the

62:15

bottlenecks basically.

62:17

>> Dude, too many too many GPS are not

62:18

doing the work. The amount of GPS who've

62:20

never built anything with AI is

62:22

astonishing.

62:23

>> I agree. Completely agreed. And I don't

62:27

think you can be like I don't you'll

62:29

laugh at me like I've built with every

62:31

different like vibe code provider. I'm

62:33

trying to turn my media company into an

62:35

AI first media company. It's pathetic

62:38

compared to the [ __ ] that you do. But at

62:40

least I'm trying. I'm seeing the

62:41

bottlenecks of superbase integrations

62:44

and everything that comes with it. And

62:46

you learn by building. I think if you're

62:48

not doing that in the be beginning, you

62:50

shouldn't be investing period.

62:53

>> I completely agree. I mean I was there

62:54

there's a sovereign country that came to

62:57

me at the end of last year and said we

63:00

want to bring 26 of our ministers to

63:03

your house and do a one-year program

63:06

where we educate it's a frontier program

63:09

where we learn what's going on in AI

63:11

from from lectures and so on and then we

63:14

want to do a deployment project where

63:16

each of our ministers actually build AI

63:18

agents and I said you know what that

63:20

that like if you take take Stanford

63:22

CS153 that it's a microcosm of this

63:25

course I'm doing with this country, the

63:26

sovereign fund that we partnered with.

63:28

Um and that's the way you you have to

63:32

work like do the work to read the

63:34

literature understand what's going on in

63:35

research and then deploy yourself like

63:37

build tools uh you know the class

63:40

project the Stanford CS153 class project

63:42

is the oneperson frontier lab because I

63:46

do believe genuinely that what would

63:48

have taken 50 people to do four years

63:50

ago now with the right AI tools you can

63:51

do with one person and as a leader if

63:54

you haven't played with these tools and

63:55

deployed yourself and built your own

63:56

agent I don't think you understand

63:58

what's going on. I'm not letting the the

64:00

ministers who are taking this class with

64:01

me, I'm not letting them graduate until

64:03

they build and deploy agents. I've told

64:05

them they're not getting they're not

64:07

getting their graduate certification.

64:09

>> Have you told your wife that you've got

64:10

26 ministers coming to your house?

64:13

>> She let me co-host

64:14

>> date night. An

64:16

>> she let me co-host them at our house in

64:18

SF, you know, few weeks ago. And I'm

64:21

very lucky to Viv. I don't deserve Viv,

64:23

I'll tell you that. But she's very very

64:25

she she's missional aligned and we both

64:27

believe that the best thing we could be

64:29

doing with our time is is educating at

64:32

scale.

64:32

>> What makes Dario so good that other

64:36

people don't see from the outside?

64:38

>> One sheer scientific brilliance truly

64:43

like world-class technical ability in

64:45

his domain. an obsessive

64:48

um desire for truth seeeking to

64:52

admit like to to to keep reasoning

64:56

reasoning reasoning doing to keep doing

64:58

experiments until he's he's a physicist

65:01

at heart right like I I think Dario is a

65:04

physicist at the end of the day he's not

65:06

actually a computer scientist um and so

65:08

a physicist a world-class physicist

65:12

tries to derive and and and he's an

65:14

applied physicist um derive laws,

65:17

general laws of reality by looking at

65:19

data and running empirical experiments.

65:20

He's an empiricist and he has an

65:22

obsessive desire to be a good

65:24

empiricist. And the third is mission

65:28

alignment culture. He says this is our

65:30

focus. This is our mission. No drift. We

65:34

won't take shortcuts.

65:37

We we are willing to make huge tradeoffs

65:39

to hit this mission. And that attracts

65:41

the best talent, incredible talent. In

65:43

the face of criticism of people saying

65:45

you're a mercenary, you're blah blah

65:47

blah. You're just doing this for profit.

65:50

No, actually, it turns out there's a

65:52

ruthless desire to to stay focused on

65:55

the mission. And that results in hard

65:58

trade-offs and priorities. And if you

66:00

don't if you're not aligned on that

66:01

mission, then you'll just think he's

66:03

crazy or, you know, he's evil or

66:05

whatever. It's crazy how much ad

66:07

hominemum attacks people I've seen

66:08

against him. But he's that got that

66:10

clarity of mission. What have you

66:11

changed your mind on in the last 12

66:13

months?

66:14

>> You know, the biggest one is um health.

66:17

Um I

66:19

I've had some health experiences between

66:21

both my family members and myself have

66:23

had health experiences that made me

66:25

realize we all just don't know how much

66:27

time we have on Earth.

66:30

And that makes you stop taking for

66:32

granted how much time we have. And so I

66:35

started taking time much more seriously.

66:38

But I would say and this was my my kind

66:40

of and you know every lecture I do at

66:42

Stanford um we talk a lot about scaling

66:45

laws and technical stuff but I also give

66:46

the kids like an Andre's life scaling

66:49

laws lesson you know at every I'm very

66:51

inspired by Richard Fineman uh Fineman's

66:54

lectures you know always kind of combine

66:56

technical education with a little bit of

66:58

life coaching for them and and my f my

67:00

my like number one scaling law for them

67:01

for the students was take life seriously

67:04

but don't take it so seriously that you

67:07

forget what makes it worth living, which

67:08

is have fun with friends, work on

67:11

interesting projects with people you

67:12

love. Don't take relationships for

67:15

granted. It's humans that make the world

67:16

go around. And if you're so focused on

67:19

your next fund or your next raise or

67:21

whatever, you just take for granted the

67:23

one thing we all don't know how much we

67:25

have, which is time with each other. And

67:28

so I just start valuing my time more, my

67:30

relationships with people. You know,

67:31

there's so many people. I mean

67:33

my parents, you know, I left my parents

67:35

behind in India to move to college um at

67:39

Stamford and I have gone weeks of my

67:42

life not calling them or texting them

67:45

and now they're, you know, in their 60s

67:46

and I've

67:51

>> I would give you a hug if we were in

67:53

person.

67:55

>> [ __ ] I'm so sorry, man. I

67:58

>> Don't worry. It's okay.

68:00

Oh Jesus.

68:03

>> You know, the first money we ever made

68:05

from the show, we made it because my mom

68:08

has MS and we couldn't afford treatment

68:10

for her. And the only way that I could

68:13

pay for it was by putting adverts in the

68:15

show.

68:17

And that was how we did it. And they

68:20

still pay for it. Thank you to Vant for

68:22

paying for Mom's MS.

68:24

>> Thank you, Christina. Yeah. Thank you

68:26

for the corporate sponsors.

68:28

Um, yeah, man. The trade-offs, you know,

68:32

the sacrifices are

68:35

>> parents are amazing.

68:36

>> Parents are insane.

68:38

>> How do you escape the money treadmill? I

68:40

I didn't have money when I grew up and I

68:42

was like, I'll be happy when I get like,

68:44

you know, x amount of money. Any advice

68:47

on escaping that money treadmill? I was

68:49

very lucky that you know I went to

68:53

Singapore on a government scholarship

68:55

and um

68:57

Lie Kuwan Yu who is the you know was the

69:00

founding father of Singapore I'm a big

69:01

lieuanist realized that

69:05

you know the the best like they're they

69:08

didn't have many resources they had they

69:10

didn't have they didn't have money as a

69:12

founding nation they didn't have

69:15

they didn't had nothing basically other

69:16

than themselves and their location their

69:18

strategic location and he realized we

69:19

need to build a talent program. We need

69:21

to run this country like a company and I

69:23

we would recruit um the best talent from

69:26

across Asia and because I I think I was

69:28

the top 10 or something in some public

69:31

exam when in the 10th grade in India I

69:34

was tapped to be a scholar in Singapore

69:37

and I took I was a government scholar.

69:38

Now I didn't have to actually I was

69:40

lucky enough that my parents could have

69:41

paid for it. had a family business in

69:42

telecom but it was very important to me

69:46

to be independent from my parents

69:48

because in Indian culture and a lot of

69:51

cultures where like if you don't have

69:53

financial independence you are always

69:56

kind of beholden to somebody else and in

69:58

in the case of community cultures like

69:59

India like there's a lot of pressure to

70:02

adhere to their values and so on um and

70:07

I I think I did subconsciously I'm very

70:08

lucky I have a sister actually who lives

70:10

in London and who fought my battle for

70:12

me. She was she's 7 years older and I

70:14

got to see that she was a rebel and she

70:16

wanted to do all kinds of, you know,

70:18

things including she wanted to go to

70:20

fashion school and they didn't want her.

70:21

So, but she had to go to law school

70:22

because they were paying for it. And she

70:25

actually, I think, fought some of my

70:26

battles and made me realize like the

70:28

more independent I was, the more I more

70:31

freedom I had. And freedom matters to me

70:32

a lot. And so I have always found I'm

70:36

willing to I I I just define my goal, my

70:39

financial goals by through independence

70:43

is what m has always mattered to me. And

70:45

so I'm willing to make big tradeoffs in

70:48

money to retain my independence. And

70:52

anytime I find my independence feeling

70:54

threatened, I go, you know what? I need

70:57

a change. So I I that that's what

70:59

matters to me. I I I think you need to

71:00

figure out what is your mission. what

71:02

what matters to you more than anything

71:04

else. So, you're willing to just turn

71:05

down all kinds of money and job

71:07

opportunities and so on because that

71:09

clarifies a lot where you spend your

71:11

time basically.

71:12

>> My mission Yeah. My mission is really to

71:14

enrich the already very rich family

71:16

offices of Europe. That's it.

71:19

>> Okay. Well, then you've got a ways to go

71:21

on the treadmill, brother.

71:23

>> Was that not a bug? I didn't read the

71:25

memo. [ __ ] Um,

71:29

you know, I also think that people are

71:30

just not very funny anymore. Like we

71:32

lack a little bit of humor in a lot of

71:34

society. It's so sad.

71:36

>> Yeah. I I was with my partner the other

71:38

day. I'm like, "You speak like AI." And

71:40

he's like, "I know." And you know what?

71:42

I talk to my wife like I talk to Claude

71:44

and she [ __ ] me. I'm like, "Yeah,

71:46

that's not a good thing." Dude, a final

71:48

one. Um, it's a bit morbid, but like

71:52

what do you want to be remembered for?

71:54

Like what do you want Ana's legacy to

71:56

be?

71:57

>> You know, Viv asks me asked me this like

72:00

three years ago at a party. We were like

72:02

I think it was at at our like

72:04

anniversary or something. We were with

72:05

like 10 of our friends, our closest

72:07

friends, including some of the

72:09

co-founders of Entropic. And uh so there

72:11

were all these you know it it it was one

72:14

of these classic San Francisco kind of

72:15

like dinner parties and she puts me on

72:17

the spot and I just blurted out I want

72:19

she I think what she had asked was like

72:21

what do you wanted to say on your

72:22

tombstone? H and I and I blurted out he

72:25

was right.

72:30

And and the room just went dead quiet

72:32

and they were like, "Yep."

72:35

And it's because I have this obsessive

72:39

desire and need to try to learn where

72:42

the future is going and then tell

72:43

everybody about it. And then everybody

72:45

thinks I'm like some snake oil salesman

72:48

or whatever. And then like now that's

72:49

changed because now it turns out I was

72:51

so right. I made LP so much money that

72:53

they're all now asking me an can you you

72:55

know people like I was invited back to

72:57

teach this class at Stanford, right? So

72:59

people are are f I guess have realized

73:02

okay an might know a thing or two about

73:03

the future. Let's go get his take. I

73:06

went to this boarding school in India

73:08

called Rishi Valley and it was founded

73:09

by

73:10

>> seven years of no tech.

73:11

>> No tech.

73:13

>> Seven years.

73:15

>> Yeah. Rural India.

73:16

>> No wonder. No wonder you're a happy and

73:18

adjusted person.

73:21

>> It's taken me a while to get here. But

73:24

yeah.

73:24

>> Would you let your children have social

73:26

media?

73:26

>> Yes. I I I I I think it'd be crazy to

73:29

not let them have it in social have

73:31

access to social media, but I I think it

73:33

has to be done in moderation and most

73:35

parents have a really hard time

73:37

moderating it with their with their kids

73:40

and then it's really hard to moderate.

73:41

You know, I with Rishi Valley I had

73:42

access to a computer once a week and so

73:44

you need to enforce something like that

73:46

where you you don't take it for granted.

73:48

it's within a structured sort of

73:51

environment and then you develop good

73:54

habits and protocols and practices to

73:55

not be dependent on it but you like I

73:57

would plan my Wikipedia sessions like I

73:59

had to plan my like you got one hour a

74:01

week in the computer room in Rishi

74:02

Valley so you really got to plan like

74:05

the highest use of that time and so

74:07

you're not dependent on it but you just

74:09

use it as a high lever strategic asset I

74:12

that's how I think technology should be

74:13

viewed you shouldn't take it for granted

74:15

like the problem is you know people keep

74:17

saying we're we're we're going to have

74:18

the singularity. Like, have you realized

74:20

we're we've been at this for like 10

74:21

years. Half of you outsourced your brain

74:23

and thinking to this device. Anyway,

74:25

>> oh dude, uh I I so enjoyed doing this.

74:28

Thank you so much for putting up with

74:30

me. You've been utterly fantastic.

74:33

>> No problem, man. Thank you for doing the

74:35

consistency with which you've kept up

74:37

this. I mean, you're an institution now,

74:39

right? Like the hard thing is I mean, I

74:41

can't believe this. You you've been at

74:43

this we've gotten old together, right?

74:44

When you were doing this, like you said,

74:46

you were a kid. I was much younger,

74:48

>> dude. Me and Pat Grady, Pat Grady was

74:50

like one of the first people I met in

74:51

Venture and he was an associate and I

74:54

was like a 17year-old and like I I

74:57

laughed with Pat. I said to him the

74:58

other day, "Dude, you've gone from

74:59

associate to like a head of Sequoia and

75:03

I've gone from podcaster to to

75:05

podcaster.

Interactive Summary

An Midhar, a prominent AI investor and founder of AMP, discusses why scaling laws are far from reaching saturation, particularly in specialized fields like material science. He identifies four critical bottlenecks for AI: context feedback, compute, capital, and culture, with culture being the most decisive. Midhar details the early days of Anthropic, the strategic importance of sovereign AI infrastructure for Europe to counter the US Cloud Act, and his vision for a 'western grid' for compute. He also proposes a 'coordinated Iron Dome' for inference to defend against adversarial distillation and state-sponsored attacks, while advocating for a 'back to the future' approach to venture capital that prioritizes deep incubation over passive check-writing.

Suggested questions

6 ready-made prompts