HomeVideos

Dr. Fei-Fei Li, The Godmother of AI — Asking Audacious Questions & Finding Your North Star

Now Playing

Dr. Fei-Fei Li, The Godmother of AI — Asking Audacious Questions & Finding Your North Star

Transcript

1611 segments

0:00

I think the ability to learn is even

0:03

more important.

0:04

>> Yeah,

0:04

>> AI has really changed it. For example,

0:07

my startup when we interview a software

0:09

engineer honestly how much I personally

0:13

feel the degree they have matters less

0:17

to us now is more about what have you

0:20

learned what tools do you use how

0:23

quickly can you superpower yourself in

0:27

using these tools and a lot of these are

0:29

AI tools what's your mindset towards

0:32

using these tools matter more to

0:36

Dr. Lee, it is nice to see you. Thanks

0:38

for making the time.

0:39

>> Hi, Tim. Very nice to be here. Very

0:41

excited.

0:42

>> And we were chatting a little bit before

0:44

we started recording about how

0:46

miraculous and I suppose unfortunate it

0:49

is. It's somehow we managed to spend

0:51

three years on the same campus and

0:52

didn't bump into each other.

0:54

>> I know. And now I'm wondering which

0:57

college you were at and which clubs.

1:00

>> Oh, yeah. I was Forbes. I was in Forbes

1:02

College. No, I was Forbes, too.

1:05

>> Okay, this is for people who don't know

1:08

what the hell we're talking about. There

1:09

are these residential colleges where

1:12

students are split up when they come

1:14

into the school. And Forbes was way out

1:17

there in the sticks, right next to a

1:20

fast food spot like 7-Eleven called Waw

1:22

Wa

1:23

>> Waw Wa

1:25

>> and next to the commuter train. And then

1:28

there's something called eating clubs at

1:30

Princeton. People can look them up, but

1:31

they're effectively co-ed

1:33

fraternity/sorities where you also eat

1:35

unless you want to make your own meals.

1:37

And I was in Terrace.

1:39

>> I was not any of that. But for those of

1:41

you wondering why we didn't meet, we

1:43

should say we were very studious

1:45

students who were only in the libraries.

1:48

>> Yeah, we were very studious. I actually

1:51

made my whatever it was $6 an hour at

1:54

guest library working up in the attic.

1:57

>> Tim, I work in the same library. I don't

1:59

understand why we did not meet.

2:02

>> That's really hilarious. Okay. So, well,

2:05

now we're meeting.

2:06

>> Did you change name or something? Maybe

2:08

we did meet.

2:10

>> I did didn't change my name, but here we

2:13

are.

2:13

>> Yes,

2:14

>> we've reunited. That's wild that we

2:16

didn't bump into each other. I was also

2:17

gone for a period of time because I went

2:20

to Princeton in Beijing

2:22

>> and went to the what was it? Capital

2:25

University of Business and Economics

2:27

after that. And so I was gone for a good

2:31

period of time and then took a year off

2:35

before graduating with the class of

2:36

2000. So still we had a lot of overlap.

2:39

>> Yes.

2:40

>> But let's hop into the conversation and

2:43

this is a very perhaps typical way to

2:48

start but in your case I think it's a

2:50

good place to start which is just with

2:52

the basics chronologically.

2:54

Where did you grow up? And could you

2:56

describe your upbringing? Because based

3:00

on my reading, your parents were pretty

3:02

atypical for Chinese parents in my

3:04

experience. Certainly.

3:06

>> Yes. You know, a lot.

3:08

>> Yeah. Could you speak to that please?

3:10

>> I would say my childhood and leading up

3:13

to the formative years is a tale of two

3:17

cities. I grew up in a town in China

3:19

called Chundu. I was born in Beijing but

3:22

most of my childhood was spent in Chundu

3:25

where it's very famous for panda bears

3:29

and at the age of 15 my mom and I joined

3:33

my dad in a town called Precipity New

3:35

Jersey. So I went from a relatively

3:40

typical middleclass Chinese family,

3:44

Chinese kid

3:46

>> to become a new immigrant in a

3:49

completely different world of all

3:51

places, New Jersey,

3:54

>> to learn a new language, to learn a new

3:56

culture, to to embrace a new country.

3:59

Mhm.

3:59

>> And then from there on I went to

4:02

Princeton as a physics major, but I did

4:06

take some of the classes you took

4:09

and and then went to Caltech as a PhD

4:12

student to study AI and and the rest is

4:15

history. I want to hear about both your

4:17

parents, but I want to hear a little bit

4:20

about your dad because he seems like

4:23

based on my reading a very whimsical

4:28

sort of creative soul which is a sharp

4:31

contrast in some ways to for instance I

4:33

had Bo Sha on the podcast amazing

4:37

entrepreneur and his father was I

4:40

suppose what some folks might think of

4:42

when they imagine not a tiger mom but

4:44

like a tiger ad. So in the case of B's

4:47

upbringing, his father was very strict,

4:49

but if he if he, meaning Bo, won a math

4:53

competition, then he would get extra

4:55

love and he would be allowed to have

4:56

certain treats and things like that.

4:59

>> Could you just describe your parents a

5:01

little bit?

5:01

>> First of all, clearly you read my book.

5:04

Thank you for that. It is true. As a

5:06

child, you don't realize as I was just

5:09

going through my my own science memo, I

5:12

was writing it, the more I wrote about

5:13

it, the more I realized, oh my god, I

5:15

really did not have a typical dad.

5:18

>> My dad loved and still loves nature.

5:22

He's just a curious. He finds humor and

5:26

fun in unserious things, you know, like

5:30

he loves bugs, insects. Mh.

5:33

>> Growing up in the 1980s in China, there

5:36

isn't much abundance in terms of

5:39

material resources.

5:41

>> But my city Chundu was expanding. So we

5:44

lived in apartment complexes at the edge

5:46

of the city. Even though my dad and my

5:49

mom worked in the middle of the city. So

5:51

on the weekends, my dad and I would just

5:54

play in the fields where there's still

5:56

rice fields. There's water buffaloos. I

5:59

had a puppy. really all my memory is

6:02

just like fighting bugs really and then

6:05

sometimes my dad and I will follow some

6:09

I don't know we took a art class I took

6:11

a kids art class and we'll go to the

6:14

mountains neighboring mountains to draw

6:18

>> but my entire childhood memory of my dad

6:21

is is just a very unserious parent who

6:25

had no interest in my grades or what I'm

6:29

doing in class. Did I achieve anything?

6:32

Did I bring back any competition awards?

6:36

Nothing to do with that. Even when I

6:40

came to New Jersey with my parents, life

6:43

became extremely tough, right? It was

6:45

immigrant life. We were in a lot of

6:49

poverty. And even that my memory is that

6:52

he had so much fun in yard sales. Like I

6:55

would just go to yard sales and and

6:57

those are our every weekend it was just

7:01

yay let's go to yard sales and just use

7:05

that as a treasure hunt almost. So so

7:08

it's just he's a very curious and

7:11

childlike mind in that way.

7:13

>> I'm asking about your parents in part

7:15

because I know you're a parent and

7:16

ultimately I'm going to want to ask how

7:19

you think about parenting. that will

7:21

come up at some point. But since

7:23

listeners will certainly be asking

7:25

themselves this question and we're not

7:26

going to get into any geopolitics

7:28

because there are plenty of people who

7:29

want to get into that and fight over

7:31

that which we're not going to do. But

7:33

why did your parents leave China? like

7:35

what was the catalyst or what were the

7:37

reasons behind leaving what you knew or

7:40

leaving what they knew and coming to a

7:43

very different foreign country where I

7:45

you're going from Chungdu

7:47

which is a city to suburban New Jersey

7:50

which is as I think you've described it

7:53

felt very empty right and then you have

7:55

the language barriers and the financial

7:56

barriers there's so many things why the

7:58

move

7:59

>> I'll give you two answers in the early

8:02

teenage fay would say I I have no idea

8:06

>> because my dad left when I was 12 and my

8:09

mom and I joined him when I was 15 and

8:12

those years you're you're a teenager,

8:15

right? Like there's so many strange

8:17

things in your head and all I knew is

8:21

that you know they said let's go to

8:23

America and I had no idea. I really did

8:27

not know what happened. There was this

8:30

vague sense of there's opportunities and

8:33

freedom. there's this that education is

8:36

very different

8:37

>> and I had a hunch that I was not a

8:42

typical kid

8:43

>> in the sense that you know I was a girl

8:46

and

8:47

>> I loved physics I loved fighter jets of

8:51

all things you know I can tell you all

8:54

the fighter jets I love from F-17 to F16

9:00

to you know to all the different things

9:03

that I I loved. So that's all I knew. In

9:06

hindsight, as a grownup fa

9:10

>> I appreciated my parents. They're very

9:13

brave people because I don't know this

9:16

age myself would just pick up and and

9:21

leave a country I'm familiar with and go

9:23

to I don't know a completely different

9:26

country that I speak zero language and I

9:29

have zero connectivity to

9:32

>> and mind you that's pre- internet pre-

9:35

AI age so when you are

9:37

>> going to a different country you're comp

9:38

you're just you might as well go to a

9:40

different planet

9:41

>> you're cut of

9:42

>> Yeah.

9:43

>> Yeah.

9:43

>> I think they're very brave. The grown-up

9:45

Fay realized that they wanted me to have

9:48

an opportunity that that they think will

9:52

be unprecedented for for my education.

9:56

>> Mhm.

9:57

>> And it turned out that's that's kind of

9:59

true.

10:00

>> Yeah.

10:02

Well, certainly looking at your bio, I

10:05

mean, it's mind-boggling to imagine all

10:09

the different sliding door events and

10:12

different paths you could have taken.

10:14

So, we're going to hop pretty closely

10:17

along chronologically, but we're going

10:18

to ultimately get to a lot of the meat

10:21

and potatoes of the conversation, but I

10:23

want to touch on maybe some other

10:26

formative figures. And I would like to

10:28

hear about your mother as well, because

10:31

just with the context of your dad, it's

10:33

like, okay, that seems fascinating and

10:35

very unusual, particularly if you've

10:37

spent any time in China, especially

10:40

during that period of time.

10:42

>> He is very unusual that way. Yeah, very

10:44

unusual. So then people might wonder,

10:46

well, where does the drive come from?

10:47

Where does the technical focus come

10:50

from? And I'd love to hear your answer

10:52

to that and also hear you explain who

10:56

Bob Sabella was, if I'm pronouncing that

10:59

correctly.

11:00

>> Yes. There are two questions mostly, you

11:03

know, is my mom the one who put in the

11:05

drive and the technical passion and and

11:08

what role did Bob play in my life? So

11:11

first of all, my mom has zero technical

11:13

jeans here. I sometimes still laugh at

11:15

her. She cannot do math. Let's put it

11:18

this way. So I think the the technical

11:21

passion is just I was born with it

11:24

>> inate.

11:24

>> My dad is more technical, but he loves

11:26

bugs more than insects, more than

11:29

equations for sure. So I think that's

11:32

you know as an educator for so many

11:35

decades now myself and also as a parent

11:38

you have to respect the wonders of

11:40

nature there is this inner love and fire

11:45

and and passion and curiosity that comes

11:48

with the with the package right so

11:51

>> but my mom is much more disciplined

11:53

person she she's still not a tiger mom

11:57

in the sense I don't remember my mom

11:59

ever going after me on grades or she

12:03

really did not. My both my parents never

12:06

ever cared about me bringing any awards

12:09

home.

12:10

>> Mhm.

12:10

>> Maybe I did, maybe I didn't. But I can

12:13

tell you in our house there's zero wall

12:17

hangings of anything which actually

12:20

carried to today. Even for myself, my

12:22

own house, my own office have zero of

12:25

those decorations of achievements or

12:28

awards. It's just uh my mom did not care

12:31

about that. But she did care about me

12:33

being a focused person. If I want to do

12:36

something, she doesn't want me to play

12:39

while doing homework. And that kind of

12:41

thing would bother her. She would say,

12:43

"Just finish your homework. Say by 6:00

12:46

p.m. If you don't finish your homework,

12:49

you're not allowed to do more homework.

12:51

You have to deal with the consequences."

12:53

So she she instilled some discipline,

12:55

but that's about it. She's tougher than

12:57

my dad. She is very rebellious. She had

13:00

a unfinished dream herself. She was very

13:02

academic when she was a a kid herself

13:05

and cultural revolution really crushed

13:08

all her dreams.

13:09

>> She became a more rebellious person in

13:12

that sense that I think I did observe

13:16

and and experience as a daughter.

13:19

>> Maybe part of immigration is even part

13:21

of that. Many years later, she would

13:23

say, "I had no plan coming to New

13:26

Jersey, but I think I'm going to

13:27

survive. I just believe I'm going to

13:30

survive and I'm going to make sure Fay

13:32

survives."

13:33

>> I think that is her strength, her

13:37

stubbornness, and her rebelliousness.

13:39

>> When does Bob enter the picture and who

13:41

is Bob?

13:42

>> Bob Sabella was a high school math

13:45

teacher in Pipony High School. He he was

13:49

my own math teacher as well as many many

13:52

students. He entered my life in my

13:55

second year in per so it's kind of

13:58

bordering sophomore to junior year in

14:00

Persip high school when I started taking

14:03

AP calculus but he quickly became the

14:07

most influential person in my formative

14:10

years as a new American kid immigrant as

14:14

a teenager because he became my mentor

14:17

my friend and eventually his entire

14:20

family became my American family

14:23

>> and he became my friend when I was a

14:26

very lonely ESL English as second

14:30

language student.

14:31

>> I was excelling in math but I think it's

14:34

more because I was lonely and he was

14:37

very friendly. He treated me more like a

14:41

friend who talks about books we love,

14:44

talk about the culture, talks about

14:45

science fiction,

14:47

>> and also listened to me as a very, you

14:50

know, I wouldn't say confused, but

14:53

teenager undergoing a lot of life's

14:57

turmoil in in my unique circumstance.

15:00

And that unconditional support made me

15:04

very close to him and his family. And

15:08

one thing he did to me that I did not

15:10

appreciate till later is that when

15:13

Pimity High School couldn't offer a full

15:17

calculus BC class because it just didn't

15:20

have that, he just sacrificed his lunch

15:24

hour, his only lunch hour to teach me

15:27

calculus BC. So it was a one-to-one

15:29

class. And I'm sure that contributed to

15:33

me a immigrant kid getting into

15:35

Princeton eventually. But later as I

15:39

became teacher myself, it's exhausting

15:42

to teach all day long. And the fact that

15:45

on top of that, he would use his lunch

15:48

hours to do that extra class for me is

15:51

just such a gift that I now appreciate

15:55

more than I was as a as a teenager.

15:58

>> Yeah. Thank God for the teachers who go

16:00

the extra mile. It's just incredible,

16:02

especially when you

16:03

>> get a bit older and you have more

16:05

context and you can look back and

16:06

realize.

16:07

>> I really think these public teachers in

16:11

America are the unsung heroes of our

16:14

society because they are dealing with

16:16

kids of all backgrounds. They're dealing

16:19

with the changing times. the kind of

16:21

stories Bob would share with me in terms

16:24

of how he went extra miles not just with

16:26

me but with many students in because

16:30

Puberty is is a heavily immigrant town.

16:33

>> Mhm.

16:34

>> So his students are from all over the

16:36

world and how he helped them and their

16:39

family. It's just those are the stories

16:41

that people don't write about and that's

16:43

part of the reason I wrote the book was

16:45

to celebrate a teacher like that.

16:48

>> Yeah. I have so much I want to cover and

16:50

I know we're going to run out of time

16:51

before we run out of topics. So, I want

16:53

to spend more time on Bob and at the

16:55

same time I want to keep the

16:57

conversation moving. So, we're we're

16:59

going to do that and I'll just perhaps

17:02

hit on a few things and then dig into a

17:05

number of questions. But certainly at

17:08

Princeton you but also your entire

17:10

family had to survive. So, you were

17:13

involved with operating a dry cleaning

17:15

shop in New Jersey as one option, right?

17:17

you ran that for 7 years. So through

17:20

that, it feels like you've gained

17:21

perspective on many different levels

17:23

that have then helped inform what you've

17:25

done professionally, right? So you you

17:28

learn to think about not just people who

17:30

are protected in an ivory tower, but

17:32

people all the way down in across in

17:34

society. So from every swath of society

17:37

your mother also although she was not

17:39

technical she imbued in you this

17:41

discipline and also seems to have had a

17:44

very broad

17:46

appreciation and knowledge of literature

17:48

and international literature. So now you

17:50

have this global perspective presumably

17:53

at the time in Chinese and

17:58

then you end up you end up at Princeton

18:00

and I know we're going to be hopping

18:02

around quite a bit but I'm curious to

18:05

know how Imagenet came about and you can

18:09

introduce this any way you like. You can

18:11

tell people what it is and what it

18:13

became and why it's important and then

18:14

talk about how it started or you can

18:16

just talk about how it started. But it's

18:18

it's such a an important chapter.

18:21

>> So let me just explain what ImageNet is.

18:23

Imagenet on the surface was built

18:26

between 2007 and 2009 when I was an

18:29

assistant professor at Princeton and

18:31

then I moved to Stanford. So during this

18:34

transitional time my student and I built

18:36

this at that time the field of AI's

18:40

largest training and benchmarking data

18:43

set for computer vision or visual

18:45

intelligence. The significance today

18:48

after almost 20 years of imageet was it

18:52

was the inflection point of big data.

18:54

Before imageet AI as a field was not

18:59

working on big data and because of that

19:02

and couple of other reasons which I'll

19:04

get into AI was stagnating. The public

19:08

thinks that was the AI winter. Even

19:10

though as a researcher, young researcher

19:12

at that time, it was the most exciting

19:14

field for me, but I get it. It wasn't

19:17

showing breakthroughs that the public

19:19

needs.

19:20

>> But imaget together with two other

19:24

modern computing ingredients. One is

19:27

called neuronet network algorithm. The

19:30

other one is modern chips called GPU

19:34

graphic processing unit. These three

19:36

things converged in a seinal work,

19:40

milestone work in 2012 called image net

19:44

classification deep convolutional

19:47

neuronet network approach. That was a

19:49

paper that a group of scientists did to

19:53

show that the combination of large data

19:56

by imageet,

19:58

fast parallel computing by GPUs and a

20:01

neuronet network algorithm could achieve

20:04

AI performances in the field of image

20:07

recognition in a way that's historically

20:11

unprecedented.

20:12

And that particular milestone is many

20:15

people call it the birth of modern AI.

20:18

and my work image that was onethird of

20:21

that if you count the elements and I

20:23

think that was the significance I feel

20:25

really very lucky and privileged that my

20:28

own work was pivotal in bringing modern

20:31

AI to life

20:33

>> but the journey to image that was longer

20:35

than that the journey to image that

20:38

started in Princeton when I was an

20:40

undergrad you were in the East Asian

20:43

study department I was hiding in Jadwin

20:45

Hall which is our physics department.

20:48

>> Yeah,

20:48

>> I loved physics since I was a young kid.

20:52

I I don't know how somehow my dad's love

20:54

of bugs and insects and nature

20:57

translated in my head into just the

21:00

curiosity for for the universe. So, I

21:03

loved, you know, looking to the stars. I

21:05

loved the speed of fighter jets and the

21:09

intricate engineering of that eventually

21:11

translated into the love of the

21:15

discipline that that asks the most

21:17

audacious question of our civilization

21:20

such as what is the smallest matter?

21:24

What is the definition of spacetime? How

21:26

big is the universe? What is the

21:28

beginning of the universe? And in that

21:31

in that early teenagehood love I loved

21:34

Einstein. I love his work.

21:37

>> I wanted to go to Princeton for that.

21:38

But it turned out what physics taught me

21:42

was not just the math and and physics.

21:45

It was really this passion to ask

21:47

audacious question. So by the end of my

21:50

undergrad years, I wanted my own

21:52

audacious question. You know, I wasn't

21:55

satisfied with just pursuing somebody

21:57

else's audacious question. And through

22:00

reading books and all that I realized my

22:02

passion was not the physical matters. It

22:06

was more about intelligence

22:08

>> I was really really enamored by the

22:11

question is of what is intelligence and

22:14

how do we make intelligent machines.

22:17

So at that time I swear I did not know

22:20

it was called AI. I just knew that I

22:24

wanted to pursue the the study of

22:26

intelligence and intelligent machines.

22:29

And then I applied to grad school and I

22:31

went to Caltech. Caltech was my PhD. I

22:34

started in the turn of the century 2000.

22:38

And I think I consider that moment I

22:40

became a budding AI scientist. You know

22:44

that was my formal training as a

22:46

computer scientist in AI. Then my

22:50

physics training continued in a sense

22:53

that physics taught me to ask audacious

22:57

question and turn them into a northstar.

23:00

>> Mhm.

23:01

>> And in scientific terms that northstar

23:03

became a hypothesis.

23:05

>> Mh.

23:06

>> And it was very important for me to

23:08

define my northstar. And my first

23:11

northstar for the following years to

23:14

come was solving the problem of visual

23:18

intelligence

23:19

is how how we can make machines see the

23:22

world. And it's not just by seeing the

23:26

RGB colors or the shades of light is

23:29

about making sense of what's seen which

23:32

is you know I'm looking at you Tim. I

23:35

see you. I see a beautiful painting

23:37

behind you. I don't know it was real. I

23:40

see you're sitting on a chair like that

23:42

is seeing. Seeing is making sense of

23:44

what this world is. So that became my

23:48

northstar question. And that hypothesis

23:50

that I had is I have to solve object

23:53

recognition.

23:54

>> And then that was in my entire PhD was

23:57

the battle with object recognition.

24:00

There were many many mathematical models

24:02

we have done and there were many

24:04

questions but me and my field was

24:08

struggling. We can write papers no

24:10

problem but we did not have a

24:12

breakthrough and then luckily for me

24:15

Princeton called me back as a faculty in

24:19

2007. It was one of my happiest moment

24:22

of my life. I feel so validated my alma

24:25

m would consider giving me a faculty

24:29

job. So I happily moved back to

24:31

Princeton as a faculty this time and I

24:34

continue to be a Forbes member actually.

24:38

So at Princeton there was an epiphany is

24:41

that I realized there was a hypothesis

24:44

that everybody missed and that

24:46

hypothesis was big data. Could I pause

24:49

you there for a second because this is

24:51

the this is the point

24:53

>> that I'm so so curious about and I just

24:56

want to pause for a second also for

24:58

people who are interested in some of the

24:59

history of Princeton. It's pretty crazy.

25:01

They should look up the history of the

25:03

Princeton Institute for Advanced Study

25:05

and I remember taking some of those East

25:08

Asian Studies classes that you referred

25:09

to in classrooms where Einstein taught

25:13

and it's just the aura, the veneer. You

25:15

want to believe that you can feel it

25:17

just permeating the uh the entire campus

25:21

and it's fun in that respect. It's very

25:23

fun. But I'm going to read something

25:25

from a Wired piece that discussed you at

25:28

length and as you mentioned big data

25:31

before and after in terms of its

25:33

integration into the type of research

25:34

they were describing as it was written

25:37

and please feel free to fact check this

25:39

or push back on it but in wire they said

25:41

the problem was a researcher might write

25:43

one algorithm to identify dogs and

25:45

another to identify cats and then you it

25:48

says you know Lee began to wonder if the

25:50

problem wasn't the model but the data

25:52

she thought that if a child learns to

25:54

see by experiencing the visual world by

25:56

observing countless objects and scenes

25:58

in her early years, maybe a computer can

26:00

learn in a similar way. I want you to

26:02

expand on that for sure. And the

26:04

question for me is like why did you see

26:06

it, right? Why didn't it happen sooner?

26:09

>> We're all students of history. One thing

26:12

I actually don't like about the telling

26:15

of scientific history is there is too

26:18

much focus on single genius.

26:21

>> Yes, agreed. We know Newton discovered

26:24

the modern laws of physics but yes he is

26:27

a genius not to take away any of that

26:29

from Newton but but science is a lineage

26:32

and science is actually a nonlinear

26:34

lineage for example why was I inspired

26:37

by this hypothesis of big data because

26:40

many other scientists inspire me in my

26:42

book I talked about this particular

26:44

lineage of work by professor Beerman who

26:47

was a psychologist who was he was not

26:51

interested AI, but he was interested in

26:53

understanding minds. And I was reading

26:55

his paper and he particularly was

26:58

talking about the massive number of

27:01

visual objects that young children was

27:04

able to learn in early ages. Right? So

27:08

that piece of work itself is not image

27:10

that. But without reading that piece of

27:12

work, I would not have formulated my

27:15

hypothesis. So while I'm proud of what I

27:18

have done, my book especially wanted to

27:21

tell the history of AI in a way that so

27:26

many unsung heroes, so many generations

27:29

of scientists, so many crossdisiplinary

27:33

ideas

27:35

pollinate each other. So I was lucky at

27:39

that time as someone who is passionate

27:42

about the problem but also someone who

27:44

benefited from all these research. So

27:48

yes something happened in my brain but I

27:50

would really attribute to many things

27:53

happened across so many people's work

27:56

throughout their lifetime devotion to

27:59

science that we got to the point of

28:01

imageet. I'm so glad that you're

28:03

underscoring this because if you really

28:05

dig as a I don't consider myself a

28:08

scientist, but I I love reading about

28:10

the history of science. There's so many

28:13

inputs, so many influences, so many

28:15

interdependencies.

28:16

>> Yes.

28:17

>> And the simplicity of the single hero's

28:20

journey is appealing and its simplicity,

28:22

but it's almost never true.

28:24

>> It probably is never true. Even my

28:27

biggest hero, Einstein, right? He

28:30

anybody who knows me, anybody who read

28:31

my book knows how much I rever him and I

28:34

just love everything he's done. The

28:37

special relativity equation is a

28:40

continuation of Lawrence transform. Even

28:42

Einstein, he builds upon so many other

28:46

people's work. So I think it's really

28:48

important especially I'm sure we'll talk

28:51

about it. I'm here calling you in the

28:53

middle of Silicon Valley and we're in

28:55

the middle of an AI hype and obviously

28:58

I'm very proud of my field but I think

29:01

that when the media or whatever tells

29:04

the story of AI it almost always just

29:08

talk about a few geniuses and it's just

29:10

not true. It's generations of computer

29:12

scientists, cognitive scientists and

29:15

engineers who who made this field happen

29:18

>> for sure. I mean, everyone knows Watson

29:20

and Crick for for instance, but without

29:22

Rosalyn Franklin and her X-ray

29:25

crystalallography, it doesn't happen.

29:27

Doesn't happen. It just doesn't happen.

29:29

Point blank. We're going to hop to

29:30

modern day in a second, but with

29:33

ImageNet, I would love for you to speak

29:35

to some of the decisions or let's say

29:41

decisions or moments that were just

29:42

formative in making that successful,

29:44

right? Because for instance, if you're

29:47

going to

29:49

try to allow a machine to, and I'm using

29:52

very simple terms cuz I'm not technical

29:53

enough to do otherwise, to learn to

29:57

identify objects

29:59

closer to the path that a child would

30:01

take, you have to label a lot of images,

30:05

right? And I was reading about how

30:08

Mechanical Turk came into play and then

30:12

there's a competitive aspect that seems

30:15

to have driven some of the watershed

30:18

moments. Could you just speak to some of

30:20

the elements or decisions that made it

30:23

successful?

30:24

>> A lot of people ask me this question

30:26

because after image that many many

30:28

people have attempted to make data sets

30:31

but still only very few are successful.

30:34

So what made image less successful? I

30:37

think one of the success was timing is

30:39

that we truly were the first people who

30:41

see the impact of big data. So that very

30:45

categorical or qualitative change itself

30:48

is part of the success but it's also as

30:52

you were asking the hypothesis of big

30:54

data is not just size. A lot of people

30:58

actually misunderstands image nets

31:01

significance as well as other data sets

31:03

significance coming with the data set is

31:08

a scientific hypothesis of what is the

31:12

question to ask. For example, in visual

31:14

recognition you could talk about you

31:16

could make a data set of discerning RGB

31:20

and that would not be as impactful of a

31:23

data set that is organized around

31:25

objects. Mhm.

31:26

>> We can go down a rabbit hole of why not

31:29

because RGB is easier per se. It's

31:32

because you have to ask the scientific

31:34

question in the right way. Another

31:36

example is instead of making a data set

31:38

of objects, why don't you make a data

31:41

set of cities,

31:43

>> you know, that's even more complicated

31:45

that objects. But then that's dialing

31:47

too complicated. So every scientific

31:49

quest, you have to have the right

31:51

hypothesis and and asking the right

31:54

question. So that's one part of the

31:56

success is we defined visual object

31:59

categorization as the right hypothesis.

32:04

>> That was one rightness I guess. Another

32:07

rightness is that people just think oh

32:11

it's easy you just collect a lot of

32:12

data. Well first of all it's laborious.

32:16

But even aside from being laborious how

32:19

do you define the quality?

32:21

>> Mhm. You could say well if quality is

32:24

big enough we don't care about quality.

32:26

But how do you dial between what is big,

32:30

what is good and how do you trade off

32:33

that is a deeply scientific question

32:36

that we have to do a lot of research on.

32:39

And then another decision that is a set

32:42

of decision that is really hard is what

32:46

defines quality in terms of image. Is it

32:49

every image has higher resolution? Is it

32:52

it's photorealistic?

32:54

Is it because it's everyday image that

32:58

look very cluttered? Is it all product

33:01

shots that look clean? These are

33:04

questions that if you're too far away,

33:06

you wouldn't even think about asking.

33:08

But as a scientist, as we were

33:10

formulating the deep question of object

33:13

recognition, we have to ask this in so

33:16

many dimensions. And then you mentioned

33:19

Amazon Mechanical Turk. That is actually

33:21

a consequence of desperation

33:25

because when we formulated these this

33:28

hypothesis, our conclusion is we need at

33:32

least

33:33

tens of millions of high quality images

33:37

across every possible diverse dimension.

33:42

Whether it's user photos or is it

33:45

product shots or is it stock photography

33:49

like and then we need also high quality

33:52

labels. Once we make that decision we

33:56

realize this has to be human filtered

34:00

from billions of images.

34:03

>> So with that we became very desperate.

34:05

We're like how are we going to do that?

34:07

You know, I did try to hire Princeton

34:10

undergrads and as you know, Princeton

34:13

undergrads are very smart. But

34:15

>> they have very high opinion of the value

34:17

of their time.

34:18

>> Yes. And they're expensive. But even if

34:21

I had all the money in the world, which

34:23

we didn't, it would have taken so long.

34:26

So, we were very very stuck for very,

34:28

very long. We thought we had other

34:31

shortcuts, but the truth is human

34:34

labeling is a gold standard. M

34:36

>> we want to train machines that are

34:38

measured against human capability. So we

34:42

cannot shortcut that at that time.

34:44

>> Right?

34:44

>> So we had to go to what we eventually

34:47

found out is called crowd engineering

34:50

>> crowdsourcing and that was a very new

34:53

technology

34:55

was barely a year old or so

34:59

by Amazon. They they created a lot

35:01

online marketplace for people to do

35:05

small tasks to earn money. when these

35:09

tasks can be uploaded on the internet. I

35:12

remembered when I heard about Amazon

35:15

Mechanical Turk, I logged into my Amazon

35:17

account. I checked the first task I

35:20

checked out to do just to try was

35:24

labeling wine bottles or transcribing

35:27

wine bottle labels. The task will give

35:30

you a picture of a wine bottle and you

35:31

have to say this is 1999 Berdo and and

35:35

all that. Yeah,

35:36

>> people upload these kind of micro tasks

35:39

and then online workers like someone in

35:43

their leisure time like me if I had

35:45

leisure time I would just go sign up and

35:47

get paid to do that. And we realized

35:50

that was again out of desperation that

35:53

was a massive parallel processing with

35:57

online global population to do this for

36:00

us and that's how we labeled billions of

36:04

images and distilled it down to 15

36:07

million high quality image that images.

36:10

>> It's just so wild when you look at these

36:12

stories. is I just finished a book on

36:14

Janentech and there were all these

36:15

little technical inflection points that

36:17

also allowed things to happen right so

36:19

if it had been 5 years earlier

36:22

or maybe 3 years earlier right without

36:24

mechanical turk boy like it presents a

36:27

challenge

36:27

>> y

36:28

>> but also as you pointed out in science

36:30

it's one thing to get answers but you

36:32

need the input on the front end with a

36:35

proper hypothesis or a good question and

36:39

even with mechanical turk if you're only

36:42

focused on the

36:45

the mechanics of employing that, you can

36:48

get yourself into trouble. Because if

36:51

humans are incentivized, right, to let's

36:54

just say, I think this was the example I

36:55

read about, identify pandas in

36:58

photographs and they're paid for

37:00

identifying pandas, well, what's to stop

37:01

them from identifying a panda in every

37:03

photo, whether they exist in the photos

37:05

or not? Yes.

37:06

>> Right. So, you have to follow the

37:09

incentives as well. How did you solve

37:11

for that?

37:11

>> This is where you know my student and I

37:14

had I cannot tell you how many hours and

37:18

hours of conversation we have about

37:20

controlling the quality. We have to

37:22

solve for that in multiple steps. We

37:25

need to first filter out online workers

37:27

who are serious about doing the work. So

37:30

for example, we have to have some

37:32

upfront quizzes

37:34

>> so that they understand what a panda is.

37:37

They read the question and then once

37:40

they get into they qualify for that we

37:44

ask them to label pandas but there are

37:46

some images we know the correct answers

37:48

some are true pandas some are some are

37:50

not true pandas

37:52

>> but the labelers don't know so in a way

37:55

we implicitly monitor the quality of the

37:58

work by knowing where the gold standard

38:01

answers are

38:02

>> so these are the kind of computational

38:06

tactics we have to use to ensure the

38:09

quality of labeling.

38:10

>> Amazing. Just incredible. I'll actually

38:13

just put a recommendation out there for

38:15

a book, Pattern Breakers, by a friend of

38:17

mine, Mike Maples Jr. He taught me the

38:19

ropes initially of angel investing. But

38:21

in terms of identifying inflection

38:23

points and in some cases converging

38:25

technological trends that for the first

38:27

time make something possible which then

38:29

opens an opportunity right for something

38:31

with the right prepared mind in your

38:33

case and those of your collaborators and

38:34

the people you built upon for something

38:36

like imagageet pattern breakers is a

38:39

really good read for folks. So let's

38:41

let's hop to modern day then for a

38:43

moment and I would love to ask you right

38:45

because you've been called the godmother

38:47

of AI in our alumni magazine in fact and

38:50

elsewhere but you've had such a not just

38:53

technical but historical viewpoint

38:56

meaning you've over a broad timeline

38:58

well broad by AI standards been able to

39:00

watch the development and forking and

39:05

perils and promise of this technology.

39:08

What are people missing? What do you

39:10

think is eating up all the oxygen in the

39:12

room? What are people missing? Whether

39:15

it's things they should know or things

39:17

they should be skeptical of or otherwise

39:19

>> especially I'm here calling you from the

39:22

heart of Silicon Valley and I think

39:25

people are missing the importance of

39:27

people in AI

39:29

>> and there's multiple facads or

39:32

dimensions to to this statement is that

39:35

AI is absolutely a civilizational

39:38

technology. It's I define civilizational

39:40

technology in the sense that because of

39:44

the power of this technology it'll have

39:47

or already having a profound impact in

39:50

the economic, social, cultural,

39:53

political

39:55

downstream effects of our society. So

39:59

>> I just heard this is unverified but I

40:02

just heard that 50% of the US GDP growth

40:07

last year is attributed to AI growth.

40:12

>> Apparently this number is 4% for US GDP

40:16

have grown 4%. If you take away AI it's

40:21

only 2%. That's what means

40:24

>> that's civilizational from an economic

40:26

point of view. It's obviously redefining

40:29

our culture, right? Think about you're

40:31

talking about the word sucking oxygen

40:34

out of the room everywhere from

40:36

Hollywood to Wall Street to Silicon

40:40

Valley to political campaign to Tik Tok

40:45

to YouTube to Insta.

40:47

>> Taxis in Japan. I was just there and the

40:49

videos playing on the back of the

40:51

headset and the taxi. We're all talking

40:53

about AI. It's everywhere.

40:55

>> It's culturally impactful. Not only

40:57

impactful, it's shifting our culture and

41:00

it's going to shift education. Every

41:03

parent today is wondering what what

41:07

should their kids study to have a better

41:09

future. Every grandparent is say, "I'm

41:12

so glad I'm born early. I don't have to

41:15

deal with AI." but still worry about

41:18

their grandchildren's future. So AI is a

41:21

civilizational technology, but what I

41:23

think it's missing right now is that

41:25

Silicon Valley is very eager to talk

41:28

about tech and the growth that comes

41:31

with the tech. Politicians are just

41:33

eager to talk about whatever gets the

41:36

vote, I guess. But really, at the end of

41:40

the day, people at the heart of

41:42

everything. People made AI, people will

41:44

be using AI, people will be impacted by

41:47

AI, and people should have a say in AI.

41:50

And no matter how AI advances,

41:54

people's selfdignity as individuals, as

41:57

community, as society should not be

42:00

taken away. And that's what I worry

42:02

about because I think I think there's so

42:05

much more anxiety that because the sense

42:08

of dignity and sense of agency, sense of

42:12

being part of the future is slipping in

42:16

some people and I think we need to

42:18

change that. I've heard you say that

42:21

you're an optimist because you're a

42:24

mother. And both optimism and pessimism

42:28

to an extreme can bias us in ways that

42:31

are unhelpful, right? Or create blind

42:33

spots. And I'm curious if you try to put

42:36

your most objective hat on, which is

42:38

difficult for any human, but if you try

42:40

to do that, do you think people are too

42:43

worried, not worried enough, or worrying

42:47

about the wrong things? for people who

42:49

are not CEOs and builders and engineers

42:52

behind AI because you're right of course

42:54

I mean everybody will agree with this

42:56

that a lot of people are very worried

42:59

and I'm just wondering if it's if it's

43:01

ill-placed because I don't really if you

43:03

talk to some of the VCs who are the

43:04

biggest investors of course they have

43:06

this sort of in my view beyond all

43:09

possibilities techno optimist view of

43:12

the future where AI solves everything

43:14

right and it's hard to believe there's a

43:17

free lunch

43:19

And then you have the the doomers, the

43:21

doom and gloom where suddenly it's

43:23

Skynet next year and we're all slaves to

43:25

robots or eliminated, turned into paper

43:27

clips and reality is probably in between

43:29

those two. Do you think people are

43:31

worrying about the right things or have

43:33

they lost the plot in some way?

43:35

>> First of all, I call myself a pragmatic

43:37

optimist. I'm not a utopian. So I'm

43:40

actually the boring kind. I don't

43:42

believe in the extreme on both sides. I

43:45

travel around the world. Just last month

43:47

I was in Middle East, I was in Europe, I

43:51

was in UK and I I was in Canada, I came

43:54

back home in America. I think people in

43:57

America and people in Western Europe are

44:02

more worried about AI than say people in

44:07

Middle East, in Asia.

44:10

And I think we don't have to litigate

44:13

why they're more worried

44:15

>> but just to come closer to home just in

44:19

talk about us. I wish I have a megaphone

44:22

to tell people in the US that you're

44:25

known to be one of the most innovative

44:29

people our country have innovated so

44:32

many great things for humanity for

44:35

civilization.

44:36

We have a society that is free and

44:41

vibrant and we have a political system

44:44

that we still have so much say in how we

44:48

want to build our country. I do wish

44:50

that our country has more optimism and

44:56

positivity towards the future of using

45:00

AI than what is being heard now. I think

45:04

people like me technologists living in

45:07

Silicon Valley has a lot of

45:08

responsibility

45:10

in the right kind of public

45:12

communication.

45:14

So there's a lot of things that was not

45:17

communicated in the effective way. But I

45:21

do hope that we can instill more sense

45:25

of hope and

45:29

self agency into everybody in our

45:33

country because I think there's so much

45:37

upside of using AI in the right way. And

45:40

I want not just people in Silicon Valley

45:44

or in Manhattan, but I want people in

45:46

rural communities in traditional

45:49

industries in everywhere 50 states to be

45:54

able to embrace and and benefit from AI.

45:58

>> Why are you building what you're

46:00

building? What is World Labs? Why decide

46:02

to do this?

46:03

>> I actually answer this question very

46:05

often to every member of my team. Mhm.

46:08

>> I built World Labs. There are two levels

46:11

of this answer. From a technology point

46:13

of view, World Labs is building the next

46:15

generation AI focusing on spatial

46:18

intelligence

46:19

>> because spatial intelligence just like

46:22

language intelligence is fundamental in

46:25

unlocking incredible capabilities in

46:28

machines so that it can help humans to

46:32

create better, to manufacture better, to

46:35

design better, to build better robots.

46:38

So spatial intelligence is a lynchpin

46:40

technology. Mhm.

46:42

>> But one level up, why am I still a

46:44

technologist?

46:46

Is because I believe humanity is the

46:50

only species that builds civilizations.

46:54

Animals builds colonies or herds, but we

46:58

build civilizations.

47:00

And we build civilizations because we

47:03

want to be better and better. We want to

47:05

do good. Even though along the way we do

47:07

a lot of bad things but there is a

47:09

desire of having better lives, having

47:13

better community, having better society,

47:16

live more healthily,

47:19

have more prosperity.

47:21

>> That desire is where civilization is

47:23

built upon. And because I believe that

47:26

humanity can do that, I believe science

47:29

and technology is the most powerful

47:33

tool, one of the most powerful tools in

47:36

building civilizations and I want to

47:38

contribute to that. That's why I'm still

47:41

a scientist and a technologist and I'm

47:44

building world labs for that. Can you

47:46

explain to people what spatial

47:50

intelligence is and what the product is

47:55

so to speak at least as it stands right

47:56

now that you're building?

47:57

>> Spatial intelligence is a capability

48:00

that humans have which goes beyond

48:03

language is when you pack a sandwich in

48:08

a bag when you take a run or a hike in a

48:12

mountain. When you paint your your

48:17

bedroom, everything that has to do with

48:21

seeing and turning that scene into

48:25

understanding of the 3D world,

48:27

understanding of the environment and

48:29

then in turn you can interact with it,

48:32

you can change it, you can enjoy it, you

48:34

can make things out of it. That whole

48:38

loop between seeing and doing is

48:42

supported by the capability of spatial

48:45

intelligence. Right? The fact that you

48:46

can pack a sandwich means you know what

48:49

the bread looks like. You know how to

48:50

put the knife in between. You know how

48:52

to put the lettuce leaf on the bread.

48:56

You know how to like put the bread or

48:58

sandwich into a Ziploc bag. Every part

49:01

of this is spatial intelligence. M and

49:04

does today's AI have that? It's getting

49:07

better, but compared to language

49:09

intelligence, AI is still very early in

49:11

that ability to see, to reason,

49:15

>> and also to do in world in both virtual

49:19

3D world as well as real 3D world. So,

49:22

so that's what world labs is doing. We

49:24

are creating a frontier model that can

49:29

have intelligent

49:31

capability in the model to create world

49:35

to reason around the world and to enable

49:39

for example creators or designers or or

49:42

robots to interact with the world. So

49:45

that's spatial intelligence.

49:47

>> Could you expand on the you know

49:49

designers or creatives or robots

49:51

interacting with the world? So does that

49:54

mean that you could and my team has been

49:56

playing with with some of the tools. So

49:58

thank you for that. What does that mean?

49:59

If you could paint a picture for let's

50:01

say a year from now, two years from now,

50:04

how might someone use this or how might

50:06

a robot use this?

50:08

>> I was just talking to someone a couple

50:11

of weeks ago and it was really inspiring

50:13

is that high school theaters are very

50:16

low budget, right? like, okay, sometimes

50:19

I go to San Francisco opera or musicals

50:22

and the sets that's built for theater

50:26

are just so beautiful.

50:27

>> Mhm.

50:28

>> But it's very hard for high school or

50:30

middle school to have that budget to do

50:33

that. Imagine

50:35

>> that you can take today's worldlapse

50:37

model, we call it marble.

50:39

>> Mhm.

50:40

>> And then you create a set in medieval

50:44

French town.

50:45

>> Mhm. And then you put that in the

50:47

background and use that digital form to

50:51

help transport the actors and action

50:55

into that world. And of course,

50:57

depending on the auxiliary technology,

51:01

whether you're on a computer or

51:03

eventually people can use a headset or

51:07

whatever, you can have that immersive

51:10

feeling of being in a medieval French

51:13

town. That would be an amazing creative

51:16

tool for a lot of creators.

51:19

>> That was the example. Someone and I was

51:21

talking about it a couple of weeks ago.

51:23

But we already see creators all over the

51:27

world. Some of them are VFX creators.

51:30

Some of them are interior design

51:32

creators. Some of them are gaming

51:35

creators. Some of them are educators who

51:39

want to build some worlds that transport

51:43

their students into different

51:45

experiences are already starting to use

51:48

our model

51:49

>> because they find it very powerful at

51:52

their fingertip to be able to create 3D

51:55

worlds that they can use to to immerse

51:59

either their characters or themselves

52:02

into. And just a process-wise, if if

52:05

someone's wondering how this works,

52:06

let's just say it's a a public school

52:09

teacher, let's just say, who's hoping to

52:12

inspire and teach their students going

52:15

the extra mile. What does it look like

52:17

for someone to use this? Are they typing

52:18

in text, describing the world they'd

52:21

like to create, uploading assets or

52:23

photos, almost like an image board? How

52:26

does it how does it work? If someone's

52:27

nontechnical,

52:28

>> they don't need to be technical at all.

52:30

They open our page on desktop or in

52:34

their phone, but desktop is more fun

52:36

because it has more features.

52:38

>> And then they can type, you know, a

52:40

French medieval town or or they can

52:43

actually go to anywhere. They can use

52:47

midjourney or nano banana to create a

52:50

photo of a French medieval town or they

52:52

can get an actual photo about that and

52:54

then they upload it. We call it prompt.

52:57

And then after a few minutes, our model

52:59

gives you a 3D world that is

53:04

say a part of the tab. It does have a

53:07

limit in its range. And then that 3D

53:10

world is generally 3D because you can

53:13

just use the mouse to drag and turn

53:15

around and walk around and see that

53:19

world. And then downstream if you want

53:22

to use it, you have many ways to use it.

53:26

You can actually create a movie out of

53:28

it by using one of our tools on the

53:31

website to just put cameras and you can

53:33

make a particular movie out of it.

53:36

>> You could if you're a game developer.

53:38

>> I was just going to say it sounds a lot

53:40

like a gaming engine.

53:41

>> Yes, you you can put a lot of characters

53:44

in it. If you're VFX professional, we

53:47

have a lot of VFX professional. they can

53:50

actually take this and put it in the

53:53

workflow of their movie shooting and

53:55

have real actors shooting movies. We've

53:59

also have psychology researchers using

54:02

that immersive world in particular

54:04

psychiatric studies.

54:05

>> We could also use that as the simulation

54:08

for robotic training

54:09

>> because a lot of robotic training needs

54:12

a lot of data and then use that for

54:16

generating a lot of different data. So

54:18

is it almost like a flight simulator for

54:20

robots before they go into the real

54:22

world?

54:22

>> That's part of the goal. We are still

54:24

early. So the flight simulator is not

54:27

complete yet,

54:29

>> right?

54:29

>> But that's part of the journey.

54:31

>> You mentioned psychiatric studies. I I

54:34

think that's what you just mentioned.

54:35

Yes. What might that look like? We

54:37

actually got this researcher who called

54:40

us and they're studying people who have

54:44

psychological disorders like

54:47

obsessivecompulsive disorder

54:49

>> where they're triggered by certain

54:51

environments and they want to study the

54:54

trigger and also just study how the

54:56

treatment but how do you trigger someone

54:59

who let's say is particularly have issue

55:03

with let's say a strawberry field I'm is

55:06

making it up.

55:07

>> I mean, you can take them to a

55:09

strawberry field, but what about you

55:11

want to know if it's strawberry field in

55:15

the summer or strawberry field at night

55:17

or it's strawberry or it's mating

55:21

strawberry like how do you do this?

55:23

Suddenly this researcher realized we

55:26

give them the cheapest possible way of

55:28

varying all kinds of dimensions and they

55:30

can test this out and do their studies.

55:34

>> That's really interesting. Yeah, I could

55:35

see it being applied to it might be

55:37

called exposure therapy, but now that

55:39

you're describing it, I could see how it

55:41

could be

55:42

>> added into I mean pretty much

55:44

everything, right? I mean, if you think

55:45

about how humans operate in the real

55:47

world.

55:48

>> Yes. And the boundary between real world

55:51

and digital world is less and less,

55:53

right? Thinner and thinner because we

55:56

live in many screens. We live in the

55:59

real world. We do things in virtual

56:02

world. We do things in real world. will

56:04

create machines that can do things in

56:07

real world and virtual world.

56:09

>> So there's a lot we do in digital and

56:13

physical spaces.

56:15

>> Who are some scientists or researchers

56:19

who you pay attention to who are not

56:22

necessarily kind of the big brand names

56:25

and marquee lights that are already very

56:28

public in the world? Is there anybody

56:29

who stands out where you're like, you

56:30

know, there's some really tremendous

56:31

people doing good work? Well, that's

56:33

part of the reason I wrote the book is

56:36

especially in the middle chapters where

56:38

I wrote about the journey of doing image

56:41

that combines cognitive science with

56:44

computer science and I actually talk

56:47

about psychologists and neuroscientists

56:49

and developmental psychologists in you

56:52

know some of them are still with us some

56:54

of them are not for example the the late

56:57

anman

56:59

beerman they all passed away in the last

57:01

few years But they were giants in

57:04

cognitive science whose work has

57:06

informed computer science and eventually

57:08

AI. You know there are still lots of

57:11

scientists around the world. Many of

57:13

them are in the US who are thinkers in

57:17

developmental psychology in AI. I follow

57:20

their work. Mhm.

57:21

>> I think the world of science, just to

57:24

name some names, right, Liz Beli in

57:27

>> in Harvard, Allison Gobnik in Berkeley,

57:31

I love Rodney Brookke, who was a former

57:34

MIT professor in robotics,

57:38

>> and there's just a lot of them. I I

57:40

don't mean to just single them out.

57:42

>> Sure.

57:42

>> But you're asking me for names that are

57:44

not in in the news of AI.

57:48

>> Yeah, that's perfect. Thank you. I would

57:51

also love to get your perspective on

57:55

what might be this is a very strong word

57:57

but seemingly inevitable in in terms of

58:01

developments in the near intermediate

58:04

future. And I'll give you an example of

58:06

what I mean. In 2009

58:09

2008 2009 I became involved with Shopify

58:12

the company back when they had like 10

58:14

employees. And there were a few things

58:16

happening around that time and you could

58:20

ask questions, you know, in the next 10

58:21

years or 20 years, will there be more

58:23

broadband access or less? More. Okay.

58:25

Will there be more e-commerce or less?

58:27

There'll be more. Okay. And when you

58:29

have four or five of those that seem

58:32

over a long enough time horizon,

58:34

absolutely yeses, it begins to paint a

58:36

picture of where things are going. Are

58:39

there any things that in the next

58:41

handful of years you think are perhaps

58:43

underappreciated as near

58:46

inevitabilities?

58:47

>> You want me to talk about

58:48

underappreciated? I mean, I don't know

58:50

if they're overappreciated, but

58:52

definitely appreciated. The need for for

58:54

power is appreciated.

58:56

>> Mhm.

58:57

>> The trend of more AI, not less AI is

59:00

appreciated.

59:01

>> The long-term trend of robots coming is

59:04

appreciated. So, these are appreciated.

59:08

What's underappreciated is spatial

59:11

intelligence is underappreciated in the

59:14

sense that everybody's still now talking

59:16

about language large language models but

59:18

really world modeling of pixels of 3D

59:22

worlds is underappreciated because like

59:24

you were saying it powers so many things

59:27

from storytelling to entertainment to to

59:31

experiences to robotic simulation. I

59:33

think AI and education is

59:36

underappreciated

59:38

because what we are going to see is that

59:42

AI can accelerate the learning for those

59:46

who want to learn

59:48

>> which will have downstream implication

59:51

in our school system

59:53

>> as well as in just human capital

59:56

landscape like how do we assess

60:00

qualified workers?

60:02

>> Mhm. used to be which school you

60:04

graduate from with with which degree but

60:06

that will be changing. Yeah. With AI

60:09

being at the fingertip of so many people

60:12

that's underappreciated.

60:14

I think AI's impact in our economic

60:17

structure including labor market is

60:21

underappreciated.

60:22

The nuance is underappreciated. I think

60:25

this whole rhetoric of either total

60:29

utopia post scarcity is hyperbolic.

60:33

>> Yeah.

60:33

>> Or like everybody's job will be gone is

60:37

hyperbolic.

60:38

>> But the messy middle is how

60:42

from knowledge worker to blue collar to

60:45

hospitality to all these changes that's

60:49

happening. It's underappreciated by our

60:52

policy workers, by our scholars, by just

60:56

overall society.

60:58

>> Well, what are some of the nuances from

61:00

the job perspective? Maybe this ties

61:02

into what I promised earlier I was going

61:04

to ask you, which is what you are

61:07

telling or will tell I don't know other

61:08

ages your children are recommending.

61:11

Let's just say I don't know how old they

61:13

are, but if we assume that they just for

61:15

the sake of discussion of the age where

61:17

they're trying to decide what they

61:19

should study, where they should focus,

61:21

things of that nature, how how would you

61:24

think about answering that even

61:26

provisionally?

61:27

>> I think the ability to learn is even

61:30

more important because

61:33

when there was less tools, fewer tools

61:36

to learn, it's easier to just follow

61:39

tracks. You go through elementary

61:42

school, middle school, high school,

61:43

college and then get some, you know, get

61:46

some training vocationally and that's

61:49

kind of a path and with that is a set of

61:53

structured

61:55

credentials from degrees and all that

61:58

but AI has really changed it. For

62:00

example, my my startup when we interview

62:03

a software engineer honestly how much I

62:07

personally feel the degree they have

62:10

matters less to us now is more about

62:14

what have you learned what tools do you

62:17

use how quickly can you superpower

62:20

yourself in using these tools and a lot

62:23

of these are AI tools what's your

62:25

mindset towards using these tools matter

62:29

more to Mhm.

62:31

>> At this point in 2025, hiring at World

62:35

Labs, I would not hire any software

62:38

engineer who does not embrace AI

62:41

collaborative software tools.

62:44

>> Mhm.

62:44

>> It's not because I believe AI software

62:47

tools are perfect. is because I believe

62:50

that shows first of all the ability of

62:53

the person to grow with the fast growing

62:57

toolkits the open-mindedness and also

63:01

the end result is if you're able to use

63:03

these tools you're able to learn you can

63:06

superpower yourself better

63:08

>> so that is definitely shifting so coming

63:11

back to your question what do you tell

63:13

young people tell children I think the

63:16

timeless value of learning to learn, the

63:20

ability to learn is even more important

63:23

now.

63:24

>> Yeah, it it strikes me as we're talking

63:26

that it's only going to get increasingly

63:30

easier for

63:32

the ambitious to act as superpowered

63:35

autodidacts, right? We've already seen

63:37

this

63:38

>> with certainly YouTube has a nice track

63:40

record. Now you can either entertain

63:42

yourself to death and avoid doing things

63:44

that help with self-rowth and

63:45

development or you can supercharge it.

63:48

And similarly with AI, right, you flash

63:50

forward. We don't even need to flash

63:51

forward, but it's how does a teacher

63:54

audit that their students are doing the

63:56

work they're supposed to be doing.

63:58

>> Yeah.

63:58

>> On so many levels, it's getting to the

64:00

point. There are some exceptions, but of

64:02

near impossibility.

64:03

>> Yeah. and students can either avoid all

64:05

work or they can supercharge their own

64:08

work but the output might look very

64:09

similar at least for a period of time.

64:11

So schooling is going to change a lot.

64:14

It's very very interesting.

64:15

>> I actually think Tim

64:17

if the school evaluation is structured

64:21

in a way that whatever AI gives and

64:25

whatever the student gives is the same

64:28

there's something wrong with the

64:29

structure of the evaluation.

64:31

>> Okay. Can you say more about that?

64:33

That's interesting.

64:34

>> For example, English essay.

64:36

>> This is not me. This is me hearing a

64:38

story that I so agree with. I'll retell

64:41

the story. Is that as a high school

64:44

freshman English class teacher? I heard

64:47

that someone told me the story of their

64:50

kids school. On the first day of school,

64:54

the teacher actually said to the class,

64:57

I want to show you how I would score AI.

65:00

So the teacher give an essay topic. Show

65:03

the students this is what the best AI

65:07

gave me and I'm going to show you how I

65:10

think this is good, this is bad, how

65:12

this is suboptimal and I'll give it a B

65:15

minus.

65:16

Now I will tell you this is my bar. If

65:20

you're so lazy that you ask AI to write

65:23

your essay, this is what you're going to

65:26

get. But you can use AI, that's totally

65:29

fine. But if you can do the work, learn,

65:33

think, be the best human creator you can

65:37

and work on top of that,

65:40

>> you can get to a you can get to A+es.

65:43

And that would be in my opinion the

65:45

right way to structure the evaluation is

65:49

not to pit humans against the AI and

65:52

then try to police the use or not use of

65:55

AI. is that to show where the tools the

65:58

bar of the tools are and where the bar

66:00

of the human learner should be.

66:02

>> I'm going to sit with that example and

66:04

try to think of more examples. It's very

66:06

interesting and boy oh boy I've been

66:08

shocked by how quickly the models

66:10

improve. But yes, that's like as a

66:11

thought experiment.

66:12

>> Yeah,

66:13

>> I'm going to chew on that. I know we

66:14

only have a few minutes left. Fifth, I

66:17

wanted to ask you a question I ask a

66:19

lot, which is if you could put a quote

66:22

or a message, something on a billboard,

66:24

something to get in front of millions,

66:27

billions of people. Just assume they all

66:29

understand it. Could be an image, could

66:30

be a question, could be a quote,

66:32

anything at all. A saying, a mantra,

66:34

doesn't matter. Could be almost

66:36

anything. What would you or what might

66:37

you put on that billboard?

66:41

>> What is your northstar?

66:44

>> What is your northstar? This is of

66:46

course critically important and coming

66:48

back to

66:51

how you define that or find that for

66:53

yourself. I mean you were talking about

66:54

audacious questions

66:57

and then that leading to a north star or

66:59

hypothesis. Is there another way that

67:01

you would encourage people on top of

67:03

that to think about finding their north

67:05

star? I believe that's how that makes us

67:08

so human and makes us to be so fully

67:13

alive

67:14

>> is that we as as a species can live

67:19

beyond the chasing of just basic needs

67:22

right but dreams and missions and goals

67:26

and passion and everybody's northstar is

67:29

different

67:30

>> and that's fine not everybody has to

67:32

have AI as their northstar but finding

67:35

That goes to the heart of education

67:37

again and I don't mean formal classroom

67:40

education. It's just the journey of

67:42

education. A lot of that is the ability

67:45

to learn who you are and to learn how to

67:49

formulate your northstar and how to

67:52

chase after that.

67:54

>> Last question. Did your parents ever

67:56

explain to you why they named you Fay?

67:58

>> Yes. is because when my mom was going

68:01

through labor, my dad was

68:03

characteristically late to the hospital

68:06

and along the way he caught a bird. He

68:08

let it go, but he did catch a bird. I

68:11

don't know, he was just distracted and

68:14

it was in Beijing in the city of

68:16

Beijing. My dad was bicycling to my

68:18

mom's hospital

68:19

>> that inspired him to call me Fay.

68:23

>> Feet.

68:24

>> Oh, wait. Sorry for those who don't

68:26

speak Chinese. I forgot you do speak

68:28

Chinese, but for those who don't speak

68:30

Chinese, Fay means flying.

68:32

>> Means flying.

68:32

>> Yeah. So, be inspired by a bird.

68:35

>> Really quick, I'll just say it's kind of

68:36

funny. My first Chinese name that I had

68:39

was Fay Ting Chong, which is because I

68:41

was very blunt and honest. So, Ting, but

68:44

Fing Chong, but when I was first

68:46

starting, my tones in China were not

68:49

polished and people thought I was saying

68:51

that my name was Fiji Chang, which is

68:54

airport. So, I petitioned my teachers

68:57

and we changed my name to something less

68:58

less confusing.

69:00

>> What's your new name?

69:03

>> Oh, okay.

69:05

>> It's like

69:07

but it's without the at the bottom.

69:10

>> Oh, wow.

69:11

>> Fancy name. That's way more

69:14

sophisticated than my

69:17

>> Well, I get to script it with my Chinese

69:19

teachers, so I have an unfair advantage.

69:21

Dr. Lee, thank you so much for the time.

69:24

We will link to the show notes for

69:26

everybody at tim.blog/mpodcast. They'll

69:28

be able to find you easily and everybody

69:30

should check out worldlabs.ai

69:33

and we'll put every other link, your

69:34

social and so on in the show links. But

69:37

thank you for the time. I really

69:38

appreciate it.

69:39

>> Thank you, Tim. I enjoyed our

69:41

conversation.

69:42

>> Yeah, likewise.

69:42

>> Okay, bye

69:44

>> bye.

Interactive Summary

Dr. Fay-Fei Lee, co-director of Stanford's Human-Centered AI Institute and CEO of World Labs, discusses the profound impact of AI on society, from hiring practices to its role as a 'civilizational technology'. She shares her personal journey, including her childhood in China, immigration to New Jersey at 15, and her academic path from physics at Princeton to AI at Caltech. Dr. Lee highlights the significance of ImageNet, a massive dataset she co-created that became an inflection point for modern AI, and emphasizes the collaborative nature of scientific breakthroughs. She critiques the current AI hype, stressing the often-overlooked 'human element' in AI's development and application, and advocates for a pragmatic optimism regarding its future. Her current venture, World Labs, focuses on spatial intelligence to empower creators, designers, and robotics, and she advises young people to cultivate an enduring ability to learn and adapt to new tools.

Suggested questions

6 ready-made prompts