HomeVideos

Four CEOs on the Future of AI: CoreWeave, Perplexity, Mistral, and IREN

Now Playing

Four CEOs on the Future of AI: CoreWeave, Perplexity, Mistral, and IREN

Transcript

2762 segments

0:00

I'm here at Nvidia's annual GTC

0:02

conference and I'm going to interview

0:04

four amazing AI CEOs. Stick with us.

0:15

>> Our episode is sponsored by the New York

0:17

Stock Exchange. Are you looking to

0:19

change the world and raise capital? Do

0:21

it at the NYSE. The NYSE is a modern

0:25

marketplace and a massive platform built

0:28

for scale and long-term impact. So if

0:31

you're building for the future, the NYSE

0:33

is where it happens.

0:37

>> One of the great companies of the AI era

0:40

is of course Cororeweave. They're

0:41

building massive infrastructure for

0:43

these hyperscalers. And in some ways,

0:45

Michael Intrader, welcome to the

0:47

program. You're the original

0:49

hyperscaler. you guys got in very early

0:52

and secured your I don't know which GPUs

0:56

you wound up getting but you were very

0:58

early to this trend. How did you get to

1:01

it so early and how did you build out

1:04

this you know first I guess at the time

1:06

neocloud? Yeah. So we we didn't really

1:09

start it as a Neocloud and I I uh I was

1:12

uh running an algorithmic hedge fund uh

1:15

focused on natural gas and uh when when

1:19

you build an algorithmic hedge fund um

1:21

once the algorithms are built you're

1:23

really just monitoring it and testing

1:25

different uh thesises and doing all

1:27

that. But there's also a lot of downtime

1:29

and we got super interested in crypto.

1:32

Um, and you know, we're pretty nerdy. We

1:35

kind of dig under the hood and we

1:36

started to get interested in the

1:37

security layer. Uh, we looked at Bitcoin

1:40

and the mining for Bitcoin and we didn't

1:42

like it. We just thought that like

1:43

there's some brilliant engineer that

1:46

built the ASIC and they're probably

1:47

going to be better at running it than we

1:49

are. So, we really began to focus on the

1:51

GPUs mostly because the GPUs were you

1:55

can mine Ethereum with them. uh but you

1:58

could also do all these other things and

1:59

really so right from the start we looked

2:02

at the compute as an option to be able

2:06

to deploy our computing power to

2:10

different use cases and so you know

2:12

began the company in 2017 uh you know um

2:15

spent the first kind of three years

2:19

mining crypto went through a couple of

2:21

crypto winters um because we had come

2:24

from a hedge fund were, you know, we we

2:27

have real chops in risk management and

2:29

how we think about uh capital and risk

2:32

exposure and allocation and all of that.

2:34

And so we were really careful around

2:35

that right from the start. So we

2:37

weathered crypto winter really well um

2:39

and began to scale the company and

2:41

immediately started to look for other

2:43

use cases that you could use this

2:45

compute for because crypto was pretty

2:47

volatile.

2:48

>> Yeah. And crypto was a question mark at

2:50

that time.

2:50

>> Absolutely.

2:51

>> Yeah. I mean Bitcoin was speculative and

2:53

there were many other specular projects.

2:55

the only other people using this type of

2:57

hardware quants

2:59

>> medical researchers.

3:01

>> So a good way to think about it is like

3:03

the progression of products that we kind

3:05

of started to work on. You know first

3:07

was uh um crypto but we immediately

3:10

moved from crypto to CGI rendering and

3:12

we built projects that would allow uh um

3:16

folks that were trying to animate and

3:18

render images um you know kind of what

3:22

makes the movies cool, right? and and uh

3:24

we started to work on that and then we

3:26

moved to batch computing and started to

3:27

look at medical research and different

3:29

ways of using the compute to be able to

3:31

drive science. Um, and we just kind of

3:33

kept moving up the stack in terms of

3:36

complexity uh uh on how GPUs could be

3:39

used. And ultimately uh in like call it

3:43

like 2020 2021 we started to really try

3:47

to figure out how you can go ahead and

3:49

use GPUs for neural networks and that

3:52

was not something that uh we knew how to

3:54

do. Um, and so we actually went out and

3:56

bought a bunch of A100s and donated them

3:59

to a uh a group that was working on uh a

4:02

Luther AI. They were working on an open-

4:04

source project with the thought that um

4:07

these guys are taking the GPU compute

4:11

because we're donating it. They can't

4:13

really get pissed at us if we're not

4:14

very good at it initially. And uh that

4:16

worked out really well because

4:17

>> they can't complain about the SLA.

4:18

>> They they kept telling us like we need

4:20

more of this, you got to work on this.

4:22

And that began to really uh uh give us

4:26

an understanding of what was necessary

4:28

to run scale parallelized computing. And

4:32

uh you know that that uh um we went

4:34

through it. I I I kind of feel like

4:35

buying those initial GPUs was the

4:37

tuition we paid to learn how to run this

4:40

business. And then one of the

4:41

interesting things is all of those guys

4:43

went back to their day jobs because they

4:45

were all volunteers working on this.

4:47

They were like-minded scientists. And

4:49

when they got to their day jobs, they

4:51

were all like, I want that

4:52

infrastructure. It's built the right

4:54

way. That's the way that researchers are

4:56

going to want to use it. And that

4:57

launched our our business. It was an

4:59

amazing story. And

5:00

>> so you went from crypto to these

5:03

researchers into academia and deep

5:05

research. What's the next card to turn

5:07

over in the poker game?

5:09

>> Yeah. So, so um what became very clear

5:11

to us very very early on was that the

5:15

scaling laws were going to drive and

5:18

remember this is really back in the you

5:20

know 2020 2021 before uh uh chatgpt

5:24

moment occurred and we began to

5:27

understand that like computing

5:29

decommoditizes at scale right like when

5:31

when you know anybody can run a GPU but

5:34

can you run a cluster that's large

5:35

enough to train a model that can change

5:37

the world and that's a different

5:38

question. And so we really began to

5:41

think about like how do you go about

5:44

scaling up your delivery of this

5:47

computing to clients, larger and larger

5:49

clients. And that was the next card to

5:51

turn is to think about it from a okay,

5:54

you know, there's a component of this

5:55

that is going to lean into uh our

5:58

ability to access the capital to be able

6:01

to deliver our solution to the broadest

6:04

possible audience to the most

6:05

sophisticated consumers of this compute.

6:07

And and that was really the next card is

6:10

thinking about it as a business rather

6:12

than as a engineering project to be able

6:14

to deliver the the uh uh the

6:17

infrastructure and the software and

6:19

really everything between you know when

6:21

you when you're thinking about what we

6:22

do, we kind of live above the Nvidia

6:25

GPUs but below the models. Yeah. and

6:28

everything in there, all the software,

6:30

the integration of software and

6:32

operations and uh observability and all

6:35

the things that you need to be able to

6:36

build uh a cloud that's purpose-built

6:40

for this one specific use case, right?

6:42

So, we don't we don't do everything. We

6:44

really focus on one use case which

6:46

allows

6:46

>> you want to do web servers different you

6:48

got AWS,

6:48

>> you know what they do a great job. It's

6:50

like it's a it's a great solution. It

6:52

was a brilliant solution to solve a

6:53

problem. We just looked at it and said

6:55

there's a new problem and let's go about

6:57

let's go about looking at this problem

6:59

and try and come up with a solution to

7:01

deliver compute that solves that

7:03

problem.

7:03

>> And when did the language model start

7:05

dialing and calling you for you know

7:08

capacity?

7:09

>> Yeah. So uh our our our first uh um well

7:13

our our first language model was really

7:15

a Luther. Um but uh our our first like

7:18

large commercial uh was inflection. Um

7:22

and so you know we work with Mustafa and

7:25

and and and Inflection and then we we

7:27

really diversified from there uh into

7:31

the hyperscalers into you know uh open

7:35

AAI across the the the the model uh the

7:38

foundation models across um you know and

7:41

and just kept scaling and scaling with

7:43

the belief that you know once again the

7:46

the the decommoditization

7:48

of compute the ability to to deliver a

7:52

solution and the solution is building

7:54

supercomputers that can change the world

7:57

and that's really what we began to focus

7:59

on. That was the lead into training and

8:01

now the world has gone through, you

8:03

know, this this moment where we've moved

8:05

from research into the productization of

8:09

this. It's it's it's beginning to work

8:11

its way in from the the uh the fringe of

8:14

organizations into the core of what they

8:16

do. And you can see that every day in

8:19

the uh in the amount of inference

8:21

compute that is being driven through you

8:24

know our uh infrastructure layer which

8:27

is just massive which is just like one

8:28

of the shows you people are consuming it

8:31

not just building models but they're

8:33

deploying them and and utilizing them.

8:35

>> I always think of inference as the

8:37

monetization

8:39

>> of the investment in artificial

8:41

intelligence. So when when when we when

8:44

we see our compute being used uh uh to

8:48

stand up the massive scale of inference

8:50

that's hitting our compute every day and

8:52

like you know inference is when people

8:54

ask the model a question it comes back

8:57

with an answer that's an inference or

8:58

when you ask the model a question and

9:00

then to go do something that's inference

9:02

right and that's actually where you're

9:04

you're you have the opportunity to

9:06

really drive value outside of the model

9:10

itself but into the real world and

9:13

that's really exciting for us. That's

9:15

what we like to watch. That's what I

9:17

like to watch in terms of gauging the

9:18

health.

9:19

>> What chips are those?

9:20

>> Um so so really uh you know we are we

9:26

are the tip of the spear in bringing um

9:28

the new architecture uh out of Nvidia uh

9:32

into uh um into commercial production at

9:35

scale. Yeah. And uh so when when you

9:38

know we were the first ones to bring the

9:39

H100s at scale, we were the first ones

9:41

to bring the H200s at scale, first ones

9:43

with the GB uh 200s, and now you've got

9:46

the GB300s. And one of the things that's

9:48

that's that's amazing and really

9:50

fascinating for us is is you know people

9:53

are using the bleeding edge GPUs to

9:55

train models as the new architectures

9:58

come out and then they take those GPUs

10:01

and they move them into different

10:03

experiments and then over time they move

10:06

them into inference and they continue to

10:09

use them in inference for a very very

10:10

long time.

10:11

>> What is the shelf life of a 100 right

10:13

now? That's been a big debate is I think

10:16

for your company for Microsoft and I

10:19

guess Michael Bur you know who you must

10:21

have known when you were a quant you

10:23

know saying oh my god the whole industry

10:25

is the sky's falling and then we all

10:27

know in the industry that people don't

10:29

just throw this hardware away that they

10:30

find uses for it the street finds its

10:32

own use for technology so what's the

10:34

reality of the lifespan of these things

10:36

>> so so my my take on the the uh uh the

10:40

GPU uh depreciation bait is that it's

10:43

nonsense Right. It's a debate that is

10:45

being uh brought to the forefront by uh

10:48

some traders that have a short position

10:50

in the stock and they're trying to uh

10:52

talk down. Look, here's what we know,

10:54

right? Um

10:57

when when we buy infrastructure, we're a

10:59

success based company, right? We're a

11:00

small company on a relative basis

11:02

compared to the enormous companies that

11:04

we're competing with. And so they come

11:07

our clients come into us and they buy

11:09

compute for five years, for six years.

11:11

Our average contract is 5 years. So any

11:15

commentary by anyone either inside or

11:18

outside of the industry that this stuff

11:19

becomes obsolete in 16 months or

11:21

whatever nonsense they're spewing, it's

11:23

it doesn't it doesn't in any way match

11:26

up with the facts on the ground. The

11:28

facts on the ground is they're buying it

11:29

for 5 years. Right? If and my approach

11:32

to this has always been if people are

11:34

willing to pay me for it,

11:36

>> it still has value.

11:38

>> Correct.

11:38

>> Pretty simple way of of approaching it.

11:40

We use a six-year depreciation. Um, we

11:43

believe that the GPUs will last in

11:45

excess of six years, but we felt like

11:47

that was a fair and reasonable approach

11:49

to a technology cycle that's moving at

11:51

this velocity. Um, the A100s, the ampers

11:55

this year, the price has appreciated

11:57

through the year.

11:58

>> And why is that? I I think it's because

12:00

one of the things that happens is as

12:03

more installed capacity becomes

12:06

available, you have new companies that

12:08

come into existence that have new use

12:09

cases that have different size models

12:11

that are trying to uh build new

12:14

commercial ventures that maybe have been

12:16

blocked out of the H100s and never had

12:19

an opportunity to run on that. I mean to

12:20

make a very simple example for the

12:22

audience like when you trade in your

12:24

iPhone after 3 or 4 years you're like

12:27

who's going to use an iPhone 12 and it's

12:29

like have you been to South America or

12:32

Africa where you go to the store and you

12:35

buy an iPhone 12 or you buy the Pixel 7

12:38

and it costs $50 that's still got great

12:41

life left in it.

12:42

>> Absolutely.

12:42

>> Yeah. you know,

12:43

>> and so look, you know, we we find these

12:46

amazing use cases, new companies that

12:48

have come into existence or existing

12:51

companies that have integrated new

12:53

models into their workflow that are able

12:55

to use the Ampierce and so they keep

12:59

buying any GPUs that we have available.

13:02

And once again, you know, the the

13:03

concept that a GPU

13:05

>> is no longer relevant or commercially

13:07

viable after 16 more 18 months or two

13:10

years.

13:10

>> Yeah, that's it just it just doesn't

13:12

make sense.

13:12

>> It's obviously far. I think sometimes

13:14

people get caught up in Moore's law or

13:16

in just how fast our industry is growing

13:19

and that there's so much at stake that

13:22

big companies are demanding the most

13:24

recent products. That doesn't mean that

13:27

the lifespan has gotten shorter. It

13:29

means the opportunity and the surface

13:31

area of the opportunity has gotten much

13:33

larger.

13:33

>> Yeah. Uh one of the things is is like

13:36

you know the the uh the the industry has

13:40

gotten so much attention for the

13:43

unprecedented scale of capital that is

13:46

coming to bear on this. And

13:49

because of that, there tends to be a

13:51

incredible focus on

13:54

the companies that are building on these

13:57

most advanced chipsets. And the truth of

14:00

the matter is is you know even within

14:03

those companies they have a long tale of

14:05

useful life

14:06

>> to provide inference horsepower to work

14:09

on other experiments to do less bleeding

14:12

edge activity but still needs to be done

14:14

>> and yeah I mean rendering comes to mind

14:16

as well or yeah we're making images on

14:19

nano banana like there there will be a

14:21

use for it. There is a moment in time

14:23

where maybe the compute to power ratio

14:25

doesn't make sense. My my expectation is

14:29

is obsolescence will be defined by the

14:34

moment in time where the power

14:37

in the data center for me will be able

14:41

to be repurposed for a higher margin

14:44

than the existing infrastructure

14:46

provides. And you know, like I said, I I

14:49

fully expect this infrastructure to last

14:51

in excess of 6 years, but the the the

14:54

standard in the in in in the space has

14:56

really been used with one exception,

14:59

which is Amazon, which is Yeah, it's 6

15:01

years. That's that seems like the right

15:02

schedule. I'm not making it up. That's

15:04

what everybody's using.

15:05

>> Yeah. And the energy cost is the

15:09

opportunity because hey, it's just a we

15:11

need that space. there's a better uh

15:14

reward here and that might get resold

15:16

that hardware to somebody else who wants

15:18

it a hobbyist or something. It's

15:20

available

15:21

>> and or it could be sent someplace else

15:22

where they have more capacity when they

15:25

can repurpose it there. But I I I um I

15:27

kind of feel like, you know, we'll we'll

15:29

deal with that part of the business when

15:31

we get there. What I know right now is

15:34

it is extraordinarily profitable. It's

15:36

very creative to my company to continue

15:39

to keep the infrastructure that's been

15:41

up and running, that's been on these

15:42

long-term contracts, and as it rolls

15:44

off, as it's been in use for 5 years,

15:47

you know, as it becomes available, I am

15:49

still able to sell it at a higher price

15:50

than it was at a year ago. There's

15:53

competition now. When you were buying

15:54

these from Jansen back in the day, yeah,

15:57

you could buy them and have them

15:58

shipped, I would assume, within 30 days

16:01

or less. nowadays what's the weight like

16:04

even for you a loyal old customer and is

16:07

there a bit of a battle is there

16:08

politics to who gets the servers like I

16:12

you see some like very big names talking

16:14

about they got to get an allocation is

16:16

it still a little bit crazy what's it

16:18

like to be in that category having to

16:20

buy something everybody wants

16:22

>> look uh you know I I uh I I think of it

16:25

as an affirmation of the business that

16:26

we're in right like the fact that we are

16:29

attracting competitors the the means

16:31

that the business is healthy and there's

16:33

a lot of people trying to deliver this

16:34

service because the need for this

16:38

infrastructure the need to integrate the

16:40

infrastructure you know into the

16:42

software layers to deliver it to

16:44

artificial intelligence uh either at the

16:46

model level or the inference level or

16:48

the application level or whatever you

16:50

know level of the five layer cake that

16:52

Jensen's you know focused on

16:55

the the fact that there are more people

16:57

coming into this it doesn't discourage

16:58

me um as far as getting access to the GP

17:01

CPUs, we show up like everybody else

17:03

with a um you know, we'd like to buy

17:06

here's a PO and we're ready to pay. Um

17:08

the one what's the wait time like? And

17:11

is it just really competitive or not?

17:15

Because I talked to Jensen about he said

17:17

I said, "How do you manage all these

17:18

like big egos and names and companies

17:20

trying to buy stuff?" And he said,

17:22

"Well, they order it and we give it to

17:24

them in the order in which they order

17:26

it."

17:26

>> Is it really like that?

17:27

>> It really is. Right. like you know he he

17:30

doesn't want to be in the position of

17:33

playing favorites or all like that that

17:35

just seems like a bad place to be with

17:37

your clients

17:37

>> or auctioning them off. Can you imagine?

17:40

>> Yeah, that would that that

17:41

>> that would be crazy.

17:42

>> Yeah.

17:43

I don't I'm not sure that would be good

17:45

for the long-term business. No. Yeah.

17:46

So, so our our our approach is

17:49

>> you might get some sovereigns coming in

17:50

and saying I'll pay double. They do that

17:52

with Ferraris too sometimes.

17:54

>> I guess these are the Ferraris of

17:56

computing, right?

17:56

>> In a way they are. Yeah. Bugattis. Our

17:59

>> our our approach is to work with

18:02

clients across the entire space to find

18:05

opportunities that are really

18:07

interesting companies that can fit into

18:09

our contraction contracting requirements

18:12

where we're going to be able to go out

18:14

and structure the debt that we require

18:16

in order to go out and and uh build

18:18

infrastructure at this scale. And um

18:21

>> how does all that debt work? I that is

18:23

something that you guys specialize in.

18:26

um corporate debt uh I'm in the venture

18:28

business people are like why should I be

18:30

in venture when corporate debt pays so

18:31

well corporate paper's so huge I'm

18:34

curious how this fits in and like what

18:37

uh interest rate people are paying on

18:40

you know a billion dollars in

18:43

infrastructure what do they pay on that

18:44

>> yeah so so coreweave has really been the

18:49

innovator around a lot of the financing

18:52

engines that have come to bear on this

18:54

we did the first GPU based uh loans. Um,

18:58

and like I I think it's important or I'm

19:01

going to try to explain this in a way

19:02

people can understand. So what we do is

19:05

we go out and we find a client. Let's

19:07

use Microsoft. You brought them up

19:09

before, right? And Microsoft comes to us

19:11

and says, "We'd like to buy some compute

19:12

for you." And we say, "Okay, great.

19:14

We're going to sign a contract." Once I

19:15

have a contract in hand,

19:17

>> then what I do is I create something.

19:19

It's not a particularly creative name.

19:21

It's called the box. Yeah.

19:22

>> Right. And what I do with the box is I

19:25

take my contract with Microsoft and I

19:27

put it in the box. I go to Jensen and I

19:29

buy the GPUs, I put it in the box. I

19:31

take my data center contract, I put it

19:34

in the box. And now the box governs cash

19:36

flow.

19:37

>> And it has a waterfall of cash flow that

19:39

comes into it and goes out of it. And so

19:41

the way it works is then I build the

19:44

compute and then I deliver the compute

19:46

to Microsoft and they pay the box. They

19:48

don't pay me,

19:49

>> right? It goes into the box and the

19:51

first thing it does is it pays the data

19:53

center. It pays the power bill. It pays

19:56

the interest and the principal and then

19:59

whatever's left flows back to us, right?

20:02

And so it is an incredibly well

20:05

ststructured, time-tested,

20:07

pressure-etested vehicle to be able to

20:09

borrow money against client paper and

20:13

all of the other collateral around the

20:16

deal. which is why Corewave, which is a

20:18

company that many people haven't ever

20:20

heard of, was able to go out and raise

20:22

$35 billion in 18 months to build

20:25

infrastructure at scale. But what's

20:27

important to understand is the economics

20:30

in this box are such that within 2.5

20:33

years of a 5-year deal, we have paid for

20:37

everything.

20:38

>> The principal's been paid off. The well

20:40

the principal's been paid off, the

20:42

interest has been paid off. The return

20:43

into the box is such that we are able to

20:47

generate returns to our company at the

20:50

box level which gives the most

20:53

sophisticated lenders in the world

20:56

whether it's banks or private equity

20:59

funds or um you know whoever. confidence

21:03

that they're going to

21:06

be able to achieve the one rule of

21:08

lending, which is give me my money back.

21:10

>> Yes. Works better when that happens.

21:12

>> So, they look at this box and they're

21:13

like, "Wow, we're really confident we're

21:15

going to get our money back."

21:16

>> And maybe they want 10 boxes.

21:17

>> That's correct.

21:18

>> And if any one box um goes upside down,

21:22

you can deal with it and it's not as

21:24

acute.

21:25

>> That's correct. And they don't

21:26

cross-pollinate. They don't cause uh

21:29

contagion across the boxes. are all

21:31

independent and discreet. One, and

21:33

number two is as you do this and as you

21:36

show the lenders how this financing tool

21:40

and how this financing mechanism works,

21:43

what they do is they continue to lend

21:45

you money at progressively lower rates.

21:48

And so when you think about our cost of

21:51

capital over the last two years, we have

21:55

dropped our cost of capital by 600 basis

21:57

points.

21:58

>> Wow. It is enormous, right? And so

22:00

you're seeing a company that is driving

22:02

its cost of capital down towards where

22:05

the hyperscalers borrow, which will

22:08

enable us to be able to be competitive

22:10

with them over time. And we have been

22:14

extremely

22:15

uh militant and diligent about feeding,

22:19

watering, and caring for those boxes so

22:21

that we continue to have access to the

22:23

capital markets in a way that allows us

22:25

to build and drive our business.

22:26

>> Means you has to say no. You have to say

22:28

no to maybe some people who want to be

22:29

in the box.

22:30

>> Yeah. So, we we look at some deals and

22:33

we're just like, you know, they want to

22:35

buy GPUs for a year and I look at it and

22:37

say I I that's not a deal that I can do

22:39

because it's too short for me to am

22:41

advertise the expenses or and so I won't

22:44

do that. Right. Like once

22:46

>> and they can go to another provider who

22:47

maybe wants to take that risk on who has

22:49

extra capacity.

22:50

>> Absolutely. But our business is really

22:52

built about around the risk management

22:54

of being able to get to scale. Because

22:56

in my mind

22:59

during this period of disequilibrium

23:01

during this period where there are not

23:03

enough GPUs in the world to uh provide

23:06

the compute for all of the different use

23:09

cases in artificial intelligence the

23:11

part that's important for me and for my

23:13

company is to get enormously large so we

23:16

can drive down our cost of capital so

23:18

that we have information flow coming in

23:20

from all different parts of the market.

23:22

large language models, high-speed

23:24

trading, uh, uh, search, all of these

23:27

things. And they're feeding they're

23:28

feeding information back into us that is

23:31

letting us know what the next product we

23:33

need to build is or where, you know,

23:35

they need help uh, scaling or what type

23:38

of compute they need and all of that

23:40

information flow is incredibly valuable

23:43

to us.

23:44

>> What What can you tell us about demand?

23:45

There's been reports of, hey, maybe the

23:48

Oracle Starbase thing with OpenAI's been

23:52

downsized or maybe not and then you know

23:56

uh other folks Microsoft is going big

23:58

and Google's going big Meta's going big

24:01

and those people obviously have massive

24:03

cash flow Apple seems to be MIA they

24:05

don't seem to want to play you you

24:07

you've you've uh you've named a lot of

24:09

really big companies with really big

24:10

balance sheets that have the capacity to

24:12

drive a lot of demand look I I have been

24:15

truly steadfast in this

24:18

>> for years now for for for four

24:21

The depth of the demand for the service

24:23

we provide has been relentless and

24:27

overwhelms the global capacity of the

24:30

world to deliver enough compute to

24:33

enable all of the demand for artificial

24:37

intelligence to be sated and that has

24:40

been we have been relentless about that.

24:42

>> Sounds like Nick's tickets during the

24:44

Patrick Euing era like they got up to

24:46

50,000 people on the wait list. So if

24:49

magically the weight list went away, if

24:52

the if the constraint went away and we

24:53

just had a large amount of GPUs

24:56

available, lot of energy available, a

24:59

lot of data center available, how much

25:03

capacity would just all of a sudden come

25:05

out of the system.

25:05

>> So so or would be deployed I should say.

25:08

>> So remember how we build our our

25:11

business through this box

25:12

>> and it's a fiveyear box. So if we had an

25:16

air pocket, if if demand were suddenly

25:19

to disappear because of a technology

25:21

breakthrough, because of a uh a war,

25:24

anything, right? Like like the why from

25:26

a riskmanagement perspective does not

25:28

matter. You have to prepare your company

25:31

for the what happens if it happens.

25:33

Yeah. And so by entering into these

25:36

long-term contracts into entering into

25:38

contracts with counterparties that have

25:40

large balance sheets, you are or we are

25:44

protecting ourselves and our lenders.

25:47

Yeah.

25:47

>> So that we are confident and they are

25:50

confident because you can see how

25:51

confident they are by the rate that

25:53

they're charging us continuing to

25:54

decline that they're ultimately going to

25:56

get their money back. And that is the

25:58

one rule of lending. And so um you know

26:02

I if

26:02

>> but just in terms of the capacity if you

26:04

were unconstrained and Nvidia Jensen

26:06

says hey order as many as you want what

26:08

would happen

26:09

>> so um the the it's also important to

26:13

understand the constraints aren't just

26:14

GPUs right electricity it's it's power

26:17

shells it's memory it's storage it's

26:20

it's networking it's optics all of the

26:23

things and there's there's various

26:24

throttles that will limit the

26:27

>> memory is a throttle right now right

26:28

>> oh yeah it Oh yeah, it is.

26:30

>> Why? How did memory become the throttle?

26:32

>> If um

26:34

memory and uh it has historically been a

26:39

cyclical business, right? We have seen

26:41

these waves of demand driving up the

26:44

cost for memory and then it collapses

26:46

and then it drives it up. It's a very

26:48

boom and bust business. is cyclical in

26:50

its nature because the fabs are so

26:53

capital intensive that people invest in

26:55

the fabs, build a ton of capacity and

26:58

then overbuild if there's any type of

27:02

turndown. And that we've seen that cycle

27:04

again and again. What's happening right

27:06

now is the confluence of two things,

27:09

right? one is is

27:11

with all the demand for artificial

27:13

intelligence and the corresponding

27:15

demand for compute and the ancillary

27:18

services around the GPU, the demand is

27:21

through the roof. That's number one.

27:23

Number two is is that

27:25

>> there was probably an investment cycle

27:27

that needed to happen back in 2023

27:30

that would have brought on the necessary

27:32

fab capacity to be able to serve.

27:35

>> Impossible to predict what should happen

27:37

just with energy. It's impossible to

27:38

predict what just happened. And now

27:40

people are chasing energy. The data

27:42

centers are going where the energy is.

27:44

It's not based on real estate. It's

27:46

based on it's and

27:47

>> where's there's some wind.

27:48

>> And anytime you you have a uh very cap

27:52

not every any time, but many times when

27:54

you have a uh a capital inensive

27:56

business like you know building fabs,

27:58

you will get this boom and bust cycle

28:00

just like in energy they overbuild.

28:02

Yeah. And you know

28:03

>> fiber.

28:04

>> Yeah. I mean there's there's there's a

28:06

lot of examples of that our approach

28:09

>> in some ways when you look at that it's

28:11

a beautiful aspect of capitalism that

28:13

we're able to have a boom bus cycle that

28:17

we're able to weather it right if you

28:18

think just that capitalism from first

28:20

principles something like that happens

28:22

and we have too much fiber it creates an

28:24

opportunity for Google to buy it all up

28:26

or the next person

28:27

>> listen the the the um um you know it it

28:30

does it does a lot of things having a

28:32

boom cycle it clears out the underbrush.

28:35

will be able to survive and take

28:37

advantage of that and it sews the seeds

28:39

of future business.

28:45

You put the fiber into the ground which

28:47

became the backbone of how you know we

28:51

watch movies every day and how we you

28:53

know uh communicate and how we hop on a

28:55

zoom and you know co and all of these

28:58

things were based on that infrastructure

29:01

that was available to be consumed. Yeah,

29:04

people don't recognize this fact if you

29:06

the the premise of YouTube from the

29:08

founders who I knew, Chad Hurley and his

29:12

other partner. They basically had the

29:14

realization at this curve storage is

29:16

coming down so quickly we could offer

29:18

free unlimited uploads and bandwidth is

29:21

coming down. So I guess we don't have to

29:23

charge people for sharing a video

29:25

online. Before that, if your video went

29:28

viral, people are going to have their

29:30

minds blown. But your server would turn

29:33

off and it would say this person, you

29:35

know, needs to pay their bill. Yes.

29:36

Because they were getting charged for

29:38

carriage by the megabit going out.

29:41

>> Yes. I mean it look and and you know

29:43

these these the business models change

29:45

and evolve and you know like you said

29:48

Moore's law and and and certainly Jensen

29:50

will talk about the fact that like what

29:52

what is going on within the the the

29:55

accelerated comput dwarfs

29:58

>> Moore's law right and all of that is

30:00

going to lead to

30:03

>> more opportunity to build more companies

30:05

that are going to do things like you two

30:07

did which has really changed the world.

30:09

>> Yeah. I mean the the concept that I I

30:12

don't know if it was like a million

30:14

hours being uploaded every hour or

30:16

minute but at some point Susan what

30:18

Jackie rest in peace said told me just

30:21

like how much was being uploaded every

30:25

minute and it made no logical sense and

30:27

she realized

30:28

>> well there's three billion people two or

30:31

three billion people in the service and

30:32

1% upload or 0.1 10 bit bips upload and

30:36

it's like okay one in a thousand people

30:37

upload it's a big it's a big denom

30:39

denominator like

30:40

>> I I was uh sitting on a a panel uh with

30:43

Sarah Frier, CFO of uh Open AAI and u um

30:50

she every once in a while uh um she she

30:53

really puts out like interesting uh

30:55

information and so she was talking about

30:58

the cost of a million tokens when ChatG3

31:01

came out and it was $32 and change and

31:05

now a million tokens cost nine cents.

31:08

>> Yeah.

31:08

>> Right. And so you you just see like like

31:11

the incredible power of how the capital

31:15

markets, how capitalism is

31:19

uh uh fueling engineering and fueling uh

31:22

uh competition.

31:23

>> It's become recursive now too. I mean

31:25

these models if you say to the model,

31:27

hey make yourself more efficient, spend

31:29

less money and lower the cost of tokens.

31:30

It'll be like okay captain.

31:32

>> Yeah.

31:32

>> I don't know if you saw Cararpathy's

31:34

recursive

31:36

>> thing last weekend but it's like now

31:38

civilians who've never worked in a

31:40

language model or done computer science

31:41

are like, I'm going to try to do

31:42

something recursive this weekend. You

31:44

know, it's one of the things that I that

31:47

uh uh I talked to, you know, the other

31:51

founders about, you know, and it's like

31:55

when you think about some of the things

31:57

that AI does, right, it's lowering the

32:00

barrier to operations. So if you have a

32:03

good idea or a great idea, you can open

32:06

up your model and you can tell your

32:10

model, you can vibe code it, you can do

32:11

all kinds of different things and create

32:14

things that never existed before. That's

32:16

amazing, right? like that's bringing

32:18

down this incredible barrier that kept

32:20

human creativity contained and now all

32:23

of a sudden this whole new vector of uh

32:26

uh you know medical research or

32:28

different approaches to you know

32:31

baseball cards or whatever you want if

32:33

you've got a great idea if you've got a

32:34

new creative idea that's the valuable

32:37

kernel right now that allows you to to

32:40

build new things and to create new

32:41

things and I just think that's

32:42

incredibly exciting like you're bringing

32:45

the minds of 8 billion in people a tool

32:48

that allows them to overcome what was

32:50

insurmountable for

32:52

forever

32:53

>> for humanity.

32:54

>> Yeah, it's a bright new future. Michael,

32:56

appreciate you sharing the uh uh

32:58

information with us and the vision. I am

33:01

really delighted to have Arvin Shri Nas

33:04

on the program.

33:05

>> Thank you for having me here Jason.

33:07

>> It's so great. I want to go through

33:09

three stages in which I fell in love

33:11

with your product. The first phase was I

33:15

could go in pick my language model if I

33:18

wanted to choose open AI, if I wanted to

33:19

use claude, whatever it was. That was

33:22

like a real unlock for me. And on the

33:25

sidebar sidebar, I noticed you had done

33:29

essentially like what Yahoo did in the

33:32

early days, finance, sports, and when I

33:36

pulled my nickname up, it gave me a live

33:39

version of that. When I pulled my stocks

33:40

up, it summarized the news in real time.

33:42

time and I was like, "Wow, this

33:44

execution's great." And I I kind of made

33:46

you my front door, two different models,

33:48

and it made it easier for me to check

33:50

it. Then you came out with the Comet

33:52

browser and I was like, "Holy cow, I can

33:55

give this a series of instructions. Go

33:56

to my LinkedIn, find everybody from this

33:59

company, put them into a Google sheet

34:01

and boom, you were the first out of the

34:02

gate with that." And then just the last

34:04

couple of weeks I had been claw pilled

34:07

in using openclaw but you came out with

34:09

computer and I started using computer

34:12

and boy it's good uh it's a really

34:15

strong start uh allowing me to do

34:18

repetitive tasks very similar in some

34:21

ways to co-work from claude uh or

34:24

basically an engineer or developer using

34:26

it. So

34:28

are are these the evolution of the

34:31

company and I should think about it that

34:32

way. But how do you look at perplexity

34:34

now? You have a very loyal fan base.

34:37

You're making a lot of money. I don't

34:38

know if you disclose it but I think it's

34:40

hundreds of millions to billions. You

34:42

can tell us but what is perplexity in

34:44

the face of wow Claude's having a great

34:47

run, OpenAI still doing strong. Grock

34:49

doing very well. Gemini coming on

34:51

strong. There's like six or seven of you

34:53

and uh you just happen to be one of my

34:55

top twos right now.

34:57

>> Thank you. So tell me first of all,

34:58

first of all, thank you. Thank you so

34:59

much. Perplexity has always been built

35:01

for people who are always looking for

35:04

the extra edge, the curious people. So

35:06

it's very natural that you are uh one of

35:09

our power users. Uh one common theme for

35:13

us uh for the last three and a half

35:16

years is accuracy. Plexity wants to be

35:19

the company that's building the most

35:21

accurate AI. So when you want to give

35:23

somebody answers, accuracy is very

35:25

essential for building trust because

35:27

only then the user is going to ask the

35:29

next set of questions. It turns out it

35:31

was a great idea to give AI access to

35:34

the internet to be accurate. So that's

35:36

the perplexity ask product. It turns out

35:38

it's a great idea for AI to have full

35:40

access to a browser so that it can be

35:43

accurate when you task it to go do

35:45

something that you would do yourself on

35:46

a browser. Aentic browsing comet. Now

35:50

the last phase is it turns out it's a

35:52

great idea for AI to be given a full

35:55

access to a computer so that it can do

35:58

whatever you do on a computer on its own

36:01

essentially becoming the computer

36:03

itself. an orchestra of everything AI

36:07

can do today. every single capability

36:09

each individual AI model has be it GPT

36:12

or cloud or Gemini or anything else an

36:16

orchestra of all those capabilities that

36:18

what that's what perplexity computer is

36:20

and all these sub agents that are

36:23

running inside computer are the

36:24

musicians the models are essentially the

36:27

instruments and they're like hundreds of

36:30

models out there each having their own

36:32

specialization some are good at coding

36:34

some are good at writing some are good

36:35

at multimodal visual synthesis is image

36:38

generation, video generation, audio, but

36:40

what matters is the end output, the

36:42

music you play. That's the work AI gets

36:45

done for you. And that's what perplexity

36:47

computers. The AI itself is the

36:50

computer. Now,

36:51

>> still lives inside of a browser. Have

36:53

you considered giving it desktop root

36:56

access? That feels like the next place

36:57

this is going, but that comes with a lot

37:00

of security issues, a lot of trust

37:02

issues. As you mentioned, trust is

37:03

paramount. getting the right answer is

37:05

what builds it, but also not getting

37:07

hacked and not having it delete your

37:09

files. So, how do you think about root

37:12

access to my Windows machine? Obviously,

37:14

iOS, they won't let you, but with an

37:16

Android phone, it would let you.

37:18

>> Yes.

37:18

>> So, do you have that in the works?

37:20

>> Yes. So, we announced something called

37:21

personal computer. Perplexity personal

37:24

computer that's essentially going to

37:26

take all the trust and reliability and

37:28

the server side execution of perplexity

37:30

computer but synchronize it with your

37:33

local computer so that you can use it

37:36

from your phone and we're going to do

37:37

this with the Mac Mini where you

37:40

synchronize your computer with the Mac

37:41

Mini so that becomes your local server

37:44

all the agent orchestration that has to

37:46

do with your local private data will run

37:48

on that local orchestration loop that

37:50

runtime with the Mac Mini. Not on your

37:53

servers, not on anthropics.

37:55

>> Exactly.

37:55

>> Yeah.

37:56

>> It could still ping Frontier models if

37:58

it needs to with your permission,

38:01

>> but it will be orchestrating everything

38:03

on your local hardware.

38:05

>> Yeah.

38:05

>> And if it needs to run on the server

38:07

side hardware, if you don't want very

38:09

complicated, longunning stask to be

38:12

running on your local hardware. Yeah.

38:14

>> You can delegate it to run on your

38:16

server side computer, which is again

38:18

only accessible to you and you alone. So

38:21

that way we're going to bring this

38:23

perfect hybrid of trustworthy

38:25

uh hybrid between local and server side

38:28

and you

38:29

>> and you'll make it easy to do. It just

38:30

be abstracted. You install one

38:32

executable, boom, it's done.

38:33

>> It's it's like open claw for dummies.

38:35

Nobody needs to learn how to use it.

38:37

Nobody needs to manage API keys. Nobody

38:39

needs to manage separate billing across

38:40

like 100 different services. Figure out

38:43

what you can give access to and not

38:45

access to. We take care of that.

38:47

>> So it's a Steve Jobs way of doing it.

38:49

you know, end to end integration

38:51

>> and and how do you think about local

38:53

models? I have started running Kimmy 2.5

38:56

on a Mac Studio.

38:58

>> It's not as good as Claude or Gemini or

39:00

Grock, but you can probably do about 80%

39:03

there for free.

39:04

>> Yeah.

39:05

>> Essentially.

39:06

>> Yeah.

39:06

>> Uh and so that's quite compelling

39:08

considering some of my other bills,

39:09

Claude and and stuff were getting

39:11

expensive.

39:12

>> So, do you have one of those? You

39:15

started testing on your local Mac

39:17

Studio. I assume you have a Mac Studio

39:18

and you're doing this yourself. Yeah.

39:20

>> Or now, I don't know if you saw uh Dell

39:22

and Nvidia announced a giant

39:24

workstation. Um is it 3,800?

39:28

>> Something like that.

39:28

>> Something like that with 750 gigs of

39:30

RAM. So,

39:32

>> what do you think about the desktop

39:34

going back to workstation/server?

39:37

>> Yeah.

39:37

>> Status.

39:38

>> I think it's very promising. Um my my

39:41

prediction is it'll initially start off

39:43

as a sub agent. So whatever you need to

39:46

go uh like your tax returns, your

39:49

personal photos, your emails, your your

39:52

calendar, all that stuff, those local

39:55

apps, your personal notes, very personal

39:58

notes. You can make sure that the models

40:00

that access those tokens will be running

40:04

on your local hardware if you want to,

40:06

if you're that privacy conscious.

40:09

uh and more complicated stuff that

40:12

accesses your data that's already on the

40:14

server side. Example, your Google

40:15

calendar, yeah, your Gmail. This is

40:18

personal data still, but an AI runtime

40:22

can access that through your connector,

40:24

your Google calendar connector, your

40:26

Google Workspace connector, and that

40:28

could run on the server side because

40:29

anyway, the data is on the servers. It's

40:31

not even lying on your device.

40:32

>> So, that sort of hybrid orchestration is

40:35

where we're headed to. I don't think

40:37

it's a dichotomy between fully local

40:39

versus fully server. Uh it's all about

40:42

choice. And anyway, when you're on your

40:44

phone, uh you want to you don't care

40:46

actually which server that workloads

40:48

running from because it's not going to

40:49

be able to run on your phone anyway. The

40:52

chips need to exist on a Mac Studio or a

40:54

Mac Mini and or on the server

40:56

>> or this new Dell that's coming out. And

40:58

I I really think the idea of spending

41:01

$10,000 on a powerful desktop will

41:04

appeal to people if it lowers their $500

41:08

a month

41:09

>> claude bill. This is an incredible

41:11

savings. Plus, you get the benefit

41:13

>> of privacy and not educating the

41:15

language models on your personal data.

41:18

>> Yes. And it's going to be it's going to

41:19

be like you're buying a refrigerator,

41:21

your your your internet modem. Like the

41:23

cost for these will eventually go down.

41:26

>> Yeah. But it's not going to feel like

41:28

you're wasting your money. Uh every

41:31

every home has a lot of other sensors.

41:34

>> Yeah.

41:34

>> That runs your home that'll also be part

41:37

of this orchestration loop.

41:39

>> Yeah.

41:40

>> So, so that's where it gets exciting

41:41

because now you can just dictate

41:43

something to your phone and that can

41:45

control your entire home.

41:48

>> So that's the dream that everybody has

41:49

and all that orchestration loop can run

41:52

on your local hardware, no problem. And

41:54

I'm curious what you think of the

41:56

operating system. What's eventually

41:59

going to be the operating system of this

42:02

workstation?

42:03

>> AI is the operating system. Like earlier

42:05

in the traditional operating system, you

42:07

execute programmatically.

42:09

Now you start with objectives, not

42:11

specific instructions.

42:13

>> Right?

42:13

>> You come up with a highlevel objective.

42:15

go build this website for me that you

42:18

know takes all the transcripts of all in

42:20

podcast and tracks the stock price just

42:22

before the podcast and after. Yeah.

42:24

>> And charted for the max 7.

42:26

>> Yeah.

42:26

>> And and charted over time you can so

42:28

that's the objective but individually

42:31

it's running a file system a code

42:33

sandbox access to the internet. It's

42:35

having like its own HTML tools and like

42:38

so I think that's basically where you

42:40

know models systems and files and

42:43

connectors are all coming together. You

42:44

would think of that as an OS

42:46

>> except you're operating at an

42:48

abstraction about that where you're

42:50

thinking in terms of objectives.

42:52

>> Yeah. And does it need to eventually

42:55

become its own operating system in your

42:57

mind?

42:58

>> It could be like people could think

43:00

about like yeah I have a my perplexity

43:02

computer running all the time whether it

43:05

essentially it runs on Linux machines

43:07

right now. Every server side computer is

43:09

a Linux machine. Yeah.

43:11

>> So, I think Mark Anderson tweeted this

43:13

right after our release that turns out

43:15

Linux computers was the right idea.

43:17

Desktop desktop Linux computers are

43:19

finally going to work.

43:20

>> Yeah. I mean, they're stable. They're

43:22

customizable and you're not at the mercy

43:25

of Apple's desire to contain the

43:28

experience or Microsoft surface area as

43:32

for hackers.

43:33

>> Exactly.

43:34

>> You build something rock solid and it

43:36

does feel like Linux might actually

43:38

become the correct

43:39

>> the eventual winner. It may not need to

43:41

have a front end.

43:42

>> That's the thing. You could you could

43:44

access the Linux machine on your phone.

43:47

>> You could be running iOS or Android. It

43:49

doesn't matter.

43:50

>> The actual valuable runtime is running

43:53

on Linux on the server.

43:55

>> You've done great as a consumer company.

43:58

Lot of love there. Now I'm starting to

44:00

see uh corporations with computer

44:04

starting engaging it. In fact, you'll be

44:06

happy to know this. Last week, I took

44:08

two people in my back office and I said,

44:10

"Stop working on OpenClaw. Your job is

44:13

to do the back office automation at our

44:16

venture firm only using Perplexity." And

44:19

they were like, "Perplexity computer."

44:21

And they were like, "Oh, okay. Um, it

44:24

doesn't talk well in Slack. It doesn't

44:26

have an agent in Slack." I was like, "It

44:28

will. I'm going to see AR and I'll talk

44:31

to him about that." So, we need a really

44:33

strong Slack connector.

44:34

>> It's already out.

44:34

>> It is. Okay, great. computer exists as a

44:37

Slackbot right now.

44:38

>> Okay,

44:38

>> that you can add to your Slack workspace

44:40

on enterprise plan

44:41

>> and our entire company works like that.

44:44

People are talking more to computer on

44:45

Slack to other than to other people.

44:47

>> In our first volley, we were sending

44:49

reports in, but it wasn't interactive.

44:51

That's perfect.

44:54

So now you've got your company going in

44:56

two different directions. This

44:57

incredible consumer run you have. How

44:59

many people are using the product every

45:01

month?

45:01

>> Several tens of millions. So tens of

45:03

millions of people that's very much

45:05

similar to the trajectory of the Google

45:07

and Yahoo consumer business. Now you've

45:09

got corporate. How are you doing on the

45:11

corporate side? Thousands of companies.

45:13

>> The fastest growing business for us. Ah

45:15

>> it's growing faster than the consumer in

45:17

revenue and things like computer unlock

45:20

entirely new possibilities. For example,

45:22

we've saved more than $und00 million for

45:25

our uh enterprise max customers who are

45:27

on the highest tier of enterprise.

45:29

>> Explain what that is. What does it cost?

45:31

200 a month per person.

45:32

>> So there are two tiers. One is the

45:34

enterprise pro which is $40 a month and

45:36

there's the enterprise max which is $400

45:39

a month. And that that and and and and

45:42

on a computer after you run out of your

45:44

credits you would pay for the tokens.

45:46

You pay for the usage.

45:48

>> Are you making money on the $400 a

45:50

month, $5,000 a year one or at this

45:52

point in time are people going so crazy?

45:54

Our uh one thing that Perplexity has is

45:57

every revenue we make, unlike certain

45:59

other rapper companies, every revenue

46:01

Perplexity makes has positive gross

46:02

margins.

46:03

>> Got it.

46:04

>> Because uh we're not just selling

46:06

tokens,

46:06

>> right?

46:07

>> Most of our revenue is recurring because

46:09

people are paying a subscription fee

46:11

>> and because we route through multiple

46:13

different models, we're very efficient

46:15

in terms of how we spend on the tokens.

46:18

because we have all this advantage with

46:19

rag and orchestration and search. We

46:22

don't actually need to blow up the

46:23

context window of the models.

46:25

>> Yeah.

46:25

>> As a result of that, we have positive

46:28

gross margins on all the revenue. Every

46:30

single penny we make, we make profits on

46:32

that. But the overall the company is

46:34

still yet to be profitable, but we're

46:35

working towards that.

46:36

>> You've had the opportunity to exit. A

46:39

lot of rumors, Apple, other people were

46:41

like, "Hey, this is a great team." How

46:42

many people on the team now?

46:44

>> About 400.

46:45

>> Yeah. You you've got a very coveted

46:46

team. You obviously understand consumer.

46:48

You obviously understand business. It's

46:50

a product driven organization. Reports

46:52

are you declined,

46:55

but the world's getting hyper

46:56

competitive here. How do you keep up as

46:58

a 400 person organization when you got

47:01

Sam Alman over here raising a hundred

47:04

billion dollars, you know, and then you

47:06

have Elon putting data centers in space

47:08

and merging with SpaceX and Twitter. You

47:11

have Google with unlimited resources.

47:13

Amazon getting in the game and obviously

47:16

Gemini uh very strong product and Google

47:20

really good at consumer. I think we'd

47:22

all agree Facebook and Meta haven't

47:25

figured it out yet except maybe for

47:27

serving us better ads, but they they

47:28

haven't figured out the consumer case

47:30

yet, but they'll copy it. They always

47:31

do.

47:32

How do you look at the playing field?

47:34

Because the degree of difficulty, this

47:37

isn't playing checkers or this is like

47:40

playing against the 10 best chess

47:43

players in the world. That's what you

47:44

have to do every day.

47:46

>> So, how do you think about it? Long-term

47:47

and independent company. Do you think

47:49

you'll need to join forces at some

47:51

point?

47:51

>> Well, and why didn't you take the deal?

47:53

This deals were incredible that you got

47:54

offered.

47:55

>> So, one advantage we have that all these

47:59

companies you mentioned don't have is

48:01

the multimodel orchestration. We're like

48:04

Switzerland. We don't have to have one

48:06

horse in the race. If GPT wins, Gemini

48:09

wins, Claude wins, Llama wins, it

48:12

doesn't matter to us. Uh or even open

48:15

source models can win, no problem.

48:16

>> And you have them on the service. You

48:18

have DeepSeek and Kimmy.

48:20

>> We have Kimmy, we have Neotron, and we

48:22

have uh a lot of usage of Quen, Alibaba

48:25

Quen.

48:26

>> Yeah.

48:26

>> Silently under the hood. So for us like

48:29

that advantage of being able to take the

48:31

best in each model and give the user the

48:35

orchestra of everything they can do. I

48:37

don't think any of the companies you

48:38

mentioned can do that

48:39

>> right nor would they

48:40

>> nor would they it makes no sense for

48:42

them. It would be an admission that all

48:44

the data centers and capex they've built

48:46

out mean still couldn't produce them the

48:49

best model. And uh Daario uh CEO of

48:52

Anthropic said recently in an interview

48:54

uh that models are specializing. Towards

48:58

the beginning of last year people

48:59

thought models are going to commoditize

49:01

but towards the end of last year people

49:04

models started specializing. Even within

49:06

coding

49:08

u cloud code and codeex have very

49:11

different capabilities. Our iOS

49:13

engineers love using codeex. Our backend

49:16

engineers love using cloud code.

49:17

>> Yeah. So even within a specialization

49:19

like coding, models have their own

49:22

unique specialtities and there are many

49:24

other use cases outside coding where

49:26

different models are good at different

49:27

things. Which means the orchestra

49:29

conductor that has no one model to the

49:32

horse in the race can win by providing a

49:35

very unique value and service to the

49:37

customer that each of these amazing

49:40

names that you mentioned cannot. And so

49:42

you're buying tokens wholesale from them

49:45

and then you'll charge customers to do

49:48

it or do you think it's all

49:50

>> we're going to take care of all that

49:51

orchestration?

49:52

>> Yeah.

49:52

>> So you don't have to manage tokens

49:54

across different models

49:55

>> cuz I authenticate I a couple of my

49:58

different accounts my pro accounts into

49:59

perplexity. But does it I I I don't have

50:02

enough knowledge to know if you're

50:04

abstracting that and people can just

50:05

search across them and it's part of

50:07

their perplexity subscription. No, we're

50:09

not bundling subscriptions from into

50:11

other AIS.

50:12

>> We just ping the models directly.

50:14

>> Got it.

50:14

>> Uh what you get in us is the perplexity

50:17

or orchestration.

50:18

>> Got it.

50:18

>> The harness,

50:19

>> right?

50:20

>> So the when when when when models are

50:22

kind of specializing the there's a

50:25

bigger value in the one who knows how to

50:27

build a great harness,

50:28

>> right?

50:28

>> That can take the best in each model.

50:30

>> Does it auto route today or do you still

50:32

have the drop down somebody's got to

50:34

pick

50:34

>> it? It definitely auto routes the best

50:36

model for each prompt,

50:37

>> but we also give users the flexibility

50:39

to pick whatever model they want.

50:41

>> What do you think of I've seen a bunch

50:43

of startups hack this together, but

50:45

doing the same query across multiple

50:47

>> We built a thing called model council.

50:49

>> Model council. Yeah.

50:50

>> Yeah. So that's one of the one of the

50:52

modes and perplexity where I saw Jensen

50:54

say in one of the interviews that he he

50:57

puts the same prompt in five different

50:58

AIs and sees what each of them says.

51:00

>> Yes.

51:01

>> Like everybody does that. Yeah. But then

51:03

you still have to apply your biological

51:05

computers

51:10

about your trust or your

51:12

>> five different doctors.

51:13

>> Five different doctors trying to figure

51:14

it out.

51:15

>> Exactly. So it's dumb.

51:16

>> So the model council is a feature we

51:18

built where it will not just give you

51:20

the answers of each model, but it will

51:22

tell you exactly where they agree, where

51:23

they disagree, and where the nuances

51:24

are.

51:25

>> And that's in the interface. Model

51:26

council, I didn't know it was there.

51:27

>> It's there.

51:28

>> I mean, you you released product at a

51:30

pretty great cadence, huh? How where did

51:32

you learn that and what's your

51:34

philosophy of shipping product?

51:36

>> Our philosophy is like speed is our

51:38

mode. Like you know again one of the

51:40

things that big companies cannot do is

51:41

move at the speed we do serve customers

51:43

at the speed and qual it's it's very

51:45

hard to maintain quality speed and trust

51:47

at the same time.

51:48

>> Yeah.

51:49

>> Like Apple takes a long time to ship

51:51

anything

51:51

>> because they're very worried about

51:53

people not trusting them.

51:54

>> Yeah.

51:54

>> Uh and so some companies are

51:56

bureaucratic and they just take forever

51:58

to ship something. They don't maintain

52:00

what they ship. They may make a big deal

52:01

about an event but nobody even knows how

52:04

to go and use that feature.

52:05

>> Yeah. They get abandoned.

52:06

>> Exactly. So, Perplexity has those

52:08

advantages for being very small. And

52:10

towards the end of last year, we found

52:12

that like AI coding tools have made it

52:14

much faster for us to ship things

52:16

>> which is honestly one of the reasons why

52:18

we built computer because now even

52:20

non-engineers are shipping code here by

52:22

just pinging a slack bot and asking it

52:24

to fix bugs.

52:25

>> So, this the the iteration has just been

52:27

like exponential. The the moment I had

52:30

where I became clawilled was when I was

52:34

working with it and I was like, "Hey, I

52:36

want to build my network. I know these

52:38

20 people in Japan. I had dinner with

52:39

them during my recent trip. I want to

52:41

know who they know. So, check out

52:43

LinkedIn and other things and who

52:45

they're associated with and make me like

52:46

a mind map of it. And then the next trip

52:49

I want to meet with the next circle of,

52:52

you know, those connections." So, I

52:54

started asking like, "Okay, I got the

52:55

results." I was like, "Great."

52:58

Um, and they said, "Where do you want me

52:59

to put them?" And uh, I was like, "Well,

53:01

where can you put them?" And it said,

53:02

"Well, I can put it in a Google sheet. I

53:04

can put it in notion table. I can put it

53:05

here. I can give you a PDF. I can give

53:06

you a CSV file. Or I could write you a

53:09

CRM." And I was like, "Yeah, sure. Make

53:12

me a CRM system." And it may a CRM

53:15

system.

53:16

>> And I think that becomes, and I think

53:18

maybe one out of a thousand people

53:20

working with AI have had that

53:21

experience. Maybe it's one in 10,000.

53:24

Where your agent says, I'll make you

53:26

bespoke software.

53:27

>> Yeah.

53:28

>> Have you had that yet? And and do you

53:30

see that as a part of computer that when

53:33

a person needs a spreadsheet, you don't

53:35

launch Excel or Google Sheets, you just

53:38

pop up a spreadsheet?

53:40

>> Yeah. Well, we have a board meeting

53:42

tomorrow.

53:42

>> Okay, I'll come.

53:43

>> And and so

53:44

>> I'll pitch it to the board.

53:45

>> Sure.

53:46

>> Uh our computer computer made the memo.

53:50

>> Oh, wow.

53:50

>> Yeah. And um we had a partner meeting to

53:53

pitch a partnership idea and uh earlier

53:56

we would have a design team do the whole

53:58

deck.

53:58

>> Yeah.

53:59

>> Computer just oneshotted it. Uh I had a

54:02

press briefing with a bunch of

54:03

journalists. My comm's person would

54:05

>> Sorry about that. Brutal.

54:07

>> And then my comms person would usually u

54:09

give me a memo what to say.

54:12

>> Computer one-shoted him.

54:13

>> So

54:14

>> it's crazy. And it's the context is so

54:17

good because the memor is getting

54:18

better. Yeah. Yeah.

54:19

>> So it's like I know that journalist from

54:22

the last time.

54:24

>> I know the board meeting. I have all the

54:26

previous decks.

54:28

>> When did that happen?

54:30

>> I think it it happened with Opus 45

54:34

>> Opus 45. That was a inflection point

54:36

when models were started being amazingly

54:40

good at orchestration and reasoning and

54:42

tool calls and cloud code brought in

54:45

this new idea in AI that everything can

54:47

happen inside a sandbox, a console, a

54:51

terminal with access to tools where

54:54

tools are just command line tools.

54:56

>> Yeah,

54:56

>> they don't even need to have graphical

54:58

user interface. So when you did that and

55:00

when you organize around files and sub

55:03

aents and skills and CLIs, the model

55:06

started be becoming very good at

55:09

handling the context. So the context

55:11

window no longer became a problem. It

55:14

just put whatever necessary into the

55:15

context whenever it wanted to and dumped

55:17

dumped them away when it wanted to.

55:19

>> Yeah.

55:19

>> And that made it like suddenly so good

55:21

at doing very long orchestration tasks.

55:26

>> Yeah. It's it's pretty crazy. I have

55:28

every episode of this week in startups,

55:30

all the transcripts and then all of all

55:32

in

55:32

>> that was one of the tasks I did by the

55:33

way I can send it to you. I asked it I

55:36

want you to download every all-in

55:37

podcast.

55:38

>> Yeah.

55:38

>> U since the beginning and I want you to

55:42

take a mention of all the public

55:44

companies they mentioned during the

55:46

episode.

55:46

>> Yes.

55:46

>> I want you to have a histogram of the

55:48

counts and I also want you to chart it

55:51

across time and then I want you to

55:53

analyze the impact on the stock price

55:55

>> and the sentiment of what we said.

55:57

Exactly. And it did like it clearly

55:59

said,

55:59

>> "Are we moving stocks

56:00

>> around Google's stock going up?"

56:02

>> Yes.

56:03

>> Prior to that, you guys were talking a

56:04

lot about Google.

56:05

>> Yes.

56:06

>> And it clearly

56:06

>> And I said I made a bet publicly on the

56:08

thing. I said, "I am buying a bunch of

56:10

Google because I believe even though

56:12

they're behind,

56:13

>> it's because they're too precious." You

56:15

were kind of mentioning a company that

56:16

might be too precious at times and

56:18

doesn't release.

56:18

>> I was like, "That's that company. They

56:20

need to release more." Yeah.

56:21

>> And uh I told Sergey, I was like,

56:23

>> like

56:25

>> give us the good stuff. started giving

56:27

us the good stuff.

56:28

>> It literally gives you the timestamps of

56:29

every single and then I can go click on

56:31

it and actually hear

56:33

>> exactly

56:34

>> that moment.

56:34

>> Yeah.

56:35

>> Sweet.

56:35

>> Yeah. So that's when that's when I was

56:37

like damn like

56:38

>> this I would have had somebody do this

56:40

as a weekl long project.

56:42

>> It would have been 10 hours a week of

56:44

researcher. I I'm experiencing the same

56:46

thing when I do research notes. I've

56:49

created my own uh like mega prompt.

56:54

>> Yeah. and it will go and like tell me

56:57

where you worked before and who's in

56:59

your circle, who your competitors are,

57:01

who your friends are, blah blah blah,

57:02

and then go find I try to find old

57:05

podcast is one of my secrets. If you're

57:06

an interviewer watching, I try to find

57:09

what was the person talking about 5

57:10

years ago, 10 years ago, and then over

57:13

10 years ago. And I've gone into

57:14

interviews now with Michael Dell and

57:16

talked about things he was talking about

57:18

in the '9s. Yeah.

57:19

>> And it finds me some ancient stuff. Like

57:21

you would pay a research or a producer,

57:24

>> you know, $70,000 a year, $80,000 a year

57:27

to do this and they would have done a

57:29

third of the job in 10 times longer.

57:33

>> It's really gotten weird just in the

57:35

last 6 months. What do you think the

57:36

next 6 months looks like?

57:38

>> I think the the dream that what we are

57:40

going to try to do is help businesses

57:42

run as autonomously as possible. You

57:45

know, everybody talks about this AI is

57:47

going to create this one person $1

57:48

billion company. Some people say it's

57:51

already happened because people pay

57:53

researchers like 1 billion, but it's not

57:55

truly moving the GDP by 1 billion. It's

57:58

not truly creating new value. So the

58:01

best way to do that is to actually help

58:02

a small business people who would

58:04

otherwise drive Ubers for like yes

58:06

>> extra passive income to like buy like a

58:10

Mac mini set up perplexity personal

58:12

computer and run their business on that

58:14

or like run it on the server it doesn't

58:16

matter uh and actually make real money.

58:19

>> Yeah.

58:19

>> Hundreds of thousands or even millions a

58:20

year

58:21

>> and uh grow it.

58:24

>> Have computer go and run your ad

58:25

campaigns on Instagram or Google. I mean

58:28

>> integrate with SEM and SEO tools, find

58:30

new users and uh integrate with Stripe,

58:34

charge them, ship new features, have

58:36

your own like intercom integration for

58:39

customer support and like have this all

58:41

working well. You can be sipping wine in

58:42

Napa. That's the dream that you know it

58:45

feels awesome to say. Everybody thinks

58:47

AI is already there. It's not there yet.

58:49

Someone has to do that hard work.

58:51

>> Yeah,

58:51

>> that's what we want to do. Yeah, it it's

58:53

a great vision because

58:56

when I watched startups 20 years ago,

58:58

there were so many check boxes they had

59:00

to do. I have to find an office space. I

59:01

got to put up a bunch of servers. I I

59:04

got to hire hire an HR firm. I I got to

59:07

hire a PR person. All this stuff. And

59:09

now I talk to young founders. They got a

59:11

three-person team. They've come out of

59:13

A16Z, my program, Launch Accelerator,

59:16

whatever it is, Y Combinator. And I'm

59:18

like, "Okay, you raised a half million,

59:19

you raised a million. Who are you

59:21

hiring?" And they're like, "Um, I don't

59:23

know if we need to hire anybody." I'm

59:24

like, "If you could hire somebody, would

59:26

you hire?" They're like, "Well, I do my

59:28

own HR. I have this partner." And

59:30

they're I'm like, "How are you doing

59:32

hiring anyway?" And they're like, "Well,

59:34

I put out an ad and then uh it sorts and

59:36

ranks the candidates and then it emails

59:40

the top 10, asks them a bunch of

59:41

questions, and then I meet with the last

59:42

two." And I'm like, "That's what a

59:44

recruiter did."

59:45

>> Like, the entire recruiting job has been

59:47

abstracted. And like a a tool like

59:49

computer is going to make that even

59:51

faster.

59:52

>> Much work to do. Uh lot of connectors, a

59:55

lot of specific workflows. People don't

59:58

want to like learn how to write like,

60:00

you know, essay long prompts. You know,

60:01

it needs to be so quick and fast and

60:03

autonomous. You just set it up and done.

60:06

>> And you have an idea, you can turn it

60:07

into a business and start making money.

60:09

>> Yeah. It's it's an incredible future. Uh

60:12

and it feels like it's right here. Do

60:13

you how do you think about job

60:15

displacement? is you're actually making

60:17

the tool that enables people

60:20

>> to be a solo entrepreneur and get to a

60:22

million in revenue, but it's also the

60:23

same tool that doesn't require them to

60:25

hire. And we've had this debate a

60:26

million times on the podcast.

60:28

>> Do you

60:30

I'm wondering if like me, you have

60:32

moments where you're like, "Oh my god,

60:33

this is really terrifying." Yeah.

60:35

>> A lot of people are going to lose their

60:36

jobs really fast.

60:37

>> Yeah.

60:38

>> And then, oh my god, you can learn any

60:40

skill you want and all the things that

60:42

were hard are now easy.

60:44

>> Yeah. I I go back and forth. I'm 70 80%

60:48

super positive about this, but I do

60:49

worry about like 20% of the time I'm a

60:51

little worried. Yeah. Where do you sit?

60:53

>> I mean, America has always been about

60:54

like entrepreneur entrepreneurship,

60:57

right? Like we we've been about like

60:58

trying to build new things, discover new

61:00

things, go explore.

61:02

>> Uh I think this whole like Henry Ford

61:04

came and built factories and brought in

61:06

jobs and things like that and like put

61:09

people into a box. But u I think the

61:12

reality is people most people don't

61:15

enjoy their jobs. They're doing it for

61:16

they hate them.

61:17

>> Exactly.

61:18

>> So there is suddenly a new possibility a

61:20

new opportunity to go use these tools,

61:22

learn them and start your own mini

61:24

business. And if it pays for your needs

61:27

for year or multiple years and lets you

61:30

have a high quality life and good work

61:32

life balance and true feeling of agency

61:34

and ownership and passion to like get

61:36

your ideas out there. I think that is

61:39

even if there is temporary job

61:41

displacement to deal with that sort of

61:43

glorious future is what we should look

61:45

forward to.

61:45

>> I I I think you're exactly right. If

61:48

there will be some displacement, but

61:49

then there's also going to be so many

61:51

opportunities open up and it requires

61:53

the individual to not be passive.

61:55

>> Exactly.

61:55

>> They have to be rugged individualists.

61:58

They have to be resilient. Yeah.

61:59

>> And they have to be resourceful. And I

62:01

think once you start playing with these

62:02

tools, that's what happens.

62:04

>> Exactly. you you all of a sudden feel

62:06

like

62:06

>> it brings out the best in you if you

62:07

truly are in a good space.

62:09

>> Yeah.

62:09

>> Yeah.

62:10

>> I today uh Comet for iOS is out.

62:14

>> Yeah.

62:14

>> I'm a Comet super fan. I required

62:17

everybody. You were nice enough when I I

62:18

emailed you. I was like, "Can you send

62:19

me some licenses?" You sent You don't

62:21

may not remember. You sent me a bunch of

62:22

licenses. I said, "Everybody put this on

62:24

because it was $300 a month when you

62:26

first came out with the common browser.

62:28

Now it's free, I think, for all users.

62:30

>> Highly recommend it. Highly recommend

62:32

getting a pro account. It's only 20

62:34

bucks a month to get into perplexity,

62:35

which is a joke. So, you can get on

62:37

board for nothing, less than a dollar a

62:39

day.

62:40

>> But what does iOS allow me to do? And

62:43

and how does it connect to computer?

62:45

Because that's another thing I'm having.

62:47

>> Yeah.

62:48

>> Cloud code. Uh computer, there's not a

62:51

good enough integration with this mobile

62:53

device yet.

62:54

>> Yeah. So, computer is already on the

62:56

perplexity app. So, you can just toggle

62:58

the computer and start using it.

63:00

uh comet's uniqueness and perplexity for

63:03

the company uh and and and the strategy

63:05

is the fact that you can control the

63:08

browser. So the browser also becomes a

63:10

tool for computer

63:12

>> just like your Google workspace and all

63:14

these other things. uh until the whole

63:17

world is organized around CLI and tools.

63:20

>> Yeah,

63:20

>> there's still a lot of tasks we have to

63:22

do manually on the web on the browser.

63:25

Open tabs, fill up forms, click on

63:26

things, upload stuff, all that stuff. If

63:29

you want to automate, you need a

63:31

browser. You need an AI that can

63:33

natively control the browser. So that is

63:35

comet. And that's why no matter how many

63:38

other tools in the market exist like

63:40

open claw or like claw co-work

63:43

>> executing tasks on a browser on the

63:45

server side along with all the other

63:47

things is something uniquely perplexity

63:49

can do.

63:50

>> Yeah. My dream is that you'll create an

63:53

Android app that roots my Android phone.

63:56

>> Yeah.

63:57

>> And that you just take over and see

63:58

everything because one of the blockers I

64:00

have now is some of the websites have

64:03

gotten a little pnicity.

64:05

>> Yeah. I don't want to mention too many,

64:07

but Reddit, LinkedIn.

64:09

>> Yeah.

64:10

>> And like they're just I I am a great

64:13

Reddit user. I'm a great LinkedIn

64:15

supporter, but sometimes like I need to

64:17

get my inmail.

64:18

>> Yeah.

64:19

>> From my LinkedIn and I just need to, you

64:22

know, find seven people at company. I is

64:24

there going to be a solution

64:27

>> between the LinkedIn and Reddits of the

64:29

world and the claws and perplexities? Is

64:32

how is that

64:33

>> I mean

64:33

>> negotiation going? You don't have to

64:35

speak about any specific ones unless you

64:36

want to,

64:37

>> but it feels like there's got to be a

64:39

solution

64:40

>> and I'm willing to pay for it as a user.

64:42

I'm willing to play Reddit to allow my

64:44

bot to show up and behave properly.

64:46

>> Well, I I I cannot speak about any

64:48

particular company, but we are happy to

64:51

work with anyone, right? So, um I think

64:54

with with Comet, our idea is to give

64:56

people the flexibility to set things up

64:58

on their own.

64:58

>> Yeah. and uh any um official APIs that

65:02

anyone's willing to offer, we're always

65:04

happy to put that as part of computer.

65:07

Here's what I think should happen. Let

65:08

me see if you agree. Um and this is for

65:11

Steve Huffman at Reddit.

65:14

I go on Reddit. I do a pro account for

65:18

20 bucks a month. And when I do that, I

65:20

can authenticate whatever tool I want um

65:24

to do a series of well- behaved things a

65:28

certain number of times a day.

65:30

>> Yeah.

65:30

>> So, it's not unlimited. I'm not going to

65:31

scrape the whole site, but I would like

65:33

it to just let Perplexi or computer go

65:37

and just tell me, hey,

65:39

>> what are people saying on the this

65:40

weekend startups and all-in subreddits?

65:42

Summarize it for me so I get the

65:44

customer feedback. And I would literally

65:46

name my

65:48

uh agent and I would say I it won't post

65:51

on my behalf. It won't vote on my

65:53

behalf. Just needed to do a couple of

65:54

little readonly things. This would be an

65:56

easy solution. Or LinkedIn I would like

65:58

if you I have I already pay LinkedIn

66:00

like 50 bucks a month. Like they should

66:01

just let the $50 a month one work with

66:04

computer.

66:04

>> Yeah, absolutely. I mean,

66:06

>> okay, this is for Satia Nadella. Let

66:09

LinkedIn work with Perplexity and the

66:11

other players and we'll pay you extra.

66:14

>> Perfect. It's a revenue stream. Don't

66:15

you think API access for our customers

66:17

is a revenue stream?

66:18

>> I think so. I think so. I think I think

66:20

fundamentally giving users a choice

66:23

>> and setting it up as a win-win for both

66:24

the business and the user

66:26

>> Yeah.

66:26

>> is where the world should head to.

66:28

>> And and I I would say the same thing

66:30

applies to any any website in the world.

66:32

Like if if you want an AI to use it on

66:34

your behalf, it should be okay for cuz

66:36

that's what the user wants.

66:37

>> I mean, I have a paid New York Times

66:40

subscription. like let me go in there

66:42

and do you know whatever 100 searches a

66:45

day, a week, a month, whatever they

66:47

choose, but that would make the

66:49

subscription that much more sticky.

66:51

>> Exactly.

66:51

>> Uh all right, Arvin, love the product.

66:55

Anybody at home,

66:56

>> it's just tremendous. Go learn computer

66:59

and get the Comet browser. It has

67:01

changed my business for the last two

67:03

years. Love the product and we'll have

67:06

you back soon when you launch your

67:07

operating system and come up with your

67:09

own server and desktop server but

67:11

business is the focus. Yeah.

67:13

>> Yes.

67:13

>> All right. Great seeing you.

67:15

>> We have an amazing guest Arthur

67:16

Manchester here the CEO of Mistral AI.

67:19

How are you doing sir?

67:21

>> Great. Thank you for having me.

67:22

>> And so you're here at Nvidia's big

67:26

conference,

67:28

big announcement. You're going to be

67:30

working with Nvidia to build models.

67:33

uh to open source them. What is the uh

67:36

big announcement here?

67:37

>> Well, we're announcing that we are going

67:39

to be training the next generation of

67:41

frontier models with uh with Nvidia. Um

67:44

it's something that we've been doing

67:45

before with Nvidia with MLMO, something

67:47

we did like 18 months ago. And the point

67:49

for us is really to be able to produce

67:51

the best open source models out there so

67:53

that we can actually use those assets to

67:55

specialize them through products that we

67:57

do for our customers like Forge that

68:00

helps us customize the models for the

68:02

enterprise we work with in engineering

68:04

in physics in science uh in making them

68:06

better at certain languages when we work

68:08

with governments etc.

68:10

>> And and Michel obviously based in uh

68:12

France you're the leading AI company

68:15

there. What's it like running the

68:17

company and building a large language

68:19

model in Europe? Obviously, there's

68:20

regulations and all kinds of

68:23

considerations. Privacy, the French are

68:24

known for protecting privacy. In the

68:26

United States, we're known for taking it

68:28

away. How is the landscape there and

68:31

what do you have to deal with there that

68:32

maybe you wouldn't have to deal with in

68:34

America? And what's the pros and the

68:35

cons? I'd say first, we have 25% of our

68:38

business in the US. Uh, and 25% of our

68:40

researchers are actually here. So I

68:42

actually spend a lot of time here as

68:44

well as in France as well as in the UK

68:46

in Singapore where we are. So of course

68:48

it's it's different markets. Uh it's

68:51

markets where you have language which is

68:52

a topic uh where there's much more

68:55

manufact manufacturing is a bigger piece

68:57

of the cake than it is here. uh and I'd

69:00

say the our strength has been to also

69:02

work with European companies that are a

69:04

bit lagging behind uh and that wants to

69:06

adopt the technology to to leap forward

69:09

and we've been able to do that through a

69:10

forward deployment engineering

69:11

engagement through our forge product for

69:13

our studio product that allows to deploy

69:15

agents that do end to end automation but

69:18

on top of that the thing that we have

69:19

announced today like forge is something

69:21

that is actually being used today uh

69:23

with customers in the US because they

69:25

come to us with uh needs for post

69:28

training for making mod specifically

69:29

good at financial services and what's

69:32

happening is that we have this product

69:33

and we can bring the models to

69:34

specialize them as well.

69:36

>> And so your belief is specialized

69:39

verticalized models healthcare finance

69:42

engineering different verticals will win

69:44

the day or a a global model will win the

69:47

day that does everything.

69:49

>> Well you need general purpose models to

69:51

do the orchestration parts etc. But at

69:53

some point you enterprises sits on a lot

69:56

of intellectual property on a lot of

69:58

signals coming from physical systems

70:00

from factories from tools and the it's

70:03

actually not trivial to connect those

70:05

systems to connect those data to models

70:07

that are closed source. If you have open

70:09

models you can actually add uh new

70:10

parameters you can make a lot of deeper

70:13

things that you cannot do with closed

70:14

models. You can also and that's what

70:16

something that we do. We don't we not

70:18

only do we work at the model side but

70:19

also at the orchestration side. We see

70:21

it with subject matter experts to

70:23

understand their needs and we build

70:24

business applications that are fully

70:26

bespoke to their needs by modifying the

70:28

models but also modifying the harness on

70:30

top etc. So we believe that eventually

70:33

building on open source technology is a

70:35

way to save cost is a way to have better

70:36

control because you can sit the thing on

70:38

every cloud that you want on your

70:40

hardware if you want you can deploy it

70:41

on the edge if you want and eventually

70:43

uh from a from a customization

70:45

perspective and from leveraging your

70:47

decades of IP that you've been acrewing

70:49

in financial services in heavy

70:51

manufacturing like companies like SML

70:53

for instance they do benefit from

70:54

working with us because we take their

70:56

data and we build models that are

70:57

specifically good for their

70:59

>> um just training data using experts to

71:03

come in and refine a model. Most people

71:05

don't know this business that well, but

71:07

this has become a very large part of the

71:10

industry. Obviously, scale AI was doing

71:12

it. They went to Facebook, lost a lot of

71:14

the customer base who didn't want to uh

71:17

send their data, I guess, over to Meta.

71:19

Uh we're investors in a company called

71:20

Micro One that's doing pretty well in

71:22

this space. There's other folks doing

71:23

it. explain to the audience what you're

71:27

doing specifically for companies and how

71:29

this training works in a verticalized

71:31

way and then how you silo that data

71:33

because if you're working with one

71:34

customer in aerospace or fintech they

71:37

might have a need set but they may not

71:39

want that training to go to a

71:42

competitor. I can use a few examples. I

71:45

think overall the data segregation is

71:47

super important and the way we have

71:48

solved that is through a portable

71:50

platform. So our technology is a set of

71:52

services, a set of training tools, a set

71:55

of data processing tools that I can take

71:57

and that I can put on the infrastructure

71:59

of my customers. So suddenly from an IT

72:01

perspective and when we talk to the

72:02

CIOS, they realize that from security

72:04

perspective, the flow of data doesn't go

72:07

there's no data flow coming back to

72:08

Mistral because everything stays there.

72:10

Now uh the way we we then use that

72:13

technology that has been deployed is

72:15

that we're going to be working with uh

72:17

the team that is doing uh image scanning

72:20

and default detection with ISML for

72:22

instance and we're going to be sending

72:24

forward deployment engineers scientists

72:26

they're all PhDs they know how to train

72:27

models and they spend some time with the

72:29

subject matter experts that can explain

72:31

how an image is being detected what how

72:33

do you def detect defaults etc and based

72:36

on that we're going to work out what

72:37

kind of data needs to be used to train

72:40

the models that it's going to solve the

72:41

task in itself. And so the we we send

72:44

the technology typically we send a

72:46

little bit of scientists because uh you

72:49

do need that expertise transfer and that

72:51

knowledge transfer in between our teams

72:53

and the vertical experts and then we

72:55

make sure that eventually our team no

72:57

longer needs to be there to retrain the

72:58

models to get more data access etc. So

73:00

that combination of data segregation,

73:03

expertise transfer, knowledge transfer

73:05

is the one thing that makes us quite

73:06

unique and allows us to serve the most

73:09

critical use cases, the most critical

73:10

processes in industries that actually

73:12

need to take their data and put it into

73:14

models for it to work. Yeah, this seems

73:17

to be once the entire open web, what was

73:21

available legally, gray market, etc. I

73:25

wouldn't have you comment on that

73:26

controversy. Uh but we we kind of

73:29

exhausted what's in the open crawl.

73:31

Yeah,

73:31

>> we have.

73:31

>> And and it's time to actually

73:34

either make synthetic data or actually

73:37

use experts. Do you believe in synthetic

73:39

data and where does that work and where

73:41

does it fail? We use synthetic data as a

73:44

way to warm up the models. It's a way to

73:46

actually be quite efficient at the

73:48

beginning. If you have a large model and

73:49

you want to train a small model, you

73:50

would you will use your large model to

73:52

pro to process and to produce a lot of

73:55

synthetic data at the beginning. uh and

73:57

then but eventually you do need to have

73:58

human signal. Uh so the human signal is

74:01

something that is always a bit costly to

74:02

acquire because you need to talk to the

74:04

experts they need to give feedback to

74:06

the machines and so at the beginning

74:08

synthetic data allows you to do the

74:10

compression to to further compress the

74:11

models. At the end you do need to go and

74:13

get data that is uh produced by humans.

74:16

So yeah, it's a it's a way to have uh

74:18

it's it's mostly an efficient way of

74:20

training models to have big bigger

74:22

models that are used as as teachers for

74:25

smaller models, but it's not enough. And

74:26

so you also need human signal. Arthur,

74:28

we've seen um an incredible explosion.

74:31

We're sitting here on AO52

74:34

after OpenClaw, the year of our Lord, 52

74:37

days.

74:39

when you first saw Open Claw and saw the

74:41

reaction of hackers,

74:45

founders, startups, CEOs, just the

74:47

amount of energy and it racing to the

74:49

top of GitHub with the most number of

74:51

stars and likes and and all these

74:54

contributors. What did that say to you

74:56

as an executive in the space who's been

74:59

grinding on this for many years? What

75:01

what does that openclaw moment mean?

75:04

Well, it resonated a lot with what we

75:05

were doing with our customers uh because

75:08

pretty quickly uh enterprises realized

75:10

that if they wanted to make some gains

75:12

with artificial intelligence geni, they

75:14

would need to automate full processes.

75:16

And to automate a full process as an

75:18

enterprise, well, you can use open

75:19

cloud, but it's going to be uh it's

75:21

actually not really enough because you

75:22

you have data problems, you have

75:24

governance problems, you can't observe

75:26

uh the process that is running and you

75:28

can't can't control it in um in many

75:30

cases when you run a KYC process. So if

75:32

you're HSBC for instance, one of our

75:34

customers, uh you will want to have

75:36

deterministic gates that are going to

75:38

always do the same thing in a way that

75:40

is observable and that you can guarantee

75:43

the CIO that it's always going to go

75:44

through these gates and that's not

75:46

something that Openflow is providing

75:48

because it doesn't have this the kind of

75:50

primitives that you need to work on

75:52

collective productivity, observable

75:54

productivity and to work on mission

75:55

critical systems. On the other hand, uh

75:58

the autonomy it gives and the autonomy

76:00

it brings to to people that are just

76:02

individuals that are hacking together

76:04

things is a way to also show to

76:05

enterprises that if you set up the right

76:07

control plane, if you set up the right

76:09

sandboxes, if you connect to the right

76:11

data sources, if you make sure that you

76:13

your access controls are well respected,

76:15

then you can actually unleash the power

76:17

of agents doing things for your

76:19

employees and that's going to work. Work

76:21

on the platform cuz otherwise you will

76:23

not be at ease when you're sleeping. It

76:24

is um definitely something you have to

76:26

be thoughtful about. When I installed

76:28

it, I gave it just for my agent root

76:31

access to my Google Docs and my G Suite,

76:35

my notion, my Zoom and uh my notion and

76:40

uh GCAL, everything. And then I

76:42

realized, wow, I can with my enterprise

76:45

edition of Gmail essentially, I can just

76:48

summarize for my entire 21 person

76:50

investment company every conversation

76:52

going on in Gmail and then correlate it

76:55

with every conversation in Slack. And

76:57

then I realized, oh my gosh, there's

76:59

compensation discussions going on.

77:01

There's a person on a PIP who we put

77:03

them on a perform performance

77:05

improvement plan perhaps or something

77:06

like that. Oh, I have to make sure

77:09

nobody else can access this because the

77:12

power comes from giving it access to

77:15

data. But with great power comes great

77:17

responsibility and I think people are

77:19

learning that in real time. Yeah, it's a

77:21

big problem because the enterprise data

77:23

is not a single thing that you want to

77:24

put into a single system that is going

77:26

to be accessible by by everyone and so

77:28

you need to have this layer that

77:30

actually understands what is the is what

77:32

is in the data. you need to have a

77:34

semantic of what can actually be

77:35

proposed to uh HR or what can be

77:39

proposed to uh engineering and typically

77:42

compensation is one of these things you

77:43

want to make sure that the compensation

77:45

data does not flow back to all of the

77:47

all of the enterprise because you're

77:48

going to have a lot of problems uh if if

77:50

that's the case and so what you actually

77:52

need and which is hard to do is what we

77:54

call context engine so a mapping of

77:56

where the data sits that comes with a

77:58

certain number of metadata that is

78:00

telling you that this data is actually

78:01

not accessible to part of the company

78:04

and if you actually have someone in

78:06

engineering that is asking for something

78:08

related to comp the thing is actually

78:10

going to tell you look you actually

78:11

can't access that data so so that's uh

78:14

that's hard it's actually hard you need

78:16

to rethink entirely the way your IT

78:17

systems are being connected and uh at

78:20

some point you also need to think about

78:21

your management because your influ your

78:23

information flow is completely different

78:25

today uh if you're connecting agents

78:28

together with your data sources than it

78:29

used to be and suddenly maybe you don't

78:31

need that manager whose only purpose was

78:33

to take information from the bottom and

78:35

put the information on top etc. So

78:37

there's some IT problems to solve and

78:39

you need the right primitives, you need

78:41

sandboxes, you need airback based access

78:44

control and these kind of things and uh

78:46

you have change to do. You you need to

78:48

rethink your entire customer service uh

78:50

department cuz suddenly you actually

78:52

don't need that much transfer of

78:53

information operated by humans.

78:55

>> All right. Uh you have to go. You got a

78:57

flight to catch. It is so great to see

78:58

you Arthur. Continued success with

79:00

Mishril.

79:01

>> Thank you very much. Cheers. I'm really

79:02

lucky to have Daniel Roberts here. He's

79:04

the co-CEO and co-founder along with his

79:07

brother of Iron. They are a publicly

79:10

traded company. They started in BTC.

79:12

Welcome to the All-In Interview program.

79:14

>> Thanks, Jason. Pleasure to be here.

79:16

>> Yeah. And so you started in Sydney. You

79:19

and your brother um was seven, eight

79:22

years ago. And you got in early on

79:25

Bitcoin and all these Bitcoin monitor uh

79:28

miners wanted to have data centers. Huh.

79:30

>> Yeah, that that's directionally right.

79:32

So the thesis we saw was this explosion

79:35

of the digital world, the growth in the

79:36

online and at some point the real world

79:39

was going to struggle. So we set about

79:41

to build out largecale data centers.

79:43

Yes, the first use case was Bitcoin

79:45

mining. But as we said to our seed

79:47

investors, use that to bootstrap the

79:49

platform, generate cash flow, layer in

79:52

higher and better use cases over time as

79:53

they emerge. Here we are today with AI,

79:56

we are swapping out all the Bitcoin for

79:57

AI chips. When did you first start

79:59

seeing the demand in the company shift

80:02

from hey Bitcoin miners we need some

80:06

H100s whatever it is uh to hey we're

80:10

this nonprofit open AAI hey we're this

80:13

research lab we need some AI compute

80:16

when did that start hitting

80:17

>> look we had a bit of a false dawn I

80:19

would say back in 2020 we signed anou

80:21

with Dell to start bringing out

80:22

customers and compute but in hindsight

80:25

it was too early so we went back to

80:26

Bitcoin kept bootstrapping in the

80:28

platform. Look, I would say about 2

80:30

years ago and month by month, the demand

80:33

just continues to escalate.

80:34

>> And you were in so early that when you

80:37

were looking at data center space in the

80:39

United States,

80:42

you were one of one looking at the

80:43

space, one of two or three people

80:45

looking at the space, they they were

80:47

trying to sell you on space. Yeah.

80:49

>> Yeah. So, we actually develop the data

80:51

centers ourselves. So, we go and find

80:53

the land, we go and get the permits, we

80:54

go and apply for grid connections. And

80:57

we were doing it at a scale that just

80:59

amazed people at the time. Like 750

81:01

megawatts is our flagship Texas site

81:04

four years ago was unheard of. In the

81:06

middle of the desert, we're building

81:07

these big data centers. The traditional

81:08

data center industry going what are you

81:10

guys doing? We're saying we believe in

81:12

the future digitization, high

81:14

performance computing and obviously now

81:16

today it's paying dividends.

81:17

>> Yeah. I don't think anybody could have

81:19

predicted when chat GPT came out, Open

81:22

Claw recently as a turning point. Um,

81:25

and then you know, Microsoft, Google,

81:28

and everybody embracing this. Uh, and

81:30

that's your big partner, Microsoft.

81:33

>> Yes, Microsoft's one of our early

81:34

partners. We signed a $9.7 billion

81:37

contract with them late last year, but

81:39

as I was explaining to you before the

81:41

show, that's 5% of our capacity. So,

81:44

things are busy at the moment.

81:45

>> Yeah.

81:46

>> And when you do these buildouts,

81:49

the big conversation today is not is no

81:52

longer the number of GPUs putting in.

81:54

It's just power. Power is the uh

81:58

constraint today. Yeah.

81:59

>> Look, for many of the industry it is,

82:01

but for us, because we started 8 years

82:03

ago tying up all this land and power,

82:05

it's not. So, we've got 4 1/2 gawatt.

82:08

For context, that's almost as much power

82:11

annually as the Bay Area uses in its

82:13

entirety each. Wow. It's huge. So, for

82:16

us, the hurdle or the constraint is

82:18

really time to compute. And that's

82:21

emerging across the industry as well.

82:23

And time to compute means trades people

82:27

coming to West Texas living in a a

82:30

trailer that you set up to then break

82:34

ground on a data center, build

82:36

foundations, build water cooling

82:38

systems. Like this is hard manual labor

82:41

going on. Yeah,

82:42

>> exactly. And this is the whole real

82:44

world challenge to respond to these

82:46

digital exponential demand curves that

82:48

are unconstrained by the real world in

82:50

terms of their appetite. And it just

82:52

compounds. You need thousands of people

82:54

out in these locations that haven't

82:56

supported it. You put stress on supply

82:58

chains. We're seeing what's happening

82:59

with the memory, every aspect of it. So,

83:02

it's just permanent whack-a-ole,

83:03

permanent solving fires to try and be

83:05

bring online this compute.

83:07

>> And you get to spend time there.

83:10

>> What's it like when you set up a town or

83:13

you bring a thousand people or 2,000

83:16

people to what's a pretty much remote

83:19

small town? you I'm assuming that like

83:21

when you bring a thousand there might

83:22

only be 500 living there right now. So

83:26

what are those towns like? I'm it sounds

83:28

to me like something out of like the

83:30

gold mining era when people first you

83:32

know uh went and and were prospectors

83:35

prospecting town

83:37

>> pretty pretty much. I mean the

83:38

barbecue's great that was a draw card

83:40

but apart from that uh look we've always

83:42

had a policy of hiring local supporting

83:44

the local community. Uh this year we're

83:46

hitting a million dollars in community

83:48

grants cumulatively. That's things like

83:50

local playgrounds, supporting the fire

83:52

departments, but we will hire locally.

83:54

Once we can't find that trade locally,

83:56

we will expand the radius by 20 mi and

83:58

hire out of that and so on and so on for

84:00

us.

84:00

>> That's very thoughtful. Yeah. And and

84:02

these folks are coming say an

84:04

electrician or a construction worker.

84:07

They're coming having built houses or

84:10

you know uh maybe building um corporate

84:15

offices and now they come for a tour of

84:18

duty here and the salaries go up

84:20

massively but they got to leave their

84:22

family for a 3-month tour or something.

84:24

>> Yeah. Yes and no. Because typically

84:26

where we locate is where there's heavy

84:28

electrical infrastructure. Where there's

84:31

heavy electrical infrastructure is

84:32

typically where old manufacturing and

84:35

industry has closed down. Ah,

84:37

>> so we go in, leverage that sunk capex,

84:40

rehire, retrain local workforces and

84:43

bring a new industry to town in these

84:44

data centers. H

84:45

>> has has that workforce now been

84:48

completely depleted and we need to train

84:50

another generation, a younger generation

84:53

to be generation tool belt and really

84:55

embrace the trades

84:57

>> 100%. We're partnering with

84:58

universities, trade colleges.

85:00

Absolutely. And you go to a trade

85:02

school, you got you go to a college,

85:05

people are getting degrees in philosophy

85:07

and English literature, they're going

85:10

50k a year in debt, 200k a year in debt.

85:13

What's the starting salary for a trades

85:15

person working on a data center doing

85:18

electrical or construction or HVAC?

85:20

What's the ballpark range?

85:22

>> Uh, look, I won't talk specifics, but

85:24

they they are going up. The price is

85:26

going up. Depends on the level, but yes,

85:28

there is a rush for good. hearing 150 to

85:31

like 300K. Am I in the ballpark?

85:34

>> The lower end directionally, you're

85:35

right. Yeah.

85:35

>> Yeah. I mean, it's incredible when you

85:37

think about it. There's concern about,

85:40

hey, AI taking jobs and then on this

85:42

other side of the ledger can't find

85:45

enough talent to to to service it. Talk

85:48

to me about energy sources and how you

85:51

think about that. Uh, President Trump,

85:53

Chris Wright, the administration that

85:56

kind of started with, hey, clean,

85:57

beautiful coal. Year two, they're like,

86:00

"All sources matter." Nuclear,

86:02

obviously, nack gas is plentiful in that

86:05

area. We obviously got a lot of oil.

86:06

People don't know this about Texas in

86:08

the United States, the number one uh

86:11

source of solar installations. Yeah.

86:14

>> Yeah. Talk to us about energy.

86:15

>> So, so our our philosophy has been

86:17

sustainability from day one. We have

86:18

used 100% renewable energy since

86:21

inception.

86:22

>> What?

86:22

>> 100%.

86:23

>> Wait, how is that possible? It's

86:25

>> We use hydro in British Columbia. We use

86:27

wind and solar in West Texas. In West

86:29

Texas, where we're located, there's

86:31

around 45 to 50 GW of wind and solar.

86:34

>> Yeah.

86:34

>> The transmission line to export that

86:36

down to the load centers in Dallas and

86:38

Houston is 12 GW.

86:40

>> Oh.

86:40

>> So you go and locate to the source of

86:42

lowcost excess renewable energy,

86:44

monetize it into this digital commodity,

86:46

export it at the speed of light as

86:48

token.

86:49

>> Great arbitrage. And the wind is

86:51

producing a lot, but it it's harder to

86:53

get from those areas where people are

86:56

willing to put up. I mean, people don't

86:57

understand how big West Texas is. It is

87:01

an incredible amount of land. And you're

87:02

coming from Australia

87:04

where also on the west side, people

87:06

don't understand exactly how much just

87:09

pure nature land there is. Yeah.

87:10

Undeveloped.

87:11

>> So much land. And the issue is distance.

87:13

You've got to spend billions of dollars

87:14

on this transmission connection

87:16

infrastructure to move that power to

87:18

where people actually want it. You can

87:19

build wind farms, you can build solar

87:21

farms, but if you build it in the desert

87:23

and no one can use it, then what's the

87:25

point? So the whole opportunity for our

87:26

industry is to go to the source of that

87:28

power and monetize it.

87:30

>> So the data centers follow the wind

87:33

turbines, the solar installations.

87:36

How do you think about batteries and are

87:38

you able to put those online? Because

87:40

obviously you're going to have periods

87:41

where, hey, it's not a windy day. In

87:43

Texas, we have very few days when it's

87:45

overcast, so that problem's pretty much

87:47

solved. But you're going to have 50 days

87:48

where the sun's not beating down. So,

87:51

how do you deal with the demand and and

87:53

and softening that duck curve?

87:55

>> We don't need to.

87:56

>> The utility does that on our behalf. So,

87:58

this is why these grid connections are

88:00

so scarce, so hard to get, and so highly

88:02

valued because once you get that grid

88:04

connection, the utility underwrites all

88:06

of that variability. They guarantee you

88:09

24/7 reliable power.

88:11

>> Got it. So, on their side, they're

88:13

figuring it out. something goes down and

88:15

they could fall back even though you're

88:18

100% committed to renewables if they

88:20

needed to fall back to gas or whatever

88:22

they have that ability out there. So you

88:24

have that as a backup.

88:26

>> A lot of talk about or a debate. Are we

88:29

getting ahead of our skis? Are people

88:32

slowing down? There was some talk about

88:34

the OpenAI project maybe downscaling a

88:36

little bit. Is OpenAI a partner as well

88:38

or

88:39

>> uh can't comment.

88:40

>> Can't comment. Okay. So we'll we'll read

88:42

into that whatever we want.

88:44

But

88:46

are there pockets where people are

88:47

saying, "Hey, let's slow down." Or is it

88:49

still gang busters?

88:51

>> It's right up the end of the spectrum.

88:53

It's gang busters. We we cannot meet

88:55

demand. That's why the whole industry

88:57

now is around time to compute. There are

88:59

no idle GPUs in the world sitting in a

89:01

data center.

89:02

>> Yeah. And what's your take on when

89:06

software makes and this is a big uh

89:10

discussion from Jensen himself during

89:12

his two and a half hour keynote

89:13

yesterday. Uh we're sitting here

89:15

Wednesday. I think he did his keynote on

89:16

Tuesday. He was talking about hey

89:18

software is going to make it 50 times

89:19

more uh you know lower the cost of

89:21

tokens 50x and then you have um

89:25

transport also contributing to that.

89:28

When do you think the curve goes from

89:30

parabolic to simply growing at a

89:33

ridiculous level? Is is there a slowdown

89:36

coming or how are you planning for the

89:38

future?

89:38

>> Look, I think it's actually the

89:39

opposite. I think it feeds on itself.

89:41

So, I'll give you one example. You go

89:43

into chat GPT today and you generate an

89:45

image. You enter to the prompt. It's

89:47

like the dialup internet days.

89:48

>> It is

89:49

>> right. It takes minutes and you're like,

89:50

I better get this prompt right.

89:52

>> Yeah.

89:52

>> Finally, 2 minutes later, it comes. Now,

89:54

I'll give you an example. If we 10x the

89:56

amount of compute available, which is an

89:58

enormous task from where we are today,

90:00

and those images take 5 to 10 seconds,

90:02

are we going to generate more or less

90:04

images?

90:05

>> Oh, many more. Uh, this is Jevans

90:07

paradox. This is the theory of induced

90:09

traffic. You know, you build a couple

90:11

more lanes, people start to think, well,

90:13

maybe the uh distance from Bondai Beach

90:16

to to the central business district in

90:18

Sydney terms would be an acceptable

90:20

commute.

90:20

>> Love the analogy.

90:21

>> Yeah. Uh, so what do you think about or

90:25

or what are you seeing? I mean, we're

90:26

here at Nvidia. Obviously, they make the

90:29

leading edge chips. They just bought

90:30

Grock, so now you've got, you know, two

90:33

of the leading edge chips uh coming out

90:35

of the same company, but custom silicon

90:38

becoming a big discussion. Has that

90:40

started to land in the data centers yet?

90:42

Obviously, Google, don't know if they're

90:44

a customer, you can tell us, but they're

90:46

making custom silicon. Amazon is making

90:48

custom silicon. Meta is making custom

90:51

silicon. Talk to me about that

90:53

revolution and is it actually making it

90:55

to the data centers yet?

90:56

>> Look, it to various degrees it is.

90:58

They're promoting their products.

90:59

They're trying to tie up data center

91:01

capacity. So yes, there's multiple

91:03

silicon looking for homes. I think I

91:05

think it's fair to say Nvidia has a

91:07

massive head start. The ecosystem

91:09

they've incubated the standards that

91:11

they're setting. So I would say the

91:13

safest pathway to build out at scale

91:14

early is to follow the Nvidia road maps.

91:17

But absolutely over time we are seeing

91:19

these chips emerge.

91:20

>> A and in terms of desktop computing I

91:25

don't know if you saw the announcement

91:27

that um Dell and Nvidia are making a

91:31

really powerful desktop 750 gigs of RAM

91:34

lot of power. You're going to be able to

91:35

run some local models open source and

91:39

with openclaw and open source coming

91:42

from uh Kimmy and a bunch of the models

91:45

out in China.

91:47

has the hacker group, which I think you

91:49

started in like I did probably in

91:51

similar time periods. People are

91:53

starting to get really obsessed with

91:55

having a 10 or $20,000 desktop setup and

91:58

running this local. What do you think of

91:59

that trend? I'm curious.

92:00

>> Yeah, I mean the breakthroughs we're

92:01

seeing in software, the way it's

92:03

distributing power to every man in every

92:06

and woman in every house and their

92:07

ability to code and use products like

92:09

open core, the generation of demand and

92:12

appetite for compute at a local level

92:14

all the way through to these mega data

92:15

centers. It's absolutely real and as we

92:18

see the emergence of agents using more

92:20

and more as we see autonomous vehicles

92:22

and other automation, robotics, it's

92:24

absolutely going to compound.

92:26

>> And what about nuclear? Uh the Trump

92:29

administration

92:31

really seemed to flip the switch on a

92:35

growing

92:36

uh belief that hey wait, nuclear is

92:39

pretty great. It's clean. It's the

92:41

original renewable in a way. Uh, and

92:43

these new modular reactors have nothing

92:45

to do with Chernobyl, Fukushima, or

92:48

Three-Mile Island. They're much safer.

92:50

They're a completely different

92:51

architecture. Have the Have those

92:54

started to land yet? And are you since

92:56

you followed correctly in in the great

92:58

state of Texas where I'm from, you

93:00

followed correctly that time, are you

93:02

following nuclear?

93:03

>> I I think you have to. I think the

93:05

reality is it's going to take a decade,

93:07

a bit longer by the time big projects

93:09

can come into commissioning, but now is

93:11

the time to start that conversation. and

93:13

put in place policies, mobilize capital,

93:15

and start that ball rolling.

93:17

>> Yeah. Have you do you have a data center

93:19

going up near nuclear?

93:21

>> No, not at the moment.

93:22

>> Not at the moment. But you're actively

93:24

tracking that activity cuz

93:26

>> Yeah, this seems uh pretty inevitable.

93:30

Yeah,

93:30

>> feels like it.

93:31

>> And if that happens, what impact does it

93:33

have on your industry? If if you could

93:35

because obviously it's happening in

93:37

China and people always put the Bitcoin

93:39

miners they were like the canary in the

93:41

coal mine near the hydro dams and near

93:44

the nuclear where there was excess

93:45

capacity. What impact do you think this

93:48

has if you could actually have small

93:50

modular reactors next to data centers?

93:52

>> Well, I I think it just opens up the

93:54

market and enhances the US's competitive

93:55

advantage in this space. Like AI is

93:57

inevitable, robotics is inevitable. The

94:00

reality is the correlation between human

94:02

progress and energy consumption is

94:05

really really high over a very long time

94:07

period. So if we can find a way to

94:09

unlock new generation, clean generation

94:11

as nuclear and locate that more at the

94:13

source and enable more compute on a

94:15

distributed basis, all those use cases

94:17

we just discussed become easier, more

94:19

fluid, faster and then you get that

94:22

positive flywheel around Jebans's

94:23

paradox and demand. Talk to me about the

94:26

architecture today of

94:29

Ethernet and data moving between data

94:33

centers within data centers. That

94:35

backbone is going through a paradigm

94:38

shift as well. Yeah.

94:39

>> Yeah. Yeah, it is. And Jensen coins

94:41

coins the term the data center is the

94:43

new computer.

94:44

>> So you need to step back and you say

94:46

right this big building is essentially

94:48

the old desktop PC we had under our desk

94:51

at home. You go right how does that

94:53

work? So all the cabling, the latency,

94:56

the number of hops between each GPU, how

94:58

they talk to each other, the fabric

94:59

around Infiniband, Ethernet, it's

95:02

absolutely critical because every

95:04

millisecond matters in terms of

95:05

performance of that cluster.

95:07

>> Yeah. And where do you think uh or or

95:11

what do you think of Elon's uh vision?

95:16

It's obviously a a longer term vision of

95:18

putting data centers in space and

95:20

there's a couple other people working on

95:21

it as well. Yeah, I mean it's very hard

95:24

to argue with Elon. He's been very right

95:27

on a number of things for a very long

95:28

time. I think sitting here today, it

95:30

feels exceptionally difficult given the

95:32

cost of moving things to space, the

95:34

challenges around radiation. There's a

95:37

huge amount of engineering challenges,

95:39

but that's never scared Elon before. So,

95:41

I'm not

95:41

>> qualified and he's he he's inevitably

95:44

right, but sometimes he's late. He might

95:48

be late to the party. He might be late

95:49

to the dinner party. you might show up

95:51

at dessert, but generally uh he nails

95:54

it. How much of an issue is getting the

95:56

data out of the data center to consumers

96:00

today? Is that not something people are

96:02

worried about when you're building

96:03

something out in West Texas, all that

96:06

data, fiber, all that's been taken care

96:09

of or does that become a gating issue at

96:11

some point? So this was one of the big

96:13

myths that we had to bust when we

96:15

started this business because everyone

96:16

said data centers must be located close

96:18

to population centers, metropolitan

96:20

areas. Latency is really important and

96:22

we say yeah that's right latency is

96:23

important but the reality is in the US

96:26

Texas especially there is fiber

96:28

everywhere underneath the ground lots

96:30

and lots and lots of it. And when you

96:32

look at latency from our site in the

96:34

middle of the desert in West Texas down

96:37

to Dallas, the big carrier hotel, six

96:39

millisecond roundtrip latency. What's

96:42

six milliseconds? There's a thousand

96:44

seconds milliseconds in a second. Yeah,

96:46

>> we're talking six.

96:47

>> It's it's adjacent.

96:49

>> Yeah, it's it's not even uh Yeah, it's

96:52

definitely not material. Uh listen,

96:54

continued success uh and uh you're

96:58

hiring

96:59

>> a lot of people.

97:00

>> Yeah. Yeah, I think we got 129 job

97:03

advertisements up at the moment.

97:04

>> All right, so everybody go to the Iran

97:06

website. Uh, and listen, company's doing

97:09

fantastic. Thanks for spending some time

97:11

with us here at Allin

97:13

GTC.

97:16

>> Thanks, Jason.

97:16

>> Appreciate it.

97:33

I'm going all in.

Interactive Summary

The video features interviews with four AI CEOs at Nvidia's GTC conference. Michael Intrader from CoreWeave discusses their evolution from an algorithmic hedge fund to a major GPU infrastructure provider, highlighting their early adoption, risk management, and unique financing model. Arvin Shri Nas of Perplexity details their focus on accuracy and the progression of their products, from AI-powered search to a full AI computer, envisioning AI as a future operating system that runs locally. Arthur Manchester from Mistral AI emphasizes their commitment to open-source frontier models for enterprise specialization, focusing on data segregation and human expertise for training. Finally, Daniel Roberts of Iiron describes their journey from Bitcoin mining to building massive, renewable-energy-powered data centers for AI, addressing the relentless demand for compute and the long lifespan of GPUs, while also touching on the increasing need for skilled tradespeople.

Suggested questions

9 ready-made prompts