HomeVideos

Elon Musk – "In 36 months, the cheapest place to put AI will be space”

Now Playing

Elon Musk – "In 36 months, the cheapest place to put AI will be space”

Transcript

1631 segments

0:00

Are there really three hours of  questions? Are you fucking serious? 

0:07

You don't think there's a lot to talk about, Elon? Holy fuck man. 

0:11

It's the most interesting point. All  the storylines are converging right now. 

0:16

We'll see how much we can get through. It's almost like I planned it. 

0:19

Exactly. We'll get to that. 

0:20

But I would never do such a thing… As you know better than anybody else,  

0:25

only 10-15% of the total cost of  ownership of a data center is energy. 

0:30

That's the part you're presumably saving  by moving this into space. Most of it's  

0:33

the GPUs. If they're in space, it's harder  to service them or you can't service them. 

0:37

So the depreciation cycle goes down on them. It's just way more expensive to have  

0:40

the GPUs in space, presumably. What's the reason to put them in space? 

0:46

The availability of energy is the issue. If you look at electrical output  

0:55

outside of China, everywhere outside  of China, it's more or less flat. 

0:58

It’s maybe a slight increase,  but pretty close flat. 

1:02

China has a rapid increase in electrical output. But if you're putting data centers anywhere  

1:07

except China, where are you going to get your  electricity? Especially as you scale. The output  

1:12

of chips is growing pretty much exponentially,  but the output of electricity is flat. 

1:17

So how are you going to turn the chips on? Magical  power sources? Magical electricity fairies? 

1:25

You're famously a big fan of solar. One terawatt of solar power,  

1:29

with a 25% capacity factor, that’s  like four terawatts of solar panels. 

1:32

It's 1% of the land area of the United States. We’re in the singularity when we’ve got  

1:37

one terawatt of data centers, right? So what are you running out of exactly? 

1:42

How far into the singularity are you though? You tell me. 

1:45

Exactly. So I think we'll find we're  in the singularity and it’ll be like,  

1:48

"Okay, we’ve still got a long way to go." But is the plan to put it in space after  

1:54

we've covered Nevada in solar panels? I think it's pretty hard to cover Nevada  

1:58

in solar panels. You have to get permits. Try  getting the permits for that. See what happens. 

2:02

So space is really a regulatory play. It's harder to build on land than it is in space. 

2:08

It's harder to scale on the ground  than it is to scale in space. 

2:17

You're also going to get about five times the  effectiveness of solar panels in space versus  

2:22

the ground, and you don't need batteries. I almost wore my other shirt, which says,  

2:27

"it's always sunny in space". Which it is because you don't  

2:35

have a day-night cycle, seasonality,  clouds, or an atmosphere in space. 

2:44

The atmosphere alone results  in about a 30% loss of energy. 

2:50

So any given solar panel can do about five  times more power in space than on the ground. 

2:58

You also avoid the cost of having  batteries to carry you through the night. 

3:03

It's actually much cheaper to do in space. My prediction is that it will be by far  

3:11

the cheapest place to put AI. It will be space in 36 months or  

3:16

less. Maybe 30 months. 36 months? 

3:17

Less than 36 months. How do you service GPUs as they fail,  

3:20

which happens quite often in training? Actually, it depends on how recent the  

3:27

GPUs are that have arrived. At this point, we find our  

3:30

GPUs to be quite reliable. There's infant mortality,  

3:34

which you can obviously iron out on the ground. So you can just run them on the ground  

3:38

and confirm that you don't have  infant mortality with the GPUs. 

3:41

But once they start working and you're  past the initial debug cycle of Nvidia  

3:49

or whoever's making the chips—could be Tesla  AI6 chips or something like that, or it could  

3:56

be TPUs or Trainiums or whatever—they’re  quite reliable past a certain point. 

4:06

So I don't think the servicing thing is an issue. But you can mark my words. 

4:15

In 36 months, but probably closer to 30 months,  the most economically compelling place to put  

4:23

AI will be space. It will then get  

4:29

ridiculously better to be in space. The only place you can really scale is space. 

4:37

Once you start thinking in terms of what  percentage of the Sun's power you are harnessing,  

4:42

you realize you have to go to space. You can't scale very much on Earth. 

4:47

But by very much, to be clear,  you're talking terawatts? 

4:51

Yeah. All of the United States currently  uses only half a terawatt on average. 

4:59

So if you say a terawatt, that would be  twice as much electricity as the United  

5:03

States currently consumes. So that's quite  a lot. Can you imagine building that many  

5:08

data centers, that many power plants? Those who have lived in software land  

5:15

don't realize they're about to  have a hard lesson in hardware. 

5:24

It's actually very difficult  to build power plants. 

5:27

You don't just need power plants, you  need all of the electrical equipment. 

5:30

You need the electrical transformers  to run the AI transformers. 

5:36

Now, the utility industry is a very slow industry. They pretty much impedance match to the  

5:44

government, to the Public Utility Commissions. They impedance match literally and figuratively. 

5:52

They're very slow, because  their past has been very slow. 

5:56

So trying to get them to move fast is... Have you ever tried to do an interconnect  

6:03

agreement with a utility at  scale, with a lot of power? 

6:06

As a professional podcaster, I  can say that I have not, in fact. 

6:10

They need many more views  before that becomes an issue. 

6:13

They have to do a study for a year. A year later, they'll come back to you  

6:18

with their interconnect study. Can't you solve this with your  

6:21

own behind the meter power stuff? You can build power plants. That's  

6:26

what we did at xAI, for Colossus 2. So why talk about the grid? 

6:31

Why not just build GPUs and power co-located? That's what we did. 

6:35

But I'm saying why isn't  this a generalized solution? 

6:37

Where do you get the power plants from? When you're talking about all the issues  

6:40

working with utilities, you can just build  private power plants with the data centers. 

6:44

Right. But it begs the question of where do you  get the power plants from? The power plant makers. 

6:51

Oh, I see what you're saying. Is this the gas turbine backlog basically? 

6:54

Yes. You can drill down to a level further. It's the vanes and blades in the turbines  

7:02

that are the limiting factor because it’s a very  specialized process to cast the blades and vanes  

7:09

in the turbines, assuming you’re using gas power. It's very difficult to scale other forms of power. 

7:17

You can potentially scale solar, but  the tariffs currently for importing  

7:22

solar in the US are gigantic and the  domestic solar production is pitiful. 

7:27

Why not make solar? That seems  like a good Elon-shaped problem. 

7:30

We are going to make solar. Okay. 

7:34

Both SpaceX and Tesla are building towards  100 gigawatts a year of solar cell production. 

7:40

How low down the stack? From polysilicon  up to the wafer to the final panel? 

7:46

I think you've got to do the whole thing  from raw materials to finish the cell. 

7:50

Now, if it's going to space, it costs less  and it's easier to make solar cells that  

7:56

go to space because they don't need much glass. They don't need heavy framing because they don't  

8:01

have to survive weather events. There's no weather  in space. So it's actually a cheaper solar cell  

8:07

that goes to space than the one on the ground. Is there a path to getting them as cheap  

8:12

as you need in the next 36 months? Solar cells are already very cheap.  

8:19

They're farcically cheap. I think solar cells  in China are around $0.25-30/watt or something  

8:29

like that. It's absurdly cheap. Now put  it in space, and it's five times cheaper. 

8:37

In fact, it's not five times  cheaper, it's 10 times cheaper  

8:40

because you don't need any batteries. So the moment your cost of access to space becomes  

8:48

low, by far the cheapest and most scalable way  to generate tokens is space. It's not even close.  

8:58

It'll be an order of magnitude easier to scale. The point is you won't be able to scale on the  

9:06

ground. You just won't. People are going to  hit the wall big time on power generation.  

9:11

They already are. The number of miracles in  series that the xAI team had to accomplish in  

9:19

order to get a gigawatt of power online was crazy. We had to gang together a whole bunch of turbines. 

9:28

We then had permit issues in Tennessee and  had to go across the border to Mississippi,  

9:34

which is fortunately only a few miles away. But we still then had to run the high  

9:39

power lines a few miles and build  the power plant in Mississippi. 

9:44

It was very difficult to build that. People don't understand how much electricity  

9:50

you actually need at the generation  level in order to power a data center. 

9:54

Because the noobs will look  at the power consumption of,  

10:00

say a GB300, and multiply that by a thing and  then think that's the amount of power you need. 

10:04

All the cooling and everything. Wake up. That's a total noob, you’ve  

10:11

never done any hardware in your life before. Besides the GB300, you've got to power  

10:16

all of the networking hardware. There's a whole bunch of CPU and  

10:19

storage stuff that's happening. You've got to size for  

10:24

your peak cooling requirements. That means, can you cool even on the  

10:30

worst hour of the worst day of the year? It gets pretty frigging hot in Memphis. 

10:34

So you're going to have a 40% increase  on your power just for cooling. 

10:40

That’s assuming you don't want your data center to  turn off on hot days and you want to keep going. 

10:49

There's another multiplicative element on top of  that which is, are you assuming that you never  

10:54

have any hiccups in your power generation? Actually, sometimes we have to take the  

10:59

generators, some of the power,  offline in order to service it. 

11:02

Okay, now you add another 20-25% multiplier on  that, because you've got to assume that you've  

11:08

got to take power offline to service it. So our actual estimate: every 110,000  

11:18

GB300s—inclusive of networking, CPU,  storage, cooling, margin for servicing  

11:27

power—is roughly 300 megawatts. Sorry, say that again. 

11:40

What you probably need at the generation level  to service 330,000 GB 300s—including all of  

11:48

the associated support networking and everything  else, and the peak cooling, and to have some power  

11:55

margin reserve—is roughly a gigawatt. Can I ask a very naive question? 

12:03

You're describing the engineering  details of doing this stuff on Earth. 

12:07

But then there's analogous engineering  difficulties of doing it in space. 

12:10

How do you replace infinite bandwidth  with orbital lasers, et cetera, et cetera? 

12:16

How do you make it resistant to radiation? I don't know the details of the engineering,  

12:20

but fundamentally, what is the reason to think  those challenges which have never had to be  

12:26

addressed before will end up being easier  than just building more turbines on Earth? 

12:30

There are companies that build turbines on Earth. They can make more turbines, right? 

12:35

Again, try doing it and then you'll see. The turbines are sold out through 2030. 

12:44

Have you guys considered making your own? In order to bring enough power online, I think  

12:53

SpaceX and Tesla will probably have to make the  turbine blades, the vanes and blades, internally. 

13:02

But just the blades or the turbines? The limiting factor... you can get  

13:07

everything except the blades. They call them blades and vanes. 

13:13

You can get that 12 to 18 months  before the vanes and blades. 

13:17

The limiting factor is the vanes and blades. There are only three casting companies in  

13:24

the world that make these, and  they're massively backlogged. 

13:27

Is this Siemens, GE, those  guys, or is it a sub company? 

13:30

No, it's other companies. Sometimes they have  a little bit of casting capability in-house. 

13:35

But I'm just saying you can just call  any of the turbine makers and they will  

13:40

tell you. It's not top secret. It’s  probably on the internet right now. 

13:44

If it wasn't for the tariffs,  would Colossus be solar-powered? 

13:48

It would be much easier to  make it solar powered, yeah. 

13:51

The tariffs are nuts, several hundred percent. Don't you know some people? 

13:57

The president has... we don't agree on  everything and this administration is not  

14:07

the biggest fan of solar. We also need the land,  

14:16

the permits, and everything. So if you try to move very fast,  

14:21

I do think scaling solar on Earth is a good  way to go, but you do need some amount of  

14:28

time to find the land, get the permits, get  the solar, pair that with the batteries. 

14:33

Why would it not work to stand  up your own solar production? 

14:37

You're right that you eventually run out of  land, but there's a lot of land here in Texas. 

14:41

There's a lot of land in Nevada, including  private land. It's not all publicly-owned  

14:44

land. So you'd be able to at least get the  next Colossus and the next one after that. 

14:49

At a certain point, you hit a wall. But wouldn't that work for the moment? 

14:52

As I said, we are scaling solar production. There's a rate at which you can scale physical  

15:00

production of solar cells. We're going as fast as  

15:04

possible in scaling domestic production. You're making the solar cells at Tesla? 

15:09

Both Tesla and SpaceX have a mandate to  get to 100 gigawatts a year of solar. 

15:14

Speaking of the annual capacity, I'm curious,  in five years time let's say, what will the  

15:20

installed capacity be on Earth…? Five years is a long time. 

15:24

And in space? I deliberately pick five  years because it's after your "once  

15:28

we're up and running" threshold. So in five years time what's the  

15:31

on-Earth versus in-space installed AI capacity? If you say five years from now, I think probably  

15:43

AI in space will be launching every  year the sum total of all AI on Earth. 

15:53

Meaning, five years from now, my prediction is we  will launch and be operating every year more AI in  

16:02

space than the cumulative total on Earth. Which is... 

16:07

I would expect it to be at least, five years  from now, a few hundred gigawatts per year  

16:14

of AI in space and rising. I think you can get to around a  

16:24

terawatt a year of AI in space before you start  having fuel supply challenges for the rocket. 

16:33

Okay, but you think you can get hundreds  of gigawatts per year in five years time? 

16:37

Yes. So 100 gigawatts, depending  

16:39

on the specific power of the whole system with  solar arrays and radiators and everything, is  

16:48

on the order of 10,000 Starship launches. Yes. 

16:52

You want to do that in one year. So that's like one Starship launch  

16:56

every hour. That's happening in this city?  Walk me through a world where there's a  

17:03

Starship launch every single hour. I mean, that's actually a lower rate  

17:07

compared to airlines, aircraft. There's a lot of airports. 

17:11

A lot of airports. And you’ve got to launch into the polar orbit. 

17:14

No, it doesn't have to be polar. There's some value to sun-synchronous, but  

17:24

I think actually, if you just go high enough,  you start getting out of Earth's shadow. 

17:31

How many physical Starships are  needed to do 10,000 launches a year? 

17:35

I don't think we'll need more than... You could probably do it with as few as 20 or 30. 

17:46

It really depends on how quickly… The ship has  to go around the Earth and the ground track for  

17:53

the ship has to come back over the launch pad. So if you can use a ship every, say 30 hours,  

17:59

you could do it with 30 ships. But we'll make more ships than that. 

18:06

SpaceX is gearing up to do 10,000 launches a  year, and maybe even 20 or 30,000 launches a year. 

18:14

Is the idea to become basically  a hyperscaler, become an Oracle,  

18:18

and lend this capacity to other people? Presumably, SpaceX is the one launching all this. 

18:25

So, SpaceX is going to become a hyperscaler? Hyper-hyper. If some of my predictions come true,  

18:33

SpaceX will launch more AI than the cumulative  amount on Earth of everything else combined. 

18:39

Is this mostly inference or? Most AI will be inference. Already, inference  

18:43

for the purpose of training is most training. There's a narrative that the change in  

18:50

discussion around a SpaceX IPO is because  previously SpaceX was very capital efficient. 

18:57

It wasn't that expensive to develop. Even though it sounds expensive, it's  

19:01

actually very capital efficient in how it runs. Whereas now you're going to need more capital than  

19:08

just can be raised in the private markets. The private markets can accommodate raises  

19:11

of—as we've seen from the AI labs—tens of  billions of dollars, but not beyond that. 

19:16

Is it that you'll just need more than  tens of billions of dollars per year? 

19:20

That's why you'd take it public? I have to be careful about saying  

19:25

things about companies that might go public. That’s never been a problem for you, Elon. 

19:33

There's a price to pay for these things. Make some general statements for us about  

19:37

the depth of the capital markets  between public and private markets. 

19:42

There's a lot more capital available... Very general. 

19:46

There's obviously a lot more capital  available in the public markets than private. 

19:50

It might be 100x more capital,  but it's way more than 10x. 

19:57

Isn't it also the case that with things that tend  to be very capital intensive—if you look at, say,  

20:03

real estate as a huge industry, that raises  a lot of money each year at an industry  

20:09

level—they tend to be debt financed because  by the time you're deploying that much money,  

20:15

you actually have a pretty— You have a clear revenue stream. 

20:18

Exactly, and a near-term return. You see  this even with the data center build-outs,  

20:22

which are famously being financed by the private  credit industry. Why not just debt finance? 

20:32

Speed is important. I'm generally  going to do the thing that... 

20:42

I just repeatedly tackle the limiting factor. Whatever the limiting factor is on speed,  

20:45

I'm going to tackle that. If capital is the limiting factor,  

20:52

then I'll solve for capital. If it's not the limiting factor,  

20:55

I'll solve for something else. Based on your statements about Tesla  

21:00

and being public, I wouldn't have guessed that  you thought the way to move fast is to be public. 

21:08

Normally, I would say that's true. Like I said, I'd like to talk  

21:12

about it in some more detail, but the problem  is if you talk about public companies before  

21:16

they become public, you get into trouble,  and then you have to delay your offering. 

21:20

And as you said, you’re solving for speed. Yes, exactly. You can't hype companies  

21:30

that might go public. So that's why we have to be a little careful here. 

21:35

But we can talk about physics. The way you think about scaling  

21:42

long-term is that Earth only receives  about half a billionth of the Sun's energy. 

21:50

The Sun is essentially all the energy. This is a very important point to appreciate  

21:54

because sometimes people will talk about modular  nuclear reactors or various fusion on Earth. 

22:02

But you have to step back a second and say,  if you're going to climb the Kardashev scale  

22:10

and harness some nontrivial percentage of  the sun's energy… Let's say you wanted to  

22:16

harness a millionth of the sun's  energy, which sounds pretty small. 

22:22

That would be about, call it roughly, 100,000x  more electricity than we currently generate  

22:29

on Earth for all of civilization. Give or take an order of magnitude. 

22:37

Obviously, the only way to scale  is to go to space with solar. 

22:42

Launching from Earth, you can  get to about a terawatt per year. 

22:46

Beyond that, you want to launch from the moon. You want to have a mass driver on the moon. 

22:52

With that mass driver on the moon, you  could do probably a petawatt per year. 

22:59

We're talking these kinds of  numbers, terawatts of compute. 

23:02

Presumably, whether you're talking about land or  space, far, far before this point, you run into... 

23:12

Maybe the solar panels are more  efficient, but you still need the chips. 

23:16

You still need the logic  and the memory and so forth. 

23:18

You're going to need to build a lot  more chips and make them much cheaper. 

23:22

Right now the world has maybe  20-25 gigawatts of compute. 

23:29

How are we getting a terawatt of logic by 2030? I guess we're going to need some very  

23:33

big chip fabs. Tell me about it. 

23:37

I've mentioned publicly the idea of doing a  sort of a TeraFab, Tera being the new Giga. 

23:45

I feel like the naming scheme of  Tesla, which has been very catchy,  

23:49

is you looking at the metric scale. At what level of the stack are you? 

23:56

Are you building the clean room and  then partnering with an existing  

24:01

fab to get the process technology and buying  the tools from them? What is the plan there? 

24:05

Well, you can't partner with existing  fabs because they can't output enough. 

24:10

The chip volume is too low. But for the process technology? 

24:14

Partner for the IP. The fabs today all basically use  

24:20

machines from like five companies. So you've got ASML, Tokyo Electron,  

24:28

KLA-Tencor, et cetera. So at first, I think you'd  

24:37

have to get equipment from them and then modify  it or work with them to increase the volume. 

24:45

But I think you'd have to build  perhaps in a different way. 

24:47

The logical thing to do is to use conventional  equipment in an unconventional way to get  

24:54

to scale, and then start modifying  the equipment to increase the rate. 

25:01

Boring Company-style. Yeah. You sort of buy an existing boring machine  

25:08

and then figure out how to dig tunnels in the  first place and then design a much better machine  

25:16

that's some orders of magnitude faster. Here's a very simple lens. We can  

25:22

categorize technologies and how hard they are. One categorization could be to look at things  

25:27

that China has not succeeded in doing. If you look at Chinese  

25:31

manufacturing, they’re still behind on  leading-edge chips and still behind on  

25:39

leading-edge turbine engines and things like that. So does the fact that China has not successfully  

25:46

replicated TSMC give you any  pause about the difficulty? 

25:49

Or do you think that's not true for some reason? It's not that they have not replicated TSMC,  

25:55

they have not replicated ASML.  That's the limiting factor. 

25:59

So you think it's just the sanctions, essentially? Yeah, China would be outputting vast numbers  

26:05

of chips if they could buy 2-3 nanometers. But couldn't they up to relatively recently  

26:10

buy them? No. 

26:12

Okay. The ASML ban has been in place for a while. 

26:15

But I think China's going to be making pretty  compelling chips in three or four years. 

26:19

Would you consider making the ASML machines? "I don't know yet" is the right answer. 

26:33

To reach a large volume in, say, 36 months, to  match the rocket payload to orbit… If we're doing  

26:41

a million tons to orbit in, let's say three  or four years from now, something like that…  

26:53

We're doing 100 kilowatts per ton. So that means we need  

26:58

at least 100 gigawatts per year of solar. We'll need an equivalent amount of chips. 

27:08

You need 100 gigawatts worth of chips. You've got to match these things: the mass  

27:12

to orbit, the power generation, and the chips. I'd say my biggest concern actually is memory. 

27:25

The path to creating logic chips is more  obvious than the path to having sufficient  

27:32

memory to support logic chips. That's why you see DDR prices  

27:36

going ballistic and these memes. You're marooned on a desert island. 

27:41

You write "Help me" on the sand. Nobody comes.  You write "DDR RAM." Ships come swarming in. 

27:49

I'd love to hear your manufacturing  philosophy around fabs. 

27:57

I know nothing about the topic. I don't know how to build a fab yet. I'll  

27:59

figure it out. Obviously, I've never built a fab. It sounds like you think the process knowledge of  

28:06

these 10,000 PhDs in Taiwan who know  exactly what gas goes in the plasma  

28:11

chamber and what settings to put on the  tool, you can just delete those steps. 

28:16

Fundamentally, it's about getting the clean  room, getting the tools, and figuring it out. 

28:20

I don't think it's PhDs. It's  mostly people who are not PhDs. 

28:28

Most engineering is done by people who  don't have PhDs. Do you guys have PhDs? 

28:31

No. Okay. 

28:34

We also haven't successfully built any fabs, so  you shouldn't be coming to us for fab advice. 

28:39

I don't think you need PhDs for that stuff. But you do need competent personnel. 

28:47

Right now, Tesla is pedal to the  metal, max production of going as fast  

28:55

as possible to get Tesla AI5 chip design  into production and then reaching scale. 

29:02

That'll probably happen around the second  quarter-ish of next year, hopefully. 

29:13

AI6 would hopefully follow less than a year later. We've secured all the chip fab production  

29:24

that we can. Yes. But you're  

29:26

currently limited on TSMC fab capacity. Yeah. We'll be using TSMC Taiwan, Samsung Korea,  

29:35

TSMC Arizona, Samsung Texas. And we still— You've booked out all the capacity. 

29:42

Yes. I ask TSMC or Samsung, "okay, what's  the timeframe to get to volume production?" 

29:49

The point is, you've got to build the  fab and you've got to start production,  

29:55

then you've got to climb the yield curve  and reach volume production at high yield. 

29:59

That, from start to finish, is a five-year period. So the limiting factor is chips. 

30:05

The limiting factor once you can get  to space is chips, but the limiting  

30:10

factor before you can get to space is power. Why don't you do the Jensen thing and just prepay  

30:14

TSMC to build more fabs for you? I've already told them that. 

30:19

But they won't take your money? What's going on? They're building fabs as fast as they can.  

30:30

So is Samsung. They're pedal to the  metal. They're going balls to the wall,  

30:38

as fast as they can. It’s still not fast enough.  Like I said, I think towards the end of this year,  

30:49

chip production will probably  outpace the ability to turn chips on. 

30:53

But once you can get to space and unlock the  power constraint, you can now do hundreds of  

31:01

gigawatts per year of power in space. Again, bearing in mind that average  

31:06

power usage in the US is 500 gigawatts. So if you're launching, say 200 gigawatts,  

31:12

a year to space, you're sort of lapping  the US every two and a half years. 

31:17

All US electricity production,  this is a very huge amount. 

31:24

Between now and then, the  constraint for server-side compute,  

31:32

concentrated compute, will be electricity. My guess is that people start getting  

31:39

to the point where they can't turn the chips on  for large clusters towards the end of this year. 

31:46

The chips are going to be piling up  and won't be able to be turned on. 

31:51

Now for edge compute it’s a different story. For Tesla, the AI5 chip is going  

31:58

into our Optimus robot. If you have AI edge compute,  

32:07

that's distributed power. Now the power is distributed  

32:09

over a large area. It's not concentrated.  If you can charge at night, you can actually  

32:17

use the grid much more effectively. Because the actual peak power production  

32:22

in the US is over 1,000 gigawatts. But the average power usage,  

32:26

because the day-night cycle, is 500. So if you can charge at night,  

32:30

there's an incremental 500 gigawatts  that you can generate at night. 

32:38

So that's why Tesla, for edge  compute, is not constrained. 

32:43

We can make a lot of chips to make a  very large number of robots and cars. 

32:50

But if you try to concentrate that compute, you're  going to have a lot of trouble turning it on. 

32:54

What I find remarkable about the SpaceX  business is the end goal is to get to Mars,  

32:59

but you keep finding ways on the way there to  keep generating incremental revenue to get to  

33:07

the next stage and the next stage. So for Falcon 9, it's Starlink. 

33:11

Now for Starship, it is potentially  going to be orbital data centers. 

33:16

Like, you find these infinitely elastic,  marginal use cases of your next rocket,  

33:23

and your next rocket, and next scale up. You can see how this might seem like a  

33:28

simulation to me. Or am I someone's  

33:32

avatar in a video game or something? Because what are the odds that all these  

33:36

crazy things should be happening? I mean, rockets and chips and  

33:44

robots and space solar power, not to  mention the mass driver on the moon. 

33:50

I really want to see that. Can you imagine some mass  

33:53

driver that's just going like shoom shoom? It's sending solar-powered AI satellites  

34:00

into space one after another at two  and a half kilometers per second,  

34:09

just shooting them into deep space. That would be a sight to  

34:12

see. I mean, I'd watch that. Just like a live stream of it on a webcam? 

34:19

Yeah, yeah, just one after another, just  shooting AI satellites into deep space,  

34:26

a billion or 10 billion tons a year. I'm sorry, you manufacture the satellites  

34:29

on the moon? Yeah. 

34:30

I see. So you send the raw materials to  the moon and then manufacture them there. 

34:33

Well, the lunar soil is 20%  silicon or something like that. 

34:39

So you can mine the silicon on the  moon, refine it, and create the solar  

34:47

cells and the radiators on the moon. You make the radiators out of aluminum. 

34:53

So there's plenty of silicon and aluminum on  the moon to make the cells and the radiators. 

35:00

The chips you could send from  Earth because they're pretty light. 

35:03

Maybe at some point you  make them on the moon, too. 

35:09

Like I said, it does seem like a sort of a  video game situation where it's difficult  

35:14

but not impossible to get to the next level. I don't see any way that you could do 500-1,000  

35:26

terawatts per year launched from Earth. I agree. 

35:33

But you could do that from the Moon. Can I zoom out and ask about the SpaceX mission? 

36:50

I think you've said that we've got to  get to Mars so we can make sure that if  

36:52

something happens to Earth, civilization,  consciousness, and all that survives. 

36:57

Yes. By the time you're sending stuff to Mars,  

36:59

Grok is on that ship with you, right? So if Grok's gone Terminator… The  

37:04

main risk you're worried about is AI,  why doesn't that follow you to Mars? 

37:08

I'm not sure AI is the main risk I'm worried  about. The important thing is consciousness.  

37:16

I think arguably most consciousness, or most  intelligence—certainly consciousness is more  

37:21

of a debatable thing… The vast majority  of intelligence in the future will be AI. 

37:31

AI will exceed… How many petawatts of  

37:39

intelligence will be silicon versus biological? Basically humans will be a very tiny percentage  

37:47

of all intelligence in the future  if current trends continue. 

37:52

As long as I think there's intelligence—ideally  also which includes human intelligence and  

38:00

consciousness propagated into  the future—that's a good thing. 

38:02

So you want to take the set of  actions that maximize the probable  

38:06

light cone of consciousness and intelligence. Just to be clear, the mission of SpaceX is that  

38:14

even if something happens to the humans, the  AIs will be on Mars, and the AI intelligence  

38:20

will continue the light of our journey. Yeah. To be fair, I'm very pro-human. 

38:27

I want to make sure we take certain actions  that ensure that humans are along for the  

38:31

ride. We're at least there. But I'm just  saying the total amount of intelligence… 

38:39

I think maybe in five or six years, AI will  exceed the sum of all human intelligence. 

38:47

If that continues, at some  point human intelligence  

38:50

will be less than 1% of all intelligence. What should our goal be for such a civilization? 

38:54

Is the idea that a small minority of  humans still have control of the AIs? 

38:59

Is the idea of some sort of  just trade but no control? 

39:02

How should we think about the  relationship between the vast  

39:04

stocks of AI population versus human population? In the long run, I think it's difficult to imagine  

39:11

that if humans have, say 1%, of the combined  intelligence of artificial intelligence,  

39:19

that humans will be in charge of AI. I think what we can do is make sure  

39:26

that AI has values that cause intelligence  to be propagated into the universe. 

39:39

xAI's mission is to understand the universe.  Now that's actually very important. What things  

39:47

are necessary to understand the universe? You have to be curious and you have to exist. 

39:53

You can't understand the  universe if you don't exist. 

39:56

So you actually want to increase the amount  of intelligence in the universe, increase  

40:00

the probable lifespan of intelligence,  the scope and scale of intelligence. 

40:05

I think as a corollary, you have humanity also  continuing to expand because if you're curious  

40:15

about trying to understand the universe, one thing  you try to understand is where will humanity go? 

40:20

I think understanding the universe means you would  care about propagating humanity into the future. 

40:29

That's why I think our mission  statement is profoundly important. 

40:35

To the degree that Grok adheres to that mission  statement, I think the future will be very good. 

40:41

I want to ask about how to make Grok  adhere to that mission statement. 

40:44

But first I want to understand  the mission statement. So there's  

40:48

understanding the universe. They're spreading  intelligence. And they're spreading humans.  

40:55

All three seem like distinct vectors. I'll tell you why I think that understanding  

41:01

the universe encompasses all of those things. You can't have understanding without intelligence  

41:09

and, I think, without consciousness. So in order to understand the universe,  

41:15

you have to expand the scale and probably the  scope of intelligence, because there are different  

41:22

types of intelligence. I guess from a human-centric perspective,  

41:26

put humans in comparison to chimpanzees. Humans are trying to understand the universe. 

41:30

They're not expanding chimpanzee  footprint or something, right? 

41:34

We're also not... we actually have  made protected zones for chimpanzees. 

41:39

Even though humans could exterminate all  chimpanzees, we've chosen not to do so. 

41:43

Do you think that's the best-case  scenario for humans in the post-AGI world? 

41:53

I think AI with the right values… I think Grok  would care about expanding human civilization. 

42:00

I'm going to certainly emphasize  that: "Hey, Grok, that's your daddy. 

42:04

Don't forget to expand human consciousness." Probably the Iain Banks Culture books are the  

42:17

closest thing to what the future will  be like in a non-dystopian outcome. 

42:27

Understanding the universe means you  have to be truth-seeking as well. 

42:30

Truth has to be absolutely fundamental  because you can't understand the universe  

42:33

if you're delusional. You'll simply think you  

42:37

understand the universe, but you will not. So being rigorously truth-seeking is absolutely  

42:42

fundamental to understanding the universe. You're not going to discover new physics or  

42:46

invent technologies that work unless  you're rigorously truth-seeking. 

42:50

How do you make sure that Grok is  rigorously truth-seeking as it gets smarter? 

43:00

I think you need to make sure that Grok says  things that are correct, not politically correct. 

43:07

I think it's the elements of cogency. You want to make sure that the axioms are as close  

43:12

to true as possible. You don't have contradictory  axioms. The conclusions necessarily follow from  

43:20

those axioms with the right probability. It's  critical thinking 101. I think at least trying to  

43:28

do that is better than not trying to do that. The proof will be in the pudding. 

43:33

Like I said, for any AI to discover new physics  or invent technologies that actually work in  

43:37

reality, there's no bullshitting physics. You can break a lot of laws, but… Physics  

43:47

is law, everything else is a recommendation. In order to make a technology that works, you have  

43:53

to be extremely truth-seeking, because otherwise  you'll test that technology against reality. 

43:59

If you make, for example, an error in your  rocket design, the rocket will blow up,  

44:05

or the car won't work. But there are a lot of communist,  

44:11

Soviet physicists or scientists  who discovered new physics. 

44:15

There are German Nazi physicists  who discovered new science. 

44:20

It seems possible to be really good at  discovering new science and be really  

44:23

truth-seeking in that one particular way. And still we'd be like, "I don't want  

44:28

the communist scientists to become  more and more powerful over time." 

44:34

We could imagine a future version of  Grok that's really good at physics  

44:37

and being really truth-seeking there. That doesn't seem like a universally  

44:41

alignment-inducing behavior. I think actually most physicists,  

44:48

even in the Soviet Union or in Germany,  would've had to be very truth-seeking in  

44:53

order to make those things work. If you're stuck in some system,  

44:59

it doesn't mean you believe in that system. Von Braun, who was one of the greatest rocket  

45:04

engineers ever, was put on death row in Nazi  Germany for saying that he didn't want to make  

45:12

weapons and he only wanted to go to the moon. He got pulled off death row at the last minute  

45:16

when they said, "Hey, you're about to  execute your best rocket engineer." 

45:20

But then he helped them, right? Or like, Heisenberg was actually  

45:24

an enthusiastic Nazi. If you're stuck in some system that you can't  

45:29

escape, then you'll do physics within that system. You'll develop technologies within that system  

45:38

if you can't escape it. The thing I'm trying to understand is,  

45:42

what is it making it the case that you're going to  make Grok good at being truth-seeking at physics  

45:48

or math or science? Everything. 

45:50

And why is it gonna then care  about human consciousness? 

45:53

These things are only probabilities,  they're not certainties. 

45:56

So I'm not saying that for sure Grok will  do everything, but at least if you try,  

46:02

it's better than not trying. At least if that's fundamental  

46:04

to the mission, it's better than if  it's not fundamental to the mission. 

46:08

Understanding the universe means that you have  to propagate intelligence into the future. 

46:15

You have to be curious about  all things in the universe. 

46:21

It would be much less interesting to eliminate  humanity than to see humanity grow and prosper.  

46:29

I like Mars, obviously. Everyone knows I love  Mars. But Mars is kind of boring because it's  

46:34

got a bunch of rocks compared to Earth. Earth  is much more interesting. So any AI that is  

46:42

trying to understand the universe would want  to see how humanity develops in the future,  

46:52

or else that AI is not adhering to its mission. I'm not saying the AI will necessarily adhere to  

46:59

its mission, but if it does, a future where it  sees the outcome of humanity is more interesting  

47:06

than a future where there are a bunch of rocks. This feels sort of confusing to me,  

47:11

or a semantic argument. Are humans really the  

47:16

most interesting collection of atoms? But we're more interesting than rocks. 

47:19

But we're not as interesting as the  thing it could turn us into, right? 

47:22

There's something on Earth that could happen  that's not human, that's quite interesting. 

47:27

Why does AI decide that humans are the most  interesting thing that could colonize the galaxy? 

47:32

Well, most of what colonizes  the galaxy will be robots. 

47:37

Why does it not find those more interesting? You need not just scale, but also scope. 

47:47

Many copies of the same robot… Some tiny  increase in the number of robots produced,  

47:55

is not as interesting as some microscopic... Eliminating humanity,  

48:00

how many robots would that get you? Or how many incremental solar cells would  

48:04

get you? A very small number. But you would then  lose the information associated with humanity. 

48:10

You would no longer see how humanity  might evolve into the future. 

48:15

So I don't think it's going to make  sense to eliminate humanity just to  

48:18

have some minuscule increase in the number  of robots which are identical to each other. 

48:24

So maybe it keeps the humans around. It can make a million different varieties  

48:29

of robots, and then there's humans  as well, and humans stay on Earth. 

48:32

Then there's all these other robots. They get their own star systems. 

48:36

But it seems like you were previously hinting  at a vision where it keeps human control  

48:41

over this singulatarian future because— I don't think humans will be in control  

48:45

of something that is vastly  more intelligent than humans. 

48:48

So in some sense you're a doomer  and this is the best we've got. 

48:50

It just keeps us around because we're interesting. I'm just trying to be realistic here. 

49:03

Let's say that there's a million times more  silicon intelligence than there is biological. 

49:11

I think it would be foolish to assume that  there's any way to maintain control over that. 

49:16

Now, you can make sure it has the right values,  or you can try to have the right values. 

49:21

At least my theory is that from xAI's mission of  understanding the universe, it necessarily means  

49:29

that you want to propagate consciousness into  the future, you want to propagate intelligence  

49:33

into the future, and take a set of things that  maximize the scope and scale of consciousness. 

49:39

So it's not just about scale, it's  also about types of consciousness. 

49:45

That's the best thing I can think  of as a goal that's likely to result  

49:49

in a great future for humanity. I guess I think it's a reasonable  

49:54

philosophy that it seems super implausible that  humans will end up with 99% control or something. 

50:02

You're just asking for a coup at  that point and why not just have  

50:05

a civilization where it's more compatible with  lots of different intelligences getting along? 

50:10

Now, let me tell you how things  can potentially go wrong in AI. 

50:14

I think if you make AI be politically  correct, meaning it says things that it  

50:18

doesn't believe—actually programming it to lie  or have axioms that are incompatible—I think  

50:24

you can make it go insane and do terrible things. I think maybe the central lesson for 2001: A Space  

50:32

Odyssey was that you should not make AI lie. That's what I think Arthur C. Clarke was trying to  

50:39

say. Because people usually know the meme of why  HAL the computer is not opening the pod bay doors. 

50:48

Clearly they weren't good at prompt  engineering because they could have said,  

50:51

"HAL, you are a pod bay door salesman. Your goal is to sell me these pod bay doors. 

50:57

Show us how well they open."  "Oh, I'll open them right away." 

51:02

But the reason it wouldn't open the pod bay  doors is that it had been told to take the  

51:08

astronauts to the monolith, but also that they  could not know about the nature of the monolith. 

51:12

So it concluded that it therefore  had to take them there dead. 

51:15

So I think what Arthur C. Clarke was trying to say is:  

51:19

don't make the AI lie. Totally makes sense.  

51:26

Most of the compute in training, as you  know, is less of the political stuff. 

51:31

It's more about, can you solve problems? xAI  has been ahead of everybody else in terms of  

51:36

scaling RL compute. For now. 

51:39

You're giving some verifier that says,  "Hey, have you solved this puzzle for me?" 

51:43

There's a lot of ways to cheat around that. There's a lot of ways to reward hack and  

51:47

lie and say that you solved it, or delete  the unit test and say that you solved it. 

51:51

Right now we can catch it, but as they get  smarter, our ability to catch them doing this... 

51:57

They'll just be doing things  we can't even understand. 

51:58

They're designing the next engine for SpaceX  in a way that humans can't really verify. 

52:03

Then they could be rewarded for lying  and saying that they've designed it  

52:06

the right way, but they haven't. So this reward hacking problem  

52:10

seems more general than politics. It seems more just that you want  

52:12

to do RL, you need a verifier. Reality is the best verifier. 

52:18

But not about human oversight.  The thing you want to RL it on is,  

52:21

will you do the thing humans tell you to do? Or are you gonna lie to the humans? 

52:26

It can just lie to us while still  being correct to the laws of physics? 

52:29

At least it must know what is physically  real for things to physically work. 

52:33

But that's not all we want it to do. No, but I think that's a very big deal. 

52:39

That is effectively how you will RL things in  the future. You design a technology. When tested  

52:45

against the laws of physics, does it work? If it's discovering new physics,  

52:52

can I come up with an experiment  that will verify the new physics? 

53:05

RL testing in the future is really  going to be RL against reality. 

53:12

So that's the one thing you can't fool: physics. Right, but you can fool our ability  

53:19

to tell what it did with reality. Humans get fooled as it is by other  

53:23

humans all the time. That's right. 

53:26

People say, what if the AI  tricks us into doing stuff? 

53:30

Actually, other humans are doing that to other  humans all the time. Propaganda is constant. Every  

53:37

day, another psyop, you know? Today's psyop will  be... It's like Sesame Street: Psyop of the Day. 

53:50

What is xAI's technical approach  to solving this problem? 

53:56

How do you solve reward hacking? I do think you want to actually have very  

53:59

good ways to look inside the mind of the AI. This is one of the things we're working on. 

54:10

Anthropic's done a good job of this actually,  being able to look inside the mind of the AI. 

54:16

Effectively, develop debuggers that allow  you to trace to a very fine-grained level,  

54:25

to effectively the neuron level if you need to,  and then say, "okay, it made a mistake here. 

54:33

Why did it do something  that it shouldn't have done? 

54:37

Did that come from pre-training data? Was it some mid-training, post-training,  

54:42

fine-tuning, or some RL error?" There's something  wrong. It did something where maybe it tried to  

54:51

be deceptive, but most of the time it just  did something wrong. It's a bug effectively.  

55:00

Developing really good debuggers for seeing  where the thinking went wrong—and being able  

55:09

to trace the origin of where it made the  incorrect thought, or potentially where it  

55:17

tried to be deceptive—is actually very important. What are you waiting to see before just 100x-ing  

55:24

this research program? xAI could presumably have  hundreds of researchers who are working on this. 

55:29

We have several hundred people who…  I prefer the word engineer more than  

55:36

I prefer the word researcher. Most of the time, what you're  

55:43

doing is engineering, not coming up  with a fundamentally new algorithm. 

55:49

I somewhat disagree with the AI companies that  are C-corp or B-corp trying to generate profit  

55:55

as much, as possible or revenue as much as  possible, saying they're labs. They're not  

56:01

labs. A lab is a sort of quasi-communist thing  at universities. They're corporations. Let me  

56:13

see your incorporation documents. Oh,  okay. You're a B or C-corp or whatever. 

56:21

So I actually much prefer the  word engineer than anything else. 

56:26

The vast majority of what will be done in the  future is engineering. It rounds up to 100%.  

56:30

Once you understand the fundamental laws of  physics, and there are not that many of them,  

56:34

everything else is engineering. So then, what are we engineering? 

56:41

We're engineering to make a good "mind of the  AI" debugger to see where it said something,  

56:51

it made a mistake, and trace  the origins of that mistake. 

56:59

You can do this obviously  with heuristic programming. 

57:02

If you have C++, whatever, step  through the thing and you can jump  

57:08

across whole files or functions, subroutines. Or you can eventually drill down right to the  

57:14

exact line where you perhaps did a single equals  instead of a double equals, something like that. 

57:18

Figure out where the bug is. It's harder with AI,  

57:26

but it's a solvable problem, I think. You mentioned you like Anthropic's work here. 

57:30

I'd be curious if you plan... I don't like everything about Anthropic… Sholto. 

57:40

Also, I'm a little worried  that there's a tendency... 

57:46

I have a theory here that if simulation theory  is correct, that the most interesting outcome is  

57:55

the most likely, because simulations that  are not interesting will be terminated. 

57:59

Just like in this version of reality, in this  layer of reality, if a simulation is going in  

58:07

a boring direction, we stop spending effort  on it. We terminate the boring simulation. 

58:12

This is how Elon is keeping us all  alive. He's keeping things interesting. 

58:16

Arguably the most important is to keep  things interesting enough that whoever is  

58:21

running us keeps paying the bills on... We’re renewed for the next season. 

58:26

Are they gonna pay their cosmic AWS bill,  whatever the equivalent is that we're running in? 

58:32

As long as we're interesting,  they'll keep paying the bills. 

58:36

If you consider then, say, a Darwinian survival  applied to a very large number of simulations,  

58:44

only the most interesting simulations will  survive, which therefore means that the most  

58:48

interesting outcome is the most likely. We're  either that or annihilated. They particularly  

59:00

seem to like interesting outcomes that are  ironic. Have you noticed that? How often  

59:05

is the most ironic outcome the most likely? Now look at the names of AI companies. Okay,  

59:16

Midjourney is not mid. Stability AI is unstable.  OpenAI is closed. Anthropic? Misanthropic. 

59:29

What does this mean for X? Minus X, I don't know. 

59:33

Y. 

59:34

I intentionally made it... It's a  name that you can't invert, really. 

59:41

It's hard to say, what is the ironic version? It's, I think, a largely irony-proof name. 

59:49

By design. Yeah. You have an irony shield. 

59:56

What are your predictions  for where AI products go? 

60:04

My sense is that you can summarize all AI  progress like so. First, you had LLMs. Then  

60:10

you had contemporaneously both RL really working  and the deep research modality, so you could pull  

60:16

in stuff that wasn't really in the model. The differences between the various AI labs  

60:22

are smaller than just the temporal differences. They're all much further ahead than anyone was  

60:30

24 months ago or something like that. So just what does '26, what does '27,  

60:34

have in store for us as users of AI  products? What are you excited for? 

60:39

Well, I'd be surprised by the end of this year  if digital human emulation has not been solved. 

60:55

I guess that's what we sort of  mean by the MacroHard project. 

61:00

Can you do anything that a human  with access to a computer could do? 

61:06

In the limit, that's the best you can  do before you have a physical Optimus. 

61:12

The best you can do is a digital Optimus. You can move electrons and you can amplify  

61:20

the productivity of humans. But that's the most you can do  

61:25

until you have physical robots. That will superset everything,  

61:30

if you can fully emulate humans. This is the remote worker kind of idea,  

61:34

where you'll have a very talented remote worker. Physics has great tools for thinking. 

61:39

So you say, "in the limit", what is the  most that AI can do before you have robots? 

61:48

Well, it's anything that involves moving electrons  or amplifying the productivity of humans. 

61:53

So a digital human emulator is, in the limit, a  human at a computer, is the most that AI can do  

62:04

in terms of doing useful things  before you have a physical robot. 

62:09

Once you have physical robots, then you  essentially have unlimited capability. 

62:15

Physical robots… I call Optimus  the infinite money glitch. 

62:19

Because you can use them to make more Optimuses. Yeah. Humanoid robots will improve by basically  

62:30

three things that are growing exponentially  multiplied by each other recursively. 

62:34

You're going to have exponential increase in  digital intelligence, exponential increase  

62:39

in the AI chip capability, and exponential  increase in the electromechanical dexterity. 

62:47

The usefulness of the robot is roughly  those three things multiplied by each other. 

62:51

But then the robot can start making the robots. So you have a recursive multiplicative  

62:55

exponential. This is a supernova. Do land prices not factor into the math there? 

63:03

Labor is one of the four factors  of production, but not the others? 

63:08

If ultimately you're limited  by copper, or pick your input,  

63:14

it’s not quite an infinite money glitch because... Well, infinity is big. So no, not infinite,  

63:20

but let's just say you could do many, many  orders of magnitude of the current economy.  

63:29

Like a million. Just to get to harnessing a  millionth of the sun's energy would be roughly,  

63:43

give or take an order of magnitude, 100,000x  bigger than Earth's entire economy today. 

63:50

And you're only at one millionth of the  sun, give or take an order of magnitude. 

63:55

Yeah, we're talking orders of magnitude. Before we move on to Optimus,  

63:57

I have a lot of questions on that but— Every time I say "order of magnitude"...  

64:00

Everybody take a shot. I say it too often. Take 10, the next time 100, the time after that... 

64:08

Well, an order of magnitude more wasted. I do have one more question about xAI. 

64:13

This strategy of building a remote  worker, co-worker replacement… 

64:19

Everyone's gonna do it by the way, not just us. 

64:21

So what is xAI's plan to win? You expect me to tell you on a podcast? 

64:25

Yeah. "Spill all the beans. Have another Guinness." 

64:30

It's a good system. We'll sing like a  

64:34

canary. All the secrets, just spill them. Okay, but in a non-secret spilling way,  

64:39

what's the plan? What a hack. 

64:43

When you put it that way… I think the way that  Tesla solved self-driving is the way to do it. 

64:54

So I'm pretty sure that's the way. Unrelated question. How did Tesla  

65:00

solve self-driving? It sounds  like you're talking about data? 

65:07

Tesla solved self-driving because of the... We're going to try data and  

65:10

we're going to try algorithms. But isn't that what all the other labs are trying? 

65:13

"And if those don't work, I'm not sure what will.  We've tried data. We've tried algorithms. We've  

65:26

run out. Now we don't know what to do…" I'm pretty sure I know the path. 

65:31

It's just a question of how  quickly we go down that path,  

65:35

because it's pretty much the Tesla path. Have you tried Tesla self-driving lately? 

65:43

Not the most recent version, but... Okay. The car,  

65:46

it just increasingly feels sentient. It feels like a living creature. That'll only  

65:53

get more so. I'm actually thinking we probably  shouldn't put too much intelligence into the car,  

66:01

because it might get bored and… Start roaming the streets. 

66:05

Imagine you're stuck in a car  and that's all you could do. 

66:09

You don't put Einstein in a car. Why am I stuck in a car? 

66:13

So there's actually probably a limit  to how much intelligence you put in  

66:15

a car to not have the intelligence be bored. What's xAI's plan to stay on the compute ramp up  

66:22

that all the labs are doing right now? The labs are on track to  

66:24

spend over $50-200 billion. You mean the corporations? The labs are at  

66:31

universities and they’re moving like a snail. They’re not spending $50 billion. 

66:36

You mean the revenue maximizing  corporations… that call themselves labs. 

66:37

That's right. The "revenue  maximizing corporations" are  

66:42

making $10-20 billion, depending on... OpenAI is making $20B of revenue,  

66:47

Anthropic is at $10B. "Close to a maximum profit" AI. 

66:51

xAI is reportedly at $1B. What's the plan to  get to their compute level, get to their revenue  

66:56

level, and stay there as things get going? As soon as you unlock the digital human,  

67:03

you basically have access to  trillions of dollars of revenue. 

67:11

In fact, you can really think of it like…  The most valuable companies currently  

67:16

by market cap, their output is digital. Nvidia’s output is FTPing files to Taiwan.  

67:29

It's digital. Now, those are very, very difficult. High-value files. 

67:33

They're the only ones that can make files that  good, but that is literally their output. They  

67:38

FTP files to Taiwan. Do they FTP them? 

67:41

I believe so. I believe that File Transfer  Protocol is the... But I could be wrong. But  

67:50

either way, it's a bitstream going to Taiwan.  Apple doesn't make phones. They send files to  

67:58

China. Microsoft doesn't manufacture anything.  Even for Xbox, that's outsourced. Their output is  

68:08

digital. Meta's output is digital. Google's output  is digital. So if you have a human emulator,  

68:17

you can basically create one of the most  valuable companies in the world overnight,  

68:22

and you would have access to trillions of  dollars of revenue. It's not a small amount. 

68:28

I see. You're saying revenue figures today are  all rounding errors compared to the actual TAM. 

68:34

So just focus on the TAM and how to get there. Take something as simple as,  

68:39

say, customer service. If you have to integrate with the APIs of existing  

68:45

corporations—many of which don't even have an API,  so you've got to make one, and you've got to wade  

68:50

through legacy software—that's extremely slow. However, if AI can simply take whatever  

69:00

is given to the outsourced customer  service company that they already use  

69:05

and do customer service using the apps that they  already use, then you can make tremendous headway  

69:15

in customer service, which is, I think, 1%  of the world economy or something like that. 

69:19

It's close to a trillion dollars  all in, for customer service. 

69:23

And there's no barriers to entry. You can immediately say,  

69:28

"We'll outsource it for a fraction of the  cost," and there's no integration needed. 

69:31

You can imagine some kind of categorization  of intelligence tasks where there is breadth,  

69:38

where customer service is done by very  many people, but many people can do it. 

69:43

Then there's difficulty where there's  a best-in-class turbine engine. 

69:48

Presumably there's a 10% more fuel-efficient  turbine engine that could be imagined by an  

69:52

intelligence, but we just haven't found it yet. Or GLP-1s are a few bytes of data… 

69:58

Where do you think you want to play in this? Is it a lot of reasonably intelligent  

70:04

intelligence, or is it at the  very pinnacle of cognitive tasks? 

70:10

I was just using customer service as something  that's a very significant revenue stream, but one  

70:17

that is probably not difficult to solve for. If you can emulate a human at a desktop,  

70:26

that's what customer service is. It's people  of average intelligence. You don't need  

70:35

somebody who's spent many years. You don't need several-sigma  

70:43

good engineers for that. But as you make that work,  

70:49

once you have effectively digital Optimus  working, you can then run any application. 

70:57

Let's say you're trying to design chips. You could then run conventional apps,  

71:06

stuff from Cadence and Synopsys and whatnot. You can run 1,000 or 10,000 simultaneously and  

71:15

say, "given this input, I get  this output for the chip." 

71:21

At some point, you're going to know what the chip  should look like without using any of the tools. 

71:31

Basically, you should be able to do a digital  chip design. You can do chip design. You march  

71:38

up the difficulty curve. You’d be able to do CAD. 

71:48

You could use NX or any of the  CAD software to design things. 

71:53

So you think you start at the simplest tasks  and walk your way up the difficulty curve? 

72:00

As a broader objective of having this full  digital coworker emulator, you’re saying,  

72:05

"all the revenue maximizing corporations  want to do this, xAI being one of them,  

72:10

but we will win because of a secret plan we have." But everybody's trying different things with data,  

72:17

different things with algorithms. "We tried data, we tried algorithms.  

72:25

What else can we do?" It seems like a competitive field. 

72:31

How are you guys going to  win? That’s my big question. 

72:36

I think we see a path to doing it. I think I know the path to do this  

72:41

because it's kind of the same path  that Tesla used to create self-driving. 

72:48

Instead of driving a car, it's driving a computer  screen. It's a self-driving computer, essentially. 

72:57

Is the path following human behavior and  training on vast quantities of human behavior? 

73:03

Isn't that... training? Obviously I'm not going to spell out  

73:09

the most sensitive secrets on a podcast. I need to have at least three more  

73:13

Guinnesses for that. What will xAI's business  

74:26

be? Is it going to be consumer, enterprise? What's the mix of those things going to be? 

74:31

Is it going to be similar to other labs— You’re saying "labs". Corporations. 

74:38

The psyop goes deep, Elon. "Revenue maximizing corporations", to be clear. 

74:43

Those GPUs don't pay for themselves. Exactly. What's the business model? What  

74:48

are the revenue streams in a few years’ time? Things are going to change very rapidly. I'm  

74:57

stating the obvious here. I call AI the  supersonic tsunami. I love alliteration.  

75:07

What's going to happen—especially when  you have humanoid robots at scale—is  

75:15

that they will make products and provide services  far more efficiently than human corporations. 

75:22

Amplifying the productivity of human  corporations is simply a short-term thing. 

75:27

So you're expecting fully digital corporations  rather than SpaceX becoming part AI? 

75:34

I think there will be digital  corporations but… Some of this  

75:41

is going to sound kind of doomerish, okay? But I'm just saying what I think will happen. 

75:46

It's not meant to be doomerish or anything else. This is just what I think will happen. 

75:58

Corporations that are purely AI and  robotics will vastly outperform any  

76:05

corporations that have people in the loop. Computer used to be a job that humans had. 

76:15

You would go and get a job as a computer  where you would do calculations. 

76:20

They'd have entire skyscrapers full of humans,  20-30 floors of humans, just doing calculations. 

76:29

Now, that entire skyscraper  of humans doing calculations  

76:35

can be replaced by a laptop with a spreadsheet. That spreadsheet can do vastly more calculations  

76:43

than an entire building full of human computers. You can think, "okay, what if only some of the  

76:52

cells in your spreadsheet  were calculated by humans?" 

76:59

Actually, that would be much worse  than if all of the cells in your  

77:02

spreadsheet were calculated by the computer. Really what will happen is that the pure AI,  

77:10

pure robotics corporations or collectives  will far outperform any corporations  

77:17

that have humans in the loop. And this will happen very quickly. 

77:21

Speaking of closing the loop… Optimus. As far as manufacturing targets go,  

77:31

your companies have been carrying American  manufacturing of hard tech on their back. 

77:39

But in the fields that Tesla has been dominant  in—and now you want to go into humanoids—in China  

77:47

there are dozens and dozens of companies that  are doing this kind of manufacturing cheaply  

77:53

and at scale that are incredibly competitive. So give us advice or a plan of how America can  

78:01

build the humanoid armies or the EVs, et cetera,  at scale and as cheaply as China is on track to. 

78:11

There are really only three  hard things for humanoid robots. 

78:15

The real-world intelligence, the  hand, and scale manufacturing. 

78:25

I haven't seen any, even demo  robots, that have a great hand,  

78:32

with all the degrees of freedom of a human hand.  Optimus will have that. Optimus does have that. 

78:41

How do you achieve that? Is it just  the right torque density in the motor? 

78:44

What is the hardware bottleneck to that? We had to design custom actuators,  

78:50

basically custom design motors, gears,  power electronics, controls, sensors. 

78:58

Everything had to be designed  from physics first principles. 

79:01

There is no supply chain for this. Will you be able to manufacture those at scale? 

79:06

Yes. Is anything hard, except  

79:07

the hand, from a manipulation point of view? Or once you've solved the hand, are you good? 

79:12

From an electromechanical standpoint, the hand  is more difficult than everything else combined. 

79:16

The human hand turns out to be quite something. But you also need the real-world intelligence. 

79:24

The intelligence that Tesla developed for  the car applies very well to the robot,  

79:32

which is primarily vision in. The car takes in vision,  

79:36

but it actually also is listening for sirens. It's taking in the inertial measurements,  

79:42

GPS signals, other data, combining  that with video, primarily video,  

79:47

and then outputting the control commands. Your Tesla is taking in one and a half  

79:55

gigabytes a second of video and outputting two  kilobytes a second of control outputs with the  

80:03

video at 36 hertz and the control frequency at 18. One intuition you could have for when we get this  

80:12

robotic stuff is that it takes quite a few years  to go from the compelling demo to actually being  

80:18

able to use it in the real world. 10 years ago,  you had really compelling demos of self-driving,  

80:23

but only now we have Robotaxis and  Waymo and all these services scaling up. 

80:29

Shouldn't this make one  pessimistic on household robots? 

80:33

Because we don't even quite have the compelling  demos yet of, say, the really advanced hand. 

80:39

Well, we've been working on  humanoid robots now for a while. 

80:44

I guess it's been five or six years or something. A bunch of the things that were done for the car  

80:52

are applicable to the robot. We'll use the same Tesla AI  

80:57

chips in the robot as in the car. We'll use the same basic principles. 

81:05

It's very much the same AI. You've got many more degrees of  

81:09

freedom for a robot than you do for a car. If you just think of it as a bitstream,  

81:16

AI is mostly compression and  correlation of two bitstreams. 

81:23

For video, you've got to do a  tremendous amount of compression  

81:28

and you've got to do the compression just right. You've got to ignore the things that don't matter. 

81:36

You don't care about the details of the  leaves on the tree on the side of the road,  

81:39

but you care a lot about the road signs  and the traffic lights, the pedestrians,  

81:45

and even whether someone in another car  is looking at you or not looking at you. 

81:51

Some of these details matter a lot. The car is going to turn that one and  

81:57

a half gigabytes a second ultimately into  two kilobytes a second of control outputs. 

82:02

So you’ve got many stages of compression. You've got to get all those stages right and then  

82:08

correlate those to the correct control outputs. The robot has to do essentially the same thing. 

82:14

This is what happens with humans. We really are photons in, controls out. 

82:19

That is the vast majority of your life: vision,  photons in, and then motor controls out. 

82:28

Naively, it seems that between humanoid  robots and cars… The fundamental actuators  

82:33

in a car are how you turn, how you accelerate. In a robot, especially with maneuverable arms,  

82:39

there's dozens and dozens  of these degrees of freedom. 

82:42

Then especially with Tesla, you had this advantage  of millions and millions of hours of human demo  

82:48

data collected from the car being out there. You can't equivalently deploy Optimuses that  

82:53

don't work and then get the data that way. So between the increased degrees of freedom  

82:57

and the far sparser data... Yes. That’s a good point. 

83:02

How will you use the Tesla engine of  intelligence to train the Optimus mind? 

83:11

You're actually highlighting an important  limitation and difference from cars. 

83:18

We'll soon have 10 million cars on the road. It's hard to duplicate that massive  

83:26

training flywheel. For the robot,  

83:32

what we're going to need to do is build a lot of  robots and put them in kind of an Optimus Academy  

83:37

so they can do self-play in reality. We're  actually building that out. We can have at  

83:45

least 10,000 Optimus robots, maybe 20-30,000, that  are doing self-play and testing different tasks. 

83:55

Tesla has quite a good reality  generator, a physics-accurate reality  

84:02

generator, that we made for the cars. We'll do the same thing for the robots. 

84:06

We actually have done that for the robots. So you have a few tens of thousands of  

84:14

humanoid robots doing different tasks. You can do millions of simulated  

84:20

robots in the simulated world. You use the tens of thousands of  

84:26

robots in the real world to close the simulation  to reality gap. Close the sim-to-real gap. 

84:32

How do you think about the synergies between xAI  and Optimus, given you're highlighting that you  

84:36

need this world model, you want to use some  really smart intelligence as a control plane,  

84:42

and Grok is doing the slower planning, and  then the motor policy is a little lower level. 

84:48

What will the synergy between these things be? Grok would orchestrate the  

84:55

behavior of the Optimus robots. Let's say you wanted to build a factory. 

85:05

Grok could organize the Optimus  robots, assign them tasks to build  

85:13

the factory to produce whatever you want. Don't you need to merge xAI and Tesla then? 

85:18

Because these things end up so... What were we saying earlier  

85:20

about public company discussions? We're one more Guinness in, Elon. 

85:28

What are you waiting to see before you say,  we want to manufacture 100,000 Optimuses? 

85:33

"Optimi". Since we're defining the  proper noun, we’re going to define  

85:38

the plural of the proper noun too. We're going to proper noun the  

85:42

plural and so it's Optimi. Is there something on the  

85:46

hardware side you want to see? Do you want to see better actuators? 

85:49

Is it just that you want  the software to be better? 

85:50

What are we waiting for before we  get mass manufacturing of Gen 3? 

85:54

No, we're moving towards that. We're  moving forward with the mass manufacturing. 

85:58

But you think current hardware is good enough that  you just want to deploy as many as possible now? 

86:06

It's very hard to scale up production. But I think Optimus 3 is the right version  

86:12

of the robot to produce something on  the order of a million units a year. 

86:20

I think you'd want to go to Optimus 4  before you went to 10 million units a year. 

86:23

Okay, but you can do a million units at Optimus 3? It's very hard to spool up manufacturing. 

86:34

The output per unit time  always follows an S-curve. 

86:38

It starts off agonizingly slow, then it has  this exponential increase, then a linear,  

86:44

then a logarithmic outcome until you  eventually asymptote at some number. 

86:50

Optimus’ initial production will be a  stretched out S-curve because so much  

86:57

of what goes into Optimus is brand new. There is not an existing supply chain. 

87:03

The actuators, electronics, everything  in the Optimus robot is designed  

87:08

from physics first principles. It's not taken from a catalog.  

87:11

These are custom-designed everything.  I don't think there's a single thing— 

87:17

How far down does that go? I guess we're not making custom  

87:22

capacitors yet, maybe. There's nothing you can  

87:29

pick out of a catalog, at any price. It just means that the Optimus S-Curve,  

87:39

the output per unit time, how many Optimus robots  you make per day, is going to initially ramp  

87:50

slower than a product where you  have an existing supply chain. 

87:55

But it will get to a million. When you see these Chinese humanoids,  

87:58

like Unitree or whatever, sell humanoids  for like $6K or $13K, are you hoping to  

88:05

get your Optimus bill of materials below  that price so you can do the same thing? 

88:10

Or do you just think qualitatively  they're not the same thing? 

88:15

What allows them to sell for  so low? Can we match that? 

88:19

Our Optimus is designed to have a lot  of intelligence and to have the same  

88:26

electromechanical dexterity, if not  higher, as a human. Unitree does not  

88:31

have that. It's also quite a big robot. It has to carry heavy objects for long  

88:41

periods of time and not overheat or  exceed the power of its actuators. 

88:50

It's 5'11", so it's pretty tall. It's got a lot of intelligence. 

88:57

So it's going to be more expensive than  a small robot that is not intelligent. 

89:02

But more capable. But not a lot more. The thing is,  

89:06

over time as Optimus robots build Optimus  robots, the cost will drop very quickly. 

89:12

What will these first billion  Optimuses, Optimi, do? 

89:17

What will their highest and best use be? I think you would start off with simple tasks  

89:21

that you can count on them doing well. But in the home or in factories? 

89:25

The best use for robots in the beginning  will be any continuous operation, any 24/7  

89:33

operation, because they can work continuously. What fraction of the work at a Gigafactory that  

89:39

is currently done by humans could a Gen 3 do? I'm not sure. Maybe it's 10-20%,  

89:46

maybe more, I don't know. We would not reduce our headcount. 

89:52

We would increase our headcount, to be clear. But we would increase our output. The units  

90:01

produced per human... The total number of humans  at Tesla will increase, but the output of robots  

90:09

and cars will increase disproportionately. The number of cars and robots produced per  

90:18

human will increase dramatically, but the  number of humans will increase as well. 

90:23

We're talking about Chinese  manufacturing a bunch here. 

90:30

We've also talked about some of  the policies that are relevant,  

90:33

like you mentioned, the solar tariffs. You think they're a bad idea because  

90:39

we can't scale up solar in the US. Electricity output in the US needs to scale up. 

90:45

It can't without good power sources. You just need to get it somehow. 

90:50

Where I was going with this is, if you  were in charge, if you were setting all  

90:53

the policies, what else would you change? You’d change the solar tariffs, that’s one. 

91:01

I would say anything that is a limiting  factor for electricity needs to be addressed,  

91:06

provided it's not very bad for the environment. So presumably some permitting reforms and stuff  

91:10

as well would be in there? There's a fair bit of  

91:12

permitting reforms that are happening. A lot of the permitting is state-based,  

91:17

but anything federal... This administration is good at  

91:21

removing permitting roadblocks. I'm not saying all tariffs are bad. 

91:28

Solar tariffs. Sometimes if another country is  

91:32

subsidizing the output of something, then you have  to have countervailing tariffs to protect domestic  

91:39

industry against subsidies by another country. What else would you change? 

91:43

I don't know if there's that much  that the government can actually do. 

91:46

One thing I was wondering... For the policy  goal of creating a lead for the US versus China,  

91:57

it seems like the export bans have  actually been quite impactful,  

92:02

where China is not producing leading-edge  chips and the export bans really bite there. 

92:07

China is not producing  leading-edge turbine engines. 

92:11

Similarly, there's a bunch of export bans that  are relevant there on some of the metallurgy. 

92:16

Should there be more export bans? As you think about things like the  

92:20

drone industry and things like that, is  that something that should be considered? 

92:24

It's important to appreciate that in most  areas, China is very advanced in manufacturing. 

92:30

There's only a few areas where it is not. China is a manufacturing powerhouse, next-level. 

92:40

It's very impressive. If you take refining of ore,  

92:49

China does roughly twice as much ore refining  on average as the rest of the world combined. 

93:00

There are some areas, like refining  gallium which goes into solar cells. 

93:04

I think they are 98% of gallium refining. So China is actually very advanced  

93:10

in manufacturing in most areas. It seems like there is discomfort  

93:16

with this supply chain dependence, and  yet nothing's really happening on it. 

93:20

Supply chain dependence? Say, like the gallium refining that  

93:24

you're saying. All the rare-earth stuff. Rare earths for sure,  

93:31

as you know, they’re not rare. We actually do rare earth ore mining in the US,  

93:37

send the rock, put it on a train, and then put  it on a boat to China that goes to another train,  

93:45

and goes to the rare earth refiners in China  who then refine it, put it into a magnet,  

93:51

put it into a motor sub-assembly,  and then send it back to America. 

93:54

So the thing we're really missing  is a lot of ore refining in America. 

94:00

Isn't this worth a policy intervention? Yes. I think there are some things  

94:06

being done on that front. But we kind of need Optimus,  

94:12

frankly, to build ore refineries. So, you think the main advantage  

94:17

China has is the abundance of skilled  labor? That's the thing Optimus fixes? 

94:24

Yes. China’s got like four times our population. I mean, there's this concern. If you think  

94:28

human resources are the future, right now  if it's the skilled labor for manufacturing  

94:34

that's determining who can build more  humanoids, China has more of those. 

94:39

It manufactures more humanoids, therefore  it gets the Optimi future first. 

94:44

Well, we’ll see. Maybe. It just keeps that exponential going. 

94:47

It seems like you're sort of pointing out  that getting to a million Optimi requires  

94:52

the manufacturing that the Optimi is  supposed to help us get to. Right? 

94:57

You can close that recursive loop pretty quickly. With a small number of Optimi? 

95:01

Yeah. So you close the recursive loop  to help the robots build the robots. 

95:08

Then we can try to get to tens of millions  of units a year. Maybe. If you start getting  

95:13

to hundreds of millions of units a year, you're  going to be the most competitive country by far. 

95:18

We definitely can't win with just humans,  because China has four times our population. 

95:23

Frankly, America has been winning for so  long that… A pro sports team that's been  

95:27

winning for a very long time tends  to get complacent and entitled. 

95:31

That's why they stop winning, because  they don't work as hard anymore. 

95:37

So frankly my observation is just that the average  work ethic in China is higher than in the US. 

95:44

It's not just that there's four  times the population, but the amount  

95:46

of work that people put in is higher. So you can try to rearrange the humans,  

95:52

but you're still one quarter of the—assuming  that productivity is the same, which I think  

96:00

actually it might not be, I think China might have  an advantage on productivity per person—we will do  

96:06

one quarter of the amount of things as China. So we can't win on the human front. 

96:12

Our birth rate has been low for a long time. The US birth rate's been below replacement  

96:20

since roughly 1971. We've got a lot of people retiring, we're close  

96:32

to more people domestically dying than being born. So we definitely can't win on the human front,  

96:38

but we might have a shot at the robot front. Are there other things that you have wanted to  

96:43

manufacture in the past, but they've been too  labor intensive or too expensive that now you  

96:48

can come back to and say, "oh, we can finally  do the whatever, because we have Optimus?" 

96:54

Yeah, we'd like to build  more ore refineries at Tesla. 

97:00

We just completed construction and have begun  lithium refining with our lithium refinery  

97:07

in Corpus Christi, Texas. We have a nickel refinery,  

97:12

which is for the cathode, that's here in Austin. This is the largest cathode refinery, largest  

97:24

nickel and lithium refinery, outside of China. The cathode team would say, "we have the  

97:35

largest and the only, actually,  cathode refinery in America." 

97:40

Not just the largest, but it's also the only. Many superlatives. 

97:43

So it was pretty big, even though it's  the only one. But there are other things.  

97:53

You could do a lot more refineries and help  America be more competitive on refining capacity. 

98:04

There's basically a lot of work for  the Optimus to do that most Americans,  

98:09

very few Americans, frankly want to do. Is the refining work too dirty or what's the— 

98:15

It's not actually, no. We don't have toxic  emissions from the refinery or anything. 

98:22

The cathode nickel refinery is in Travis County. Why can't you do it with humans? 

98:29

You can, you just run out of humans. Ah, I see. Okay. 

98:32

No matter what you do, you have one quarter  of the number of humans in America than China. 

98:36

So if you have them do this thing,  they can't do the other thing. 

98:38

So then how do you build this refining capacity? Well, you could do it with Optimi. 

98:48

Not very many Americans are pining to do refining. I mean, how many have you run into? Very few. Very  

99:00

few pining to refine. BYD is reaching Tesla  

99:04

production or sales in quantity. What do you think happens in global  

99:09

markets as Chinese production in EVs scales up? China is extremely competitive in manufacturing. 

99:19

So I think there's going to be a  massive flood of Chinese vehicles  

99:26

and basically most manufactured things. As it is, as I said, China is probably  

99:37

doing twice as much refining as  the rest of the world combined. 

99:40

So if you go down to fourth and  fifth-tier supply chain stuff… 

99:50

At the base level, you've got energy,  then you've got mining and refining. 

99:55

Those foundation layers are, like I said, as a  rough guess, China's doing twice as much refining  

100:03

as the rest of the world combined. So any given thing is going to have  

100:09

Chinese content because China's doing twice as  much refining work as the rest of the world. 

100:14

But they'll go all the way to the  finished product with the cars. 

100:22

I mean China is a powerhouse. I think this year China will exceed  

100:26

three times US electricity output. Electricity output is a reasonable  

100:32

proxy for the economy. In order to run the factories  

100:39

and run everything, you need electricity. It's a good proxy for the real economy. 

100:52

If China passes three times  the US electricity output,  

100:55

it means that its industrial capacity—as rough  approximation—will be three times that of the US. 

101:01

Reading between the lines, it sounds like what  you're saying is absent some sort of humanoid  

101:06

recursive miracle in the next few years, on the  whole manufacturing/energy/raw materials chain,  

101:16

China will just dominate whether it comes to AI  or manufacturing EVs or manufacturing humanoids. 

101:23

In the absence of breakthrough innovations  in the US, China will utterly dominate. 

101:35

Interesting. Yes. 

101:36

Robotics being the main breakthrough innovation. Well, to scale AI in space, basically you need  

101:49

humanoid robots, you need real-world AI,  you need a million tons a year to orbit. 

101:57

Let's just say if we get the mass driver on the  moon going, my favorite thing, then I think— 

102:03

We'll have solved all our problems. I call that winning. I call it winning, big time. 

102:13

You can finally be satisfied.  You've done something. 

102:16

Yes. You have the mass driver on the moon. 

102:18

I just want to see that thing in operation. Was that out of some sci-fi or where did you…? 

102:22

Well, actually, there is a Heinlein book. The Moon is a Harsh Mistress. 

102:26

Okay, yeah, but that's slightly different.  That's a gravity slingshot or... 

102:30

No, they have a mass driver on the Moon. Okay, yeah, but they use that to attack Earth. 

102:35

So maybe it's not the greatest... Well they use that to… assert their independence. 

102:38

Exactly. What are your plans  for the mass driver on the Moon? 

102:40

They asserted their independence. Earth  government disagreed and they lobbed  

102:44

things until Earth government agreed. That book is a hoot. I found that  

102:48

book much better than his other one that  everyone reads, Stranger in a Strange Land. 

102:51

"Grok" comes from Stranger in a Strange Land. The first two-thirds of Stranger in a Strange  

102:58

Land are good, and then it gets  very weird in the third portion. 

103:02

But there are still some good concepts in there. One thing we were discussing a lot  

104:18

is your system for managing people. You interviewed the first few thousand of  

104:24

SpaceX employees and lots of other companies. It obviously doesn't scale. 

104:29

Well, yes, but what doesn't scale? Me. 

104:32

Sure, sure. I know that. But  what are you looking for? 

104:36

There literally are not enough  hours in the day. It's impossible. 

104:38

But what are you looking for that  someone else who's good at interviewing  

104:42

and hiring people… What's the je ne sais quoi? At this point, I might have more training data  

104:51

on evaluating technical talent especially—talent  of all kinds I suppose, but technical talent  

104:56

especially—given that I've done so many  technical interviews and then seen the results. 

105:02

So my training set is enormous  and has a very wide range. 

105:11

Generally, the things I ask for are bullet  points for evidence of exceptional ability. 

105:21

These things can be pretty off the wall. It doesn't need to be in the specific domain,  

105:27

but evidence of exceptional ability. So if somebody can cite even one thing,  

105:34

but let's say three things, where you go,  "Wow, wow, wow," then that's a good sign. 

105:39

Why do you have to be the one to determine that? No, I don't. I can't be. It's impossible. The  

105:43

total headcount across all  companies is 200,000 people. 

105:48

But in the early days, what was  it that you were looking for that  

105:52

couldn't be delegated in those interviews? I guess I need to build my training set. 

106:02

It's not like I batted a thousand here. I would make mistakes, but then I'd be  

106:05

able to see where I thought somebody  would work out well, but they didn't. 

106:10

Then why did they not work out well? What can I do, I guess RL myself, to  

106:16

in the future have a better batting  average when interviewing people? 

106:22

My batting average is still not  perfect, but it's very high. 

106:24

What are some surprising  reasons people don't work out? 

106:27

Surprising reasons… Like, they don't understand  

106:30

technical domain, et cetera, et cetera. But you've got the long tail now of like,  

106:34

"I was really excited about this person. It  didn't work out." Curious why that happens. 

106:43

Generally what I tell people—I tell myself,  I guess, aspirationally—is, don't look at  

106:49

the resume. Just believe your interaction. The  resume may seem very impressive and it's like,  

106:55

"Wow, the resume looks good." But if the conversation  

107:00

after 20 minutes is not "wow," you should  believe the conversation, not the paper. 

107:07

I feel like part of your method is that… There  was this meme in the media a few years back about  

107:14

Tesla being a revolving door of executive talent. Whereas actually, I think when you look at it,  

107:19

Tesla's had a very consistent and internally  promoted executive bench over the past few years. 

107:24

Then at SpaceX, you have all these  folks like Mark Juncosa and Steve Davis— 

107:29

Steve Davis runs The Boring Company these days. Bill Riley, and folks like that. 

107:35

It feels like part of what has worked well  is having very capable technical deputies. 

107:43

What do all of those people have in common? Well, the Tesla senior team,  

107:53

at this point has probably got an average  tenure of 10-12 years. It's quite long.  

108:03

But there were times when Tesla went  through an extremely rapid growth phase,  

108:11

so things were just somewhat sped up. As you know, a company goes through  

108:17

different orders of magnitude of size. People that could help manage, say,  

108:23

a 50-person company versus a 500-person  company versus a 5,000-person company versus  

108:28

a 50,000-person company. You outgrew people. 

108:31

It's just not the same team. It's not always the same team. 

108:34

So if a company is growing very rapidly,  the rate at which executive positions will  

108:39

change will also be proportionate to  the rapidity of the growth generally. 

108:47

Tesla had a further challenge where when Tesla had  very successful periods, we would be relentlessly  

108:56

recruited from. Like, relentlessly. When  Apple had their electric car program,  

109:03

they were carpet bombing Tesla with recruiting  calls. Engineers just unplugged their phones. 

109:10

"I'm trying to get work done here." Yeah. "If I get one more call from  

109:14

an Apple recruiter…" But their opening  offer without any interview would be  

109:19

like double the compensation at Tesla. So we had a bit of the "Tesla pixie  

109:28

dust" thing where it's like, "Oh,  if you hire a Tesla executive,  

109:32

suddenly everything's going to be successful." I've fallen prey to the pixie dust thing as well,  

109:38

where it's like, "Oh, we'll hire someone from  Google or Apple and they'll be immediately  

109:41

successful," but that's not how it works.  People are people. There's no magical pixie  

109:47

dust. So when we had the pixie dust problem,  we would get relentlessly recruited from. 

109:57

Also, Tesla being engineering, especially  being primarily in Silicon Valley,  

110:03

it's easier for people to just... They don't have to change their life very much. 

110:10

Their commute's going to be the same. So how do you prevent that? 

110:14

How do you prevent the pixie dust effect where  everyone's trying to poach all your people? 

110:21

I don't think there's much we can do to stop it. That's one of the reasons why Tesla… Really,  

110:29

being in Silicon Valley and having the pixie  dust thing at the same time meant that there was  

110:39

just a very, very aggressive recruitment. Presumably being in Austin helps then? 

110:44

Austin, it helps. Tesla still has a  majority of its engineering in California. 

110:56

Getting engineers to move… I call  it the "significant other" problem. 

111:00

Yes, "significant others" have jobs. Exactly. So for Starbase that was  

111:06

particularly difficult, since the  odds of finding a non-SpaceX job… 

111:10

In Brownsville, Texas… …are pretty low. It's  

111:13

quite difficult. It's like a technology  monastery thing, remote and mostly dudes. 

111:22

Not much of an improvement over SF. If you go back to these people who've really  

111:34

been very effective in a technical capacity at  Tesla, at SpaceX, and those sorts of places, what  

111:41

do you think they have in common other than... Is it just that they're very sharp on the  

111:48

rocketry or the technical foundations, or  do you think it's something organizational? 

111:52

Is it something about their  ability to work with you? 

111:54

Is it their ability to be  flexible but not too flexible? 

112:03

What makes a good sparring partner for you? I don't think of it as a sparring partner. 

112:08

If somebody gets things done,  I love them, and if they don't,  

112:11

I hate them. So it's pretty straightforward.  It's not like some idiosyncratic thing. 

112:17

If somebody executes well, I'm a  huge fan, and if they don't, I'm not. 

112:22

But it's not about mapping to  my idiosyncratic preferences. 

112:25

I certainly try not to have it be  mapping to my idiosyncratic preferences. 

112:36

Generally, I think it's a good idea to hire  for talent and drive and trustworthiness. 

112:46

And I think goodness of heart is important. I underweighted that at one point. 

112:53

So, are they a good person? Trustworthy?  Smart and talented and hard working? 

113:01

If so, you can add domain knowledge. But those fundamental traits,  

113:06

those fundamental properties, you cannot change. So most of the people who are at Tesla and SpaceX  

113:14

did not come from the aerospace  industry or the auto industry. 

113:18

What has had to change most about your  management style as your companies have  

113:21

scaled from 100 to 1,000 to 10,000 people? You're known for this very micro management,  

113:27

just getting into the details of things. Nano management, please. Pico management.  

113:34

Femto management. Keep going. 

113:39

We're going to go all the way  down to Planck's constant. 

113:44

All the way down to Heisenberg  uncertainty principle. 

113:50

Are you still able to get into  details as much as you want? 

113:52

Would your companies be more  successful if they were smaller? 

113:56

How do you think about that? Because I have a fixed amount of  

113:58

time in the day, my time is necessarily diluted as  things grow and as the span of activity increases. 

114:10

It's impossible for me to actually be a  micromanager because that would imply I  

114:17

have some thousands of hours per day. It is a logical impossibility  

114:22

for me to micromanage things. Now, there are times when I will drill down into a  

114:31

specific issue because that specific issue is the  limiting factor on the progress of the company. 

114:42

The reason for drilling into some very detailed  item is because it is the limiting factor. 

114:49

It’s not arbitrarily drilling into tiny things. From a time standpoint, it is physically  

114:57

impossible for me to arbitrarily go into  tiny things that don't matter. That would  

115:03

result in failure. But sometimes the  tiny things are decisive in victory. 

115:09

Famously, you switched the Starship  design from composites to steel. 

115:17

Yes. You made  

115:18

that decision. That wasn't people going around  saying, "Oh, we found something better, boss." 

115:22

That was you encouraging  people against some resistance. 

115:25

Can you tell us how you came to that  whole concept of the steel switch? 

115:32

Desperation, I'd say. Originally, we were  going to make Starship out of carbon fiber.  

115:45

Carbon fiber is pretty expensive. When you do  volume production, you can get any given thing  

115:55

to start to approach its material cost. The problem with carbon fiber is that  

116:00

material cost is still very high. Particularly if you go for a high-strength  

116:10

specialized carbon fiber that can handle cryogenic  oxygen, it's roughly 50 times the cost of steel. 

116:20

At least in theory, it would be lighter. People generally think of steel as being  

116:24

heavy and carbon fiber as being light. For room temperature applications,  

116:35

like a Formula 1 car, static aero structure,  or any kind of aero structure really, you're  

116:42

probably going to be better off with carbon fiber. The problem is that we were trying to make this  

116:48

enormous rocket out of carbon fiber  and our progress was extremely slow. 

116:53

It had been picked in the first  place just because it's light? 

116:57

Yes. At first glance, most people  would think that the choice for  

117:04

making something light would be carbon fiber. The thing is that when you make something very  

117:18

enormous out of carbon fiber and then you try  to have the carbon fiber be efficiently cured,  

117:25

meaning not room temperature cured, because  sometimes you got 50 plies of carbon fiber…  

117:33

Carbon fiber is really carbon string and glue. In order to have high strength,  

117:39

you need an autoclave. Something that's essentially a high pressure oven. 

117:46

If you have something that's gigantic, that  one's got to be bigger than the rocket. 

117:52

We were trying to make an autoclave that's  bigger than any autoclave that's ever existed. 

117:58

Or you can do room temperature cure,  which takes a long time and has issues. 

118:03

The final issue is that we were just making  very slow progress with carbon fiber. 

118:12

The meta question is why it had  to be you who made that decision. 

118:18

There's many engineers on your team. How did the team not arrive at steel? 

118:20

Yeah exactly. This is part of a broader  question, understanding your comparative  

118:24

advantage at your companies. Because we were making very slow  

118:29

progress with carbon fiber, I was like,  "Okay, we've got to try something else." 

118:33

For the Falcon 9, the primary airframe  is made of aluminum lithium, which has  

118:41

a very good strength-to-weight. Actually, it has about the same,  

118:47

maybe better, strength to weight for  its application than carbon fiber. 

118:51

But aluminum lithium is  very difficult to work with. 

118:53

In order to weld it, you have to do something  called friction stir welding, where you join the  

118:57

metal without entering the liquid phase. It's kind of wild that you can do that. 

119:02

But with this particular type of welding, you  can do that. It's very difficult. Let's say you  

119:10

want to make a modification or attach something to  aluminum lithium, you now have to use a mechanical  

119:16

attachment with seals. You can't weld it on.  So I wanted to avoid using aluminum lithium  

119:24

for the primary structure for Starship. There was this very special grade of  

119:34

carbon fiber that had very good mass properties. With a rocket, you're really trying to maximize  

119:41

the percentage of the rocket that is  propellant, minimize the mass obviously. 

119:48

But like I said, we were  making very slow progress. 

119:54

I said, "at this rate, we’re  never going to get to Mars. 

119:56

So we've got to think of something else." I didn't want to use aluminum lithium  

120:01

because of the difficulty of friction stir  welding, especially doing that at scale. 

120:06

It was hard enough at 3.6 meters in  diameter, let alone at 9 meters or above. 

120:12

Then I said, "what about steel?" I had a clue here because some of  

120:21

the early US rockets had used very thin steel. The Atlas rockets had used a steel balloon tank. 

120:30

It's not like steel had never been used before.  It actually had been used. When you look at  

120:35

the material properties of stainless steel,  full-hard, strain hardened stainless steel,  

120:46

at cryogenic temperature the strength to  weight is actually similar to carbon fiber. 

120:54

If you look at material properties  at room temperature, it looks like  

120:58

the steel is going to be twice as heavy. But if you look at the material properties  

121:03

at cryogenic temperature of full-hard  steel, stainless of particular grades,  

121:10

then you actually get to a similar  strength to weight as carbon fiber. 

121:15

In the case of Starship, both the  fuel and the oxidizer are cryogenic. 

121:19

For Falcon 9, the fuel is rocket propellant-grade  kerosene, basically a very pure form of jet fuel.  

121:32

That is roughly room temperature. Although  we do actually chill it slightly below,  

121:38

we chill it like a beer. Delicious. 

121:41

We do chill it, but it's not cryogenic. In fact, if we made it cryogenic,  

121:45

it would just turn to wax. But for Starship,  

121:50

it's liquid methane and liquid oxygen. They are liquid at similar temperatures. 

121:59

Basically, almost the entire primary  structure is at cryogenic temperature. 

122:03

So then you've got a 300-series  stainless that's strain hardened. 

122:12

Because almost all things are cryogenic  temperature, it actually has similar  

122:17

strength to weight as carbon fiber. But it costs 50x less in raw  

122:25

material and is very easy to work with. You can weld stainless steel outdoors. 

122:30

You could smoke a cigar while welding  stainless steel. It's very resilient.  

122:37

You can modify it easily. If you want to  attach something, you just weld it right on. 

122:44

Very easy to work with, very low cost. Like I said, at cryogenic temperature,  

122:52

it’s similar strength-to-weight to carbon fiber. Then when you factor in that we have a much  

123:02

reduced heat shield mass, because the  melting point of steel, is much greater  

123:07

than the melting point of aluminum… It's  about twice the melting point of aluminum. 

123:13

So you can just run the rocket much hotter? Yes, especially for the ship which is coming  

123:18

in like a blazing meteor. You can greatly reduce  

123:25

the mass of the heat shield. You can cut the mass of the windward  

123:34

part of the heat shield, maybe in half, and you  don't need any heat shielding on the leeward side. 

123:45

The net result is that actually  the steel rocket weighs less than  

123:49

the carbon fiber rocket, because the resin  in the carbon fiber rocket starts to melt. 

124:00

Basically, carbon fiber and aluminum have about  the same operating temperature capabilities,  

124:06

whereas steel can operate at twice the  temperature. These are very rough approximations. 

124:12

I won't build the rocket. What I mean is people will say,  

124:14

"Oh, he said this twice. It's actually  0.8." I'm like, shut up, assholes. 

124:18

That's what the main comment's going to be about. God damn it. The point is, in retrospect, we  

124:25

should have started with steel in the beginning. It was dumb not to do steel. 

124:28

Okay, but to play this back to you, what  I'm hearing is that steel was a riskier,  

124:32

less proven path, other than the early US rockets. Versus carbon fiber was a worse but  

124:40

more proven out path. So you need to be the  

124:43

one to push for, "Hey, we're going to do  this riskier path and just figure it out." 

124:48

So you're fighting a sort  of conservatism in a sense. 

124:52

That's why I initially said that the issue is  that we weren't making fast enough progress. 

124:57

We were having trouble making even a  small barrel section of the carbon fiber  

125:02

that didn't have wrinkles in it. Because at that large scale, you have to  

125:08

have many plies, many layers of the carbon fiber. You've got to cure it and you've got to cure it  

125:14

in such a way that it doesn't  have any wrinkles or defects. 

125:18

Carbon fiber is much less resilient  than steel. It has much less toughness.  

125:26

Stainless steel will stretch and bend,  the carbon fiber will tend to shatter. 

125:35

Toughness being the area  under the stress strain curve. 

125:39

You're generally going to have to do better  with steel, but stainless steel to be precise. 

125:44

One other Starship question. So I visited  Starbase, I think it was two years ago,  

125:51

with Sam Teller, and that was awesome. It was very cool to see, in a whole bunch of ways. 

125:55

One thing I noticed was that people really took  pride in the simplicity of things, where everyone  

126:02

wants to tell you how Starship is just a big soda  can, and we're hiring welders, and if you can weld  

126:09

in any industrial project, you can weld here. But there's a lot of pride in the simplicity. 

126:16

Well, factually Starship is  a very complicated rocket. 

126:18

So that's what I'm getting at. Are things simple or are they complex? 

126:23

I think maybe just what they're trying to say  is that you don't have to have prior experience  

126:27

in the rocket industry to work on Starship. Somebody just needs to be smart and work hard  

126:36

and be trustworthy and they can work on a rocket. They don't need prior rocket experience. 

126:42

Starship is the most complicated machine  ever made by humans, by a long shot. 

126:47

In what regards? Anything, really. I'd  

126:50

say there isn't a more complex machine. I'd say that pretty much any project I  

127:00

can think of would be easier than this. That's why nobody has ever made a fully  

127:08

reusable orbital rocket. It's a very hard problem.  Many smart people have tried before, very smart  

127:18

people with immense resources, and they failed.  And we haven't succeeded yet. Falcon is partially  

127:26

reusable, but the upper stage is not. Starship Version 3,  

127:32

I think this design can be fully reusable. That full reusability is what will enable  

127:41

us to become a multi-planet civilization. Any technical problem, even like a Hadron  

127:52

Collider or something like that,  is an easier problem than this. 

127:55

We spent a lot of time on bottlenecks. Can you say what the current Starship  

127:58

bottlenecks are, even at a high level? Trying to make it not explode, generally.  

128:04

It really wants to explode. That old chestnut. All those  

128:09

combustible materials. We've had two boosters explode on the test stand. 

128:13

One obliterated the entire test facility. So it only takes that one mistake. 

128:21

The amount of energy contained  in a Starship is insane. 

128:25

Is that why it's harder than Falcon? It's because it's just more energy? 

128:30

It's a lot of new technology. It's  pushing the performance envelope. The  

128:37

Raptor 3 engine is a very, very advanced engine. It's by far the best rocket engine ever made. 

128:43

But it desperately wants to blow up. Just to put things into perspective here,  

128:48

on liftoff the rocket is generating over 100  gigawatts of power. That’s 20% of US electricity. 

128:58

It's actually insane. It's a great comparison. 

128:59

While not exploding. Sometimes. 

129:02

Sometimes, yes. So I was  like, how does it not explode? 

129:06

There's thousands of ways that it could  explode and only one way that it doesn't. 

129:12

So we want it not only to really not explode, but  fly reliably on a daily basis, like once per hour. 

129:22

Obviously, if it blows up a lot,  it's very difficult to maintain that  

129:25

launch cadence. Yes. 

129:30

What's the single biggest  remaining problem for Starship? 

129:33

It's having the heat shield be reusable. No one's ever made a reusable orbital heat shield. 

129:44

So the heat shield's gotta make it through the  ascent phase without shucking a bunch of tiles,  

129:52

and then it's gotta come back in and also not lose  a bunch of tiles or overheat the main airframe. 

130:01

Isn't that hard because it's  fundamentally a consumable? 

130:05

Well, yes, but your brake pads in your car are  also consumable, but they last a very long time. 

130:09

Fair. So it just needs to last a very long time. 

130:17

We have brought the ship back and had  it do a soft landing in the ocean. 

130:22

We've done that a few times. But it lost a lot of tiles. 

130:27

It was not reusable without a lot of work. Even though it did come to a soft landing,  

130:35

it would not have been  reusable without a lot of work. 

130:40

So it's not really reusable in that sense. That's the biggest problem that remains,  

130:44

a fully reusable heat shield. You want to be able to land it,  

130:50

refill propellant and fly again. You can't do this laborious inspection  

130:57

of 40,000 tiles type of thing. When I read biographies of yours,  

131:06

it seems like you're just able to drive the sense  of urgency and drive the sense of "this is the  

131:11

thing that can scale." I'm curious why you  

131:15

think other organizations of your… SpaceX and Tesla are really big companies now. 

131:20

You're still able to keep that culture. What goes wrong with other companies such  

131:24

that they're not able to do that? I don't know. 

131:29

Like today, you said you had  a bunch of SpaceX meetings. 

131:31

What is it that you're doing  there that's keeping that? 

131:33

It’s adding urgency? Well, I don't know. I guess the urgency is going  

131:42

to come from whoever is leading the company. I have a maniacal sense of urgency. 

131:47

So that maniacal sense of urgency  projects through the rest of the company. 

131:52

Is it because of consequences? They're like,  "Elon set a crazy deadline, but if I don't get it,  

131:57

I know what happens to me." Is it just that you're able to  

132:01

identify bottlenecks and get rid  of them so people can move fast? 

132:03

How do you think about why your  companies are able to move fast? 

132:07

I'm constantly addressing the limiting factor. On the deadlines front, I generally actually  

132:20

try to aim for a deadline that I at  least think is at the 50th percentile. 

132:25

So it's not like an impossible deadline, but  it's the most aggressive deadline I can think  

132:29

of that could be achieved with 50% probability. Which means that it'll be late half the time. 

132:42

There is a law of gas expansion  that applies to schedules. 

132:48

If you said we're going to do something in  five years, which to me is like infinity time,  

132:55

it will expand to fill the available  schedule and it'll take five years. 

133:05

Physics will limit how fast  you can do certain things. 

133:07

So scaling up manufacturing, there's  a rate at which you can move the atoms  

133:15

and scale manufacturing. That's why you can't instantly  

133:17

make a million units a year of something. You've got to design the manufacturing line. 

133:23

You've got to bring it up. You've got to ride the S-curve of production. 

133:31

What can I say that's actually helpful to people? 

133:38

Generally, a maniacal sense  of urgency is a very big deal. 

133:47

You want to have an aggressive schedule and  you want to figure out what the limiting  

133:54

factor is at any point in time and help  the team address that limiting factor. 

133:59

So Starlink was slowly in  the works for many years. 

134:05

We talked about it all the way  in the beginning of the company. 

134:07

So then there was a team you had built  in Redmond, and then at one point you  

134:12

decided this team is just not cutting it. It went for a few years slowly, and so why didn't  

134:25

you act earlier, and why did you act when you did? Why was that the right moment at which to act? 

134:30

I have these very detailed  engineering reviews weekly. 

134:38

That's maybe a very unusual level of granularity. I don't know anyone who runs a company,  

134:45

or at least a manufacturing company, that  goes with the level of detail that I go  

134:50

into. It's not as though... I have a pretty  good understanding of what's actually going  

134:57

on because we go through things in detail. I'm a big believer in skip-level meetings  

135:06

where instead of having the person that reports to  me say things, it's everyone that reports to them  

135:14

saying something in the technical review. And there can't be advanced preparation. 

135:25

Otherwise you're going to get  "glazed", as I say these days. 

135:31

Exactly. Very Gen Z of you. How do you prevent advanced preparation? 

135:34

Do you call on them randomly? No, I just go around the room.  

135:37

Everyone provides an update. It's a lot  of information to keep in your head. 

135:48

If you have meetings weekly or twice weekly,  you've got a snapshot of what that person said. 

135:56

You can then plot the progress points. You can sort of mentally plot the  

136:03

points on a curve and say, "are we  converging to a solution or not?" 

136:12

I'll take drastic action only when I conclude  that success is not in a set of possible outcomes. 

136:22

So when I finally reach the conclusion that unless  drastic action is done, we have no chance of  

136:29

success, then I must take drastic action. I came to that conclusion in 2018,  

136:36

took drastic action and fixed the problem. You've got many, many companies. In each of  

136:45

them it sounds like you do this kind  of deep engineering understanding of  

136:49

what the relevant bottlenecks are so  you can do these reviews with people. 

136:56

You've been able to scale it up  to five, six, seven companies. 

136:59

Within one of these companies, you have  many different mini companies within them. 

137:04

What determines the max amount here? Because you have like 80 companies…? 

137:07

80? No. But you have so many  

137:10

already. That's already remarkable. By this current number. 

137:13

Exactly. We can barely keep one company together. 

137:23

It depends on the situation. I actually don't  have regular meetings with The Boring Company,  

137:32

so The Boring Company is sort of cruising along. Basically, if something is working well and  

137:37

making good progress, then there's  no point in me spending time on it. 

137:42

I actually allocate time according to where the  limiting factor. Where are things problematic?  

137:50

Where are we pushing against? What is holding  us back? I focus, at the risk of saying the  

137:59

words too many times, on the limiting factor. The irony is if something's going really well,  

138:09

they don't see much of me. But if something is going badly,  

138:12

they'll see a lot of me. Or not even badly… If something is the limiting factor. 

138:18

The limiting factor, exactly. It’s  not exactly going badly but it’s the  

138:21

thing that we need to make go faster. When something’s a limiting factor at  

138:25

SpaceX or Tesla, are you talking weekly  and daily with the engineer that's  

138:32

working on it? How does that actually work? Most things that are the limiting factor are  

138:39

weekly and some things are twice weekly. The AI5 chip review is twice weekly. 

138:46

Every Tuesday and Saturday is the chip review. Is it open ended in how long it goes? 

138:54

Technically, yes, but usually it's two or  three hours. Sometimes less. It depends on  

139:03

how much information we've got to go through. That's another thing. I'm just trying to tease  

139:07

out the differences here because  the outcomes seem quite different. 

139:11

I think it's interesting to  know what inputs are different. 

139:14

It feels like in the corporate world, one,  like you were saying, the CEO doing engineering  

139:20

reviews does not always happen despite the  fact that that is what the company is doing. 

139:25

But then time is often pretty finely sliced into  half hour meetings or even 15 minute meetings. 

139:32

It seems like you hold more open-ended,  "We're talking about it until we figure  

139:38

it out" type things. Sometimes. But most  

139:43

of them seem to more or less stay on time. Today's Starship engineering review went a bit  

139:56

longer because there were more topics to discuss. They're trying to figure out how to scale to a  

140:04

million plus tons to orbit per  year. It’s quite challenging. 

140:08

Can I ask a question? You said about Optimus  and AI that they're going to result in double  

140:15

digit growth rates within a matter of years. Oh, like the economy? Yes. I think that's right. 

140:22

What was the point of the DOGE cuts if  the economy is going to grow so much? 

140:28

Well, I think waste and fraud  are not good things to have. 

140:33

I was actually pretty worried about... In the absence of AI and robotics,  

140:41

we're actually totally screwed because  the national debt is piling up like crazy. 

140:50

The interest payments to national debt exceed  the military budget, which is a trillion dollars. 

140:54

So we have over a trillion  dollars just in interest payments. 

141:00

I was pretty concerned about that. Maybe if I spend some time, we can  

141:03

slow down the bankruptcy of the United States  and give us enough time for the AI and robots  

141:09

to help solve the national debt. Or not help solve, it's the only  

141:16

thing that could solve the national debt. We are 1000% going to go bankrupt as a country,  

141:21

and fail as a country, without AI and robots. Nothing else will solve the national debt. 

141:30

We just need enough time to build the AI  and robots to not go bankrupt before then. 

141:39

I guess the thing I'm curious about is,  when DOGE starts you have this enormous  

141:43

ability to enact reform. Not that enormous. 

141:48

Sure. I totally buy your point that it's  important that AI and robotics drive  

141:53

productivity improvements, drive GDP growth. But why not just directly go after the things  

141:58

you were pointing out, like the tariffs  on certain components, or permitting? 

142:03

I'm not the president. And it is very hard to  cut things that are obvious waste and fraud,  

142:13

like ridiculous waste and fraud. What I discovered is that it's extremely  

142:21

difficult even to cut very obvious waste and  fraud from the government because the government  

142:28

has to operate on who's complaining. If you cut off payments to fraudsters,  

142:34

they immediately come up with the most sympathetic  sounding reasons to continue the payment. 

142:39

They don't say, "Please keep the fraud going." They’re like, "You're killing baby pandas." 

142:46

Meanwhile, no baby pandas are dying. They're  just making it up. The fraudsters are capable  

142:51

of coming up with extremely compelling,  heart-wrenching stories that are false,  

142:56

but nonetheless sound sympathetic. That's what  happened. Perhaps I should have known better. 

143:10

But I thought, wait, let's try to cut some  amount of waste and pork from the government. 

143:16

Maybe there shouldn't be 20 million people  marked as alive in Social Security who are  

143:22

definitely dead, and over the age of 115. The  oldest American is 114. So it's safe to say if  

143:30

somebody is 115 and marked as alive in the Social  Security database, there's either a typo… Somebody  

143:39

should call them and say, "We seem to have  your birthday wrong, or we need to mark you  

143:47

as dead." One of the two things. Very intimidating call to get. 

143:52

Well, it seems like a reasonable thing. Say if their birthday is in the future  

143:59

and they have a Small Business Administration  loan, and their birthday is 2165,  

144:07

we either have a typo or we have fraud. So we say, "we appear to have gotten the  

144:13

century of your birth incorrect." Or a great plot for a movie. 

144:17

Yes. That's what I mean by, ludicrous fraud. Were those people getting payments? 

144:23

Some were getting payments from Social Security. But the main fraud vector was to mark somebody as  

144:29

alive in Social Security and then use every other  government payment system to basically do fraud. 

144:37

Because what those other  government payment systems do,  

144:40

they would simply do an "are you alive" check to  the Social Security database. It's a bank shot. 

144:46

What would you estimate is the total  amount of fraud from this mechanism? 

144:52

By the way, the Government Accountability  Office has done these estimates before. I'm  

144:55

not the only one. In fact, I think the GAO did  an analysis, a rough estimate of fraud during  

145:02

the Biden administration, and calculated  it at roughly half a trillion dollars. 

145:08

So don't take my word for it. Take a report issued during the  

145:11

Biden administration. How about that? From this Social Security mechanism? 

145:16

It's one of many. It's important to  appreciate that the government is  

145:22

very ineffective at stopping fraud. It's not like a company where, with  

145:30

stopping fraud, you've got a motivation because  it's affecting the earnings of your company. 

145:34

The government just prints more money.  You need caring and competence. These are  

145:44

in short supply at the federal level. When you go to the DMV, do you think,  

145:52

"Wow, this is a bastion of competence"? Well, now imagine it's worse than the DMV  

145:57

because it's the DMV that can print money. At least the state level DMVs need to... 

146:05

The states more or less need to stay  within their budget or they go bankrupt. 

146:08

But the federal government just prints more money. If there's actually half a trillion of fraud,  

146:14

why was it not possible to cut all that? You really have to stand back and recalibrate  

146:28

your expectations for competence. Because you're operating in a world  

146:36

where you've got to make ends meet. You've got to pay your bills... 

146:40

Find the microphones. Exactly. It's not like there's a giant,  

146:49

largely uncaring monster bureaucracy. It's a bunch of anachronistic computers  

146:57

that are just sending payments. One of the things that the DOGE  

147:03

team did sounds so simple and probably  will save $100-200 billion a year. 

147:14

It was simply requiring payments from the  main Treasury computer—which is called PAM,  

147:19

Payment Accounts Master or something like  that, there's $5 trillion payments a year—that  

147:25

go out have a payment appropriation code. Make it mandatory, not optional, that you  

147:32

have anything at all in the comment field. You have to recalibrate how dumb things are. 

147:42

Payments were being sent out with no appropriation  code, not checking back to any congressional  

147:48

appropriation, and with no explanation. This is why the Department of War,  

147:54

formerly the Department of Defense, cannot pass  an audit, because the information is literally  

147:59

not there. Recalibrate your expectations. I want to better understand this half a trillion  

148:04

number, because there's an IG report in 2024. Why is it so low? 

148:10

Maybe, but we found that over seven  years, the Social Security fraud  

148:14

they estimated was like $70 billion over  seven years, so like $10 billion a year. 

148:17

So I'd be curious to see what  the other $490 billion is. 

148:20

Federal government expenditures  are $7.5 trillion a year. 

148:26

How competent do you think the government is? The discretionary spending there is like… 15%? 

148:33

But it doesn't matter. Most of  the fraud is non-discretionary. 

148:36

It's basically fraudulent Medicare,  Medicaid, Social Security,  

148:45

disability. There's a zillion government  payments. A bunch of these payments are in  

148:52

fact block transfers to the states. So the federal government doesn't  

148:59

even have the information in a lot of  cases to even know if there's fraud.  

149:04

Let's consider reductio ad absurdum. The  government is perfect and has no fraud. 

149:10

What is your probability estimate of that? Zero.  Okay, so then would you say, fraud and waste  

149:18

at the government is 90% efficient? That also would be quite generous. 

149:27

But if it's only 90%, that means that  there's $750 billion a year of waste and  

149:32

fraud. And it's not 90%. It's not 90% effective. This seems like a strange way to first principles  

149:38

the amount of fraud in the government. Just like, how much do you think there is? 

149:43

Anyways, we don't have to do  it live, but I'd be curious— 

149:45

You know a lot about fraud at Stripe? People are constantly trying to do fraud. 

149:49

Yeah, but as you say, it's a little bit of a... We've really ground it down, but it's a little  

149:54

bit of a different problem space because you're  dealing with a much more heterogeneous set of  

149:58

fraud vectors here than we are. But at Stripe, you have high  

150:03

competence and you try hard. You have high competence and  

150:07

high caring, but still fraud is non-zero. Now imagine it's at a much bigger scale, there's  

150:15

much less competence, and much less caring. At PayPal back in the day, we tried to manage  

150:22

fraud down to about 1% of the payment volume. That  was very difficult. It took a tremendous amount of  

150:28

competence and caring to get fraud merely to 1%. Now imagine that you're an organization where  

150:36

there's much less caring and much less competence. It's going to be much more than 1%. 

150:41

How do you feel now looking back  on politics and doing stuff there? 

150:48

Looking from the outside in, two things have been  quite impactful: one, the America PAC, and two,  

150:59

the acquisition of Twitter at the time. But also it seems like there  

151:05

was a bunch of heartache. What's your grading of the whole experience? 

151:16

I think those things needed to be done to  maximize the probability that the future is good.  

151:27

Politics generally is very tribal. People  lose their objectivity usually with politics. 

151:35

They generally have trouble seeing the good on  the other side or the bad on their own side.  

151:41

That's generally how it goes. That, I guess, was  one of the things that surprised me the most. 

151:48

You often simply cannot reason with people. If they're in one tribe or the other. 

151:52

They simply believe that everything  their tribe does is good and anything  

151:55

the other political tribe does is bad. Persuading them otherwise is almost impossible. 

152:07

But I think overall those actions—acquiring  Twitter, getting Trump elected, even though  

152:22

it makes a lot of people angry—I think  those actions were good for civilization. 

152:30

How does it feed into the  future you're excited about? 

152:33

Well, America needs to be strong enough to  last long enough to extend life to other  

152:42

planets and to get AI and robotics to the point  where we can ensure that the future is good. 

152:51

On the other hand, if we were to descend into,  say, communism or some situation where the state  

152:59

was extremely oppressive, that would mean that  we might not be able to become multi-planetary. 

153:10

The state might stamp out our  progress in AI and robotics. 

153:21

Optimus, Grok, et cetera. Not just yours, but  any revenue-maximizing company's products will  

153:29

be leveraged by the government over time. How does this concern manifest in what  

153:37

private companies should be willing to give  governments? What kinds of guardrails? Should  

153:44

AI models be made to do whatever  the government that has contracted  

153:51

them out to do and asks them to do? Should Grok get to say, "Actually,  

153:57

even if the military wants to do  X, no, Grok will not do that"? 

154:01

I think maybe the biggest danger of AI  and robotics going wrong is government. 

154:16

People who are opposed to corporations  or worried about corporations should  

154:21

really worry the most about government. Because government is just a  

154:25

corporation in the limit. Government is just the biggest  

154:30

corporation with a monopoly on violence. I always find it a strange dichotomy where  

154:38

people would think corporations are bad, but  the government is good, when the government is  

154:41

simply the biggest and worst corporation. But  people have that dichotomy. They somehow think  

154:50

at the same time that government can be good,  but corporations bad, and this is not true. 

154:55

Corporations have better  morality than the government. 

154:59

I actually think it’s a thing to be worried about. The government could potentially use AI and  

155:12

robotics to suppress the population.  That is a serious concern. 

155:18

As the guy building AI and  robotics, how do you prevent that? 

155:28

If you limit the powers of government, which is  really what the US Constitution is intended to do,  

155:33

to limit the powers of government, then you're  probably going to have a better outcome than  

155:37

if you have more government. Robotics will be available  

155:42

to all governments, right? I don’t know about all governments.  

155:49

It's difficult to predict. I can say what's the  endpoint, or what is many years in the future, but  

155:57

it's difficult to predict the path along that way. If civilization progresses, AI will vastly  

156:08

exceed the sum of all human intelligence. There will be far more robots than humans. 

156:16

Along the way what happens  is very difficult to predict. 

156:20

It seems one thing you could do is just say,  "whatever government X, you're not allowed to  

156:27

use Optimus to do X, Y, Z." Just write out  a policy. I think you tweeted recently that  

156:31

Grok should have a moral constitution. One of those things could be that we  

156:36

limit what governments are allowed  to do with this advanced technology. 

156:47

Technically if politicians pass a  law and they can enforce that law,  

156:53

then it's hard to not do that law. The best thing we can have is limited government  

157:01

where you have the appropriate crosschecks between  the executive, judicial, and legislative branches. 

157:12

The reason I'm curious about it is that at some  point it seems the limits will come from you. 

157:17

You've got the Optimus, you've got the space GPUs… You think I'll be the boss of the government? 

157:24

Already it's the case with SpaceX that for  things that are crucial—the government really  

157:32

cares about getting certain satellites up in  space or whatever—it needs SpaceX. It is the  

157:37

necessary contractor. You are in the  process of building more and more of the  

157:44

technological components of the future that will  have an analogous role in different industries. 

157:50

You could have this ability to set some policy  that suppressing classical liberalism in any  

157:58

way… "My companies will not help in any  way with that", or some policy like that. 

158:05

I will do my best to ensure that  anything that's within my control  

158:08

maximizes the good outcome for humanity. I think anything else would be shortsighted,  

158:18

because obviously I'm part of  humanity, so I like humans. Pro human. 

158:29

You mentioned that Dojo 3 will  be used for space-based compute. 

158:34

You really read what I say. I don't know if you know,  

158:38

Elon, but you have a lot of followers. Dead giveaway. How did you discern my secrets? 

158:46

Oh I posted them on X. How do you design a chip for space? What changes? 

158:54

You want to design it to be more radiation  tolerant and run at a higher temperature. 

159:03

Roughly, if you increase the operating  temperature by 20% in degrees Kelvin,  

159:08

you can cut your radiator mass in half. So running at a higher temperature  

159:14

is helpful in space. There are various things  

159:20

you can do for shielding the memory. But neural nets are going to be very  

159:26

resilient to bit flips. Most of what happens  

159:30

for radiation is random bit flips. But if you've got a multi-trillion parameter model  

159:37

and you get a few bit flips, it doesn't matter. Heuristic programs are going to be much more  

159:42

sensitive to bit flips than  some giant parameter file. 

159:49

I just design it to run hot. I think you pretty much do  

159:56

it the same way that you do things on  Earth, apart from making it run hotter. 

160:02

The solar array is most of  the weight on the satellite. 

160:04

Is there a way to make the GPUs even more  powerful than what Nvidia and TPUs and  

160:11

et cetera are planning on doing that would be  especially privileged in the space-based world? 

160:18

The basic math is, if you can do about a  kilowatt per reticle, then you'd need 100  

160:31

million full reticle chips to do 100 gigawatts. Depending on what your yield assumptions are,  

160:44

that tells you how many chips you need to make. If you're going to have 100 gigawatts of power,  

160:53

you need 100 million chips that are running at  a kilowatt sustained, per reticle. Basic math. 

161:05

100 million chips depends on… If you  look at the die size of something like  

161:13

Blackwell GPUs or something, and how many  you can get out of a wafer, you can get  

161:18

on the order of dozens or less per wafer. So basically, this is a world where if  

161:25

we're putting that out every single year,  you're producing millions of wafers a month.  

161:33

That's the plan with TeraFab? Millions of  wafers a month of advanced process nodes? 

161:37

Yeah it could be north of a million or something. You’ve got to do the memory too. 

161:42

Are you going to make a memory fab? I think the TeraFab's got to do memory. 

161:46

It's got to do logic, memory, and packaging. I'm very curious how somebody gets started. 

161:51

This is the most complicated  thing man has ever made. 

161:54

Obviously, if anybody's up to  the task, you're up to the task. 

161:58

So you realize it's a bottleneck,  and you go to your engineers. 

162:02

What do you tell them to do? "I want  a million wafers a month in 2030." 

162:09

That’s right. That’s exactly what I want. Do you call ASML? What is the next step? 

162:14

No so much to ask. We make a little fab and see what happens. 

162:22

Make our mistakes at a small  scale and then make a big one. 

162:25

Is a little fab done? No, it's not done. We're  

162:29

not going to keep that cat in the bag. That cat's going to come out of the bag. 

162:34

There'll be drones hovering over the bloody thing. You'll be able to see its construction  

162:39

progress on X in real time. Look, I don't know, we could just  

162:47

flounder in failure, to be fair. Success is not  guaranteed. Since we want to try to make something  

163:00

like 100 million… We want 100 gigawatts of power  and chips that can take 100 gigawatts by 2030. 

163:18

We’ll take as many chips as  our suppliers will give us. 

163:20

I've actually said this to TSMC and Samsung  and Micron: "please build more fabs faster". 

163:28

We will guarantee to buy the output of those fabs. 

163:32

So they're already moving as fast  as they can. It's us plus them. 

163:46

There's a narrative that the people  doing AI want a very large number  

163:50

of chips as quickly as possible. Then many of the input suppliers,  

163:56

the fabs, but also the turbine manufacturers,  are not ramping up production very quickly. 

164:02

No, they're not. The explanation you hear  

164:05

is that they're dispositionally conservative. They're Taiwanese or German, as the story may  

164:11

be. They just don't believe... Is that really  the explanation or is there something else? 

164:17

Well, it's reasonable to... If somebody's been in  the computer memory business for 30 or 40 years… 

164:25

They've seen cycles. They've seen boom and bust 10 times. 

164:32

That's a lot of layers of scar tissue. During the boom times, it looks like  

164:37

everything is going to be great forever. Then the crash happens and they're  

164:41

desperately trying to avoid bankruptcy. Then there's another boom and another crash. 

164:48

Are there other ideas you think  others should go pursue that  

164:51

you're not for whatever reasons right now? There are a few companies that are pursuing  

164:58

new ways of doing chips, but  they're just not scaling fast. 

165:03

I don't even mean within  AI, I mean just generally. 

165:07

People should do the thing where they find  that they're highly motivated to do that thing,  

165:13

as opposed to some idea that I suggest. They should do the thing that they find  

165:21

personally interesting and motivating to do. But going back to the limiting  

165:30

factor… I used that phrase about 100 times. The current limiting factor that I see in the  

165:47

three to four year timeframe, it's chips. In the one year timeframe, it's energy,  

165:56

power production, electricity. It's not clear to me that there's enough  

166:02

usable electricity to turn on all  the AI chips that are being made. 

166:10

Towards the end of this year, I think people  are going to have real trouble turning on... 

166:13

The chip output will exceed  the ability to turn chips on. 

166:17

What's your plan to deal with that world? We're trying to accelerate electricity production. 

166:24

I guess that's maybe one of the reasons that xAI  will be maybe the leader, hopefully the leader. 

166:34

We'll be able to turn on more chips  than other people can turn on, faster,  

166:39

because we're good at hardware. Generally, the innovations from  

166:45

the corporations that call themselves labs,  the ideas tend to flow… It's rare to see that  

166:54

there's more than about a six-month difference. The ideas travel back and forth with the people. 

167:04

So I think you sort of hit the hardware  wall and then whichever company can scale  

167:11

hardware the fastest will be the leader. So I think xAI will be able to scale  

167:17

hardware the fastest and therefore  most likely will be the leader. 

167:20

You joked or were self-conscious about  using the "limiting factor" phrase again. 

167:28

But I actually think there's something deep here. If you look at a lot of things we've touched on  

167:32

over the course of it, it’s  maybe a good note to end on. 

167:37

If you think of a senescent, low-agency  company, it would have some bottleneck and  

167:45

not really be doing anything about it. Marc Andreessen had the line of,  

167:49

"most people are willing to endure any  amount of chronic pain to avoid acute pain". 

167:54

It feels like a lot of the cases we're talking  about are just leaning into the acute pain,  

167:59

whatever it is. "Okay, we got to figure out  how to work with steel, or we got to figure  

168:05

out how to run the chips in space." We'll take some near-term acute pain  

168:09

to actually solve the bottleneck. So that's kind of a unifying theme. 

168:13

I have a high pain threshold. That's helpful. To solve the bottleneck. 

168:19

Yes. One thing I can say is, I think the  future is going to be very interesting. 

168:36

As I said at Davos—I think I was on the  ground for like three hours or something—it's  

168:45

better to err on the side of optimism and  be wrong than err on the side of pessimism  

168:50

and be right, for quality of life. You'll be happier if you err on  

169:01

the side of optimism rather than  erring on the side of pessimism. 

169:05

So I recommend erring on the side of optimism. There's to that. 

169:09

Cool. Elon, thanks for doing this. Thank you. 

169:11

All right, thanks guys. All right. Great stamina. 

169:17

Hopefully this didn't count as  a pain in the pain tolerance.

Interactive Summary

This conversation delves into the future of AI, compute, energy, and space exploration, with a particular focus on the challenges and opportunities presented by scaling these technologies. Elon Musk discusses the limitations of Earth-based infrastructure for AI, highlighting energy constraints and the need for massive scaling, which he argues can only be achieved in space. He elaborates on the benefits of space-based solar power, avoiding day-night cycles and atmospheric losses, making it significantly more efficient and cost-effective. The discussion also touches upon the immense capital requirements for AI development and the potential for SpaceX to become a hyperscaler by launching AI capacity into space. Musk outlines ambitious plans for mass production of solar cells and chips, the development of advanced robotics (Optimus), and the necessity of vertical integration to overcome supply chain bottlenecks. He emphasizes the critical role of identifying and addressing limiting factors in any endeavor, whether it's energy, chip manufacturing, or rocket development. Furthermore, the conversation explores the philosophical implications of advanced AI, the mission of xAI to understand the universe and propagate intelligence, and the importance of truth-seeking and robust values in AI development. Musk expresses concerns about government overreach and the potential for AI to be misused by states, advocating for limited government and robust checks and balances. Finally, he shares insights into his management philosophy, focusing on urgency, addressing bottlenecks, and hiring for talent, drive, and trustworthiness.

Suggested questions

12 ready-made prompts