HomeVideos

The gap is widening

Now Playing

The gap is widening

Transcript

799 segments

0:00

All right. So, I want to talk about what

0:02

happens next. And this is with respect

0:04

to Open Claw and Molt Book and Rent a

0:07

Human and all that that kind of fun

0:09

stuff. So, if you're not if you're not

0:10

up to speed, here's a 202 catch you up

0:13

to speed. Uh, a few weeks ago, someone

0:16

or actually guess it was end of 2025.

0:18

Someone created OpenClaw or what was

0:20

called Claudebot at the time. Fully

0:22

autonomous uh agent that, you know,

0:25

works for you around the clock. uh runs

0:28

on scripts and cron jobs and it wakes up

0:30

at a you know regular cadence to go do

0:32

stuff autonomously for you and then

0:36

someone made claude book um or no claude

0:39

book whatever it was called um basically

0:43

read it for uh sorry mold book there we

0:46

go my brain was like no that's not right

0:48

mold book so basically read it for these

0:50

agents now that immediately became a

0:53

cess pit for cryptog and humans writing

0:56

posts and that sort of thing. But it

1:00

made open claw very popular. And then

1:03

someone extended that with rent a human,

1:05

which is basically uh AI agents can go

1:09

pay a human to do something for them.

1:12

Now what's happening? So what is what is

1:15

legitimately actually physically

1:17

happening in the real world is yes mold

1:19

book was full of grift. Um but it shows

1:24

a path forward. At the same time what is

1:27

actually physically true is that lots

1:30

and lots of people hundreds of thousands

1:31

of people if not millions of people are

1:34

using openclaw around the world uh to do

1:38

real work. And you know, people talk

1:41

about it and and the conversation is as

1:43

sharply divided about this kind of agent

1:46

as it was for chat bots when chat bots

1:49

first got big cuz you still see see

1:51

people out there saying, you know, I

1:53

don't I don't understand a legitimate

1:54

use case for chat GPT and it's like,

1:56

okay, buddy, you know, you're you're

1:58

just a lite and you're going to live in

1:59

your cave. I remember when when chat GPT

2:03

first launched and I was still on the

2:04

OpenAI forums, there were guys going

2:06

around saying, "This is just Eliza. Tell

2:08

you need to prove to me that this is

2:10

like people get so huffy and indignant

2:13

and like angry when the paradigm

2:15

shifts." And we're seeing the same exact

2:17

thing. I think that this is why it took

2:19

me a while to like figure out how to

2:21

talk about this is because the reaction

2:23

is so sharp around openclaw and these

2:27

this idea of autonomous agents that it's

2:30

it's like that basically we've had a few

2:32

paradigms. So from my perspective having

2:35

been in this since been in this like

2:37

full-time since GPT2 um so we had we had

2:40

the original LLMs which were literally

2:42

just autocomplete engines. They weren't

2:44

even chatbots yet. So that was paradigm

2:45

one. Then paradigm two was the instruct

2:48

models where I was like okay well

2:50

instead of having to give give every

2:53

single time you give a a prompt example

2:56

let's just go ahead and pre-train the

2:57

models to follow simple instructions.

3:00

Well once we had that it was a hop skip

3:01

and a step to get chat bots cuz then the

3:03

instruction is just be a chatbot with

3:05

you know x y and z personality. So we

3:08

had plain vanilla, we had instruct

3:10

aligned, then we had chat bots. And chat

3:12

bots was the first like you could say

3:14

instruct aligned was like paradigm 1.5

3:17

because it was still basically just an

3:20

autocomplete engine, but it was an

3:21

autocomplete engine that was designed to

3:23

follow at least a single instruction

3:25

relatively well. The chatbot was a was a

3:27

fundamentally different UX. So chatbot

3:30

was paradigm 2. And I didn't even use

3:32

chat GPT for the first few months cuz

3:34

I'm like I don't care. Like that's just

3:36

a chatbot. such as a distraction for

3:38

this this model. But of course, time has

3:42

proven that once you add on tool use and

3:45

you add on reasoning and a bunch of

3:47

other things, then the chatbot actually

3:49

becomes semiautonomous.

3:51

You know, like I still regularly use

3:53

chat GPT Pro when I'm doing heavy duty

3:56

math research to model like, you know,

3:58

how many how many jobs were eliminated

4:00

in 2025. And by the way, cross-checking

4:03

across multiple AIs, it looks like

4:05

artificial intelligence ultimately

4:07

either destroyed or avoided about

4:08

200,000 to 300,000 jobs in America alone

4:12

last year. Um, so anyways, just wanted

4:15

to throw that little factoid out. The

4:17

official number is 54,167

4:20

or something like that. Um, so the the

4:23

real number and the way that you detect

4:24

this is there's kind of two ways you

4:26

look at number one, you look at um

4:30

excess deaths. So during co times and

4:32

I'm sorry for this tangent but this is

4:34

actually kind of important and it's not

4:35

I don't I don't want to do a whole video

4:36

on it. So anyways during co times we did

4:39

excess deaths which is basically all

4:41

else being equal you look at how many

4:44

extra people died in a given time and

4:47

place and you say okay well the primary

4:50

difference is COVID. So then you

4:52

attribute those excess deaths to CO

4:54

exposure. So we did the same thing for

4:56

jobs which is all else being equal

4:58

looking at things like inflation data

5:01

and um and in and interest rates and

5:04

those sorts of things. You can generally

5:06

anticipate how many layoffs should have

5:08

happened and then so then you say okay

5:10

well how many excess layoffs did we

5:12

have. Then the other thing that you can

5:14

do is you can look at it the opposite

5:15

direction which is labor growth. So then

5:17

you compare things like GDP growth to

5:20

the actual number of jobs created and

5:22

you look at the at the difference

5:23

between those. And when you correct for

5:25

everything else like you know Doge

5:27

layoffs and whatever else was going on

5:29

last year um then you get somewhere

5:32

between like a 100,000 and 350,000

5:36

uh jobs were either destroyed or

5:38

avoided. And this this actually goes

5:40

back to an AI layoff is not necessarily

5:44

just someone getting fired handed a pink

5:46

slip saying AI was responsible for this.

5:49

It is that new jobs are not being

5:51

created. Uh, so anyways, um, that's an

5:55

example of what I use ChatGBT Pro for

5:59

and Gemini Pro and um, I don't use

6:02

Perplexity Max anymore, but I have a

6:04

Claude Max. So, I use all of these

6:06

different tools, but they're just

6:08

waiting for me. So, the the next big

6:10

paradigm shift has been agents. And

6:13

people have been trying to crack the nut

6:14

of agents for years now. like the ACE

6:18

framework if you remember like ACE

6:19

framework the Raven framework um we've

6:21

been trying to crack this nut for a

6:23

while and then dude with open claw

6:26

figures it out so here we are now what I

6:29

will say is that adding reasoning and

6:32

tool use was kind of paradigm 2.5 cuz

6:34

that was still fundamentally the same

6:35

form factor as chat bots so it was chat

6:38

bots with a little bit of extra bolted

6:40

on but then those chat bots with a

6:42

little bit extra bolted on has evolved

6:44

into openclaw So, Paradigm 1 was just

6:47

autocomplete engine and then we made

6:49

them a little bit better until they

6:50

evolved into chatbots. It's like

6:51

Pokemon, honestly. So, you know, the the

6:54

the the plain vanilla original Pokemon,

6:57

you know, that you get right at the

6:59

beginning of the game is just the

7:00

autocomplete engine. Stage two is the

7:02

chatbot. Stage three is the agent. Now,

7:06

the agent allows for a lot of emergence.

7:08

So you've probably heard uh everyone

7:10

from me to systems thinkers like Daniel

7:13

Schmokenberger and everyone else even

7:15

the AI safety people we talk about

7:17

emergence. So emergence has a few

7:20

connotations in in pure AI the idea is

7:24

that new abilities emerge as you get

7:26

certain levels of sophistication data

7:29

parameter counts architectures those

7:31

sorts of things. So like theory of mind

7:33

seems to have emerged. Um, but what has

7:37

what has also happened is that once you

7:39

know to look for an ability in a

7:40

language model, it's usually been there

7:41

all along. It was just not prominent

7:43

enough to be useful. So like theory of

7:45

mind, GPT2 has planning and theory of

7:47

mind. It's not very good, but it's in

7:50

there. So then it got better with GPT3

7:52

and people still didn't even know to

7:53

look for it even though a lot of us were

7:55

talking about it. Um, and then once you

7:58

get to GPT4 and GPT5, then people are

8:01

like, "Oh yeah, it's clearly got good

8:03

enough theory of mind." um and planning

8:05

and reasoning. It's always been there.

8:06

It just wasn't that good. So now it's

8:09

there. It's better. It's useful. And in

8:11

fact, in in the case of theory of mind,

8:13

um modern like frontier models are

8:16

generally better than the average human

8:18

at theory of mind. Um so that's that's

8:20

one example of emergence. But another

8:23

version of emergence because that's

8:24

within a single system. um within in in

8:28

larger systems in complex systems or or

8:31

um well yeah I guess in in systems

8:33

theory emergence can be like emergent

8:36

gameplay is an example of emergence

8:38

where you have a game that is

8:39

sophisticated enough with enough rules

8:42

that people can kind of make their own

8:44

games in this. So um some examples are

8:47

like Minecraft and Roblox and those

8:49

sorts of things where there's enough

8:51

different game mechanics. you have

8:52

enough you have a you have a

8:53

sophisticated enough alphabet of game

8:56

mechanics that you can get spontaneous

8:58

uh new forms of gameplay. So all the big

9:01

games like Fortnite and stuff where

9:02

there's a building mechanic, a survival

9:04

mechanic, a a fighting mechanic and

9:07

those sorts of things, you get emergent

9:09

gameplay. Um so that that is that is a

9:12

real life example of emergence. Now when

9:15

you apply that to things like agents, so

9:18

now you have agents interacting with

9:19

each other and they're interacting with

9:21

each other, you know, locally, they're

9:23

interacting with each other publicly,

9:24

they're interacting with humans, they're

9:26

interacting with businesses. That's

9:28

where you get another layer of emergence

9:30

because the thing is is with chat bots,

9:32

you had a very constrained environment

9:35

where a chatbot was interacting with

9:37

you. You were pretty much the the

9:39

primary variable. And then of course you

9:40

know it it had tool use and retrieval

9:43

augmented generation and reasoning

9:44

abilities as kind of the three primary

9:46

food groups. Obviously that's they have

9:48

lists of tools. Um but they can get

9:50

information from elsewhere. They can run

9:53

programming and other tools and then

9:55

they have reasoning abilities. So those

9:57

are the three big variables. But it was

9:58

still a constrained environment because

10:00

it was just going to do a process. You

10:02

know they could think I think my longest

10:04

chat GPT pro was like 58 minutes or

10:06

maybe just over an hour. and then it's

10:08

going to spit out the results and then

10:09

wait for you. So that there's a there's

10:11

a very definitive time step and it's

10:14

based on the human. However, whenever

10:16

you have complex systems, you introduce

10:19

more and more variables which which

10:22

ultimately creates more chaos and I mean

10:24

this in like the strict like the the the

10:26

legitimate like chaos theory because

10:29

with a single chatbot doing a single

10:31

thing, what we're trying to do with

10:32

safety and alignment is reduce the

10:34

chaos. Meaning we know that every time

10:37

you use Claude Max or G Chat GPT Pro or

10:40

Gemini Pro or Gro Heavy, the input

10:44

processing and output cycle is going to

10:45

be very predictable. And what I mean by

10:48

predictable is it's not going to go off

10:49

the rails. It's not going to do any

10:51

financial harm. It's not going to

10:52

accidentally teach you how to make drugs

10:54

or nukes or that sort of thing. And so

10:57

that's the time step. So we've basically

10:58

been operating up until now on an

11:01

individual loop. So the loop is human

11:04

puts something in, the AI goes and does

11:06

something and then gives the human

11:07

output. So that's the input processing

11:09

output loop and that's the loop and it

11:11

just keeps looping. But what openclaw

11:13

does is it makes the loop not dependent

11:16

on humans. So then not only do you have

11:19

incremental time steps that are not

11:21

limited to humans, you have incremental

11:23

time steps that are influenced by other

11:25

agents and other environments. And that

11:27

introduces basically irreducible

11:30

complexity or or chaos. And so this goes

11:33

back to the original work that that we

11:36

did and I say we because we had a very

11:38

large community working on Gateau. So

11:40

that's the global alignment tax taxonomy

11:42

omnibus framework which we we realized

11:46

you know anyone working on cognitive

11:47

architecture as of three plus years ago

11:50

realized and I remember one of the one

11:51

of the guys that we collaborated with he

11:53

said I realize that these things are

11:54

going to be talking to each other more

11:55

than us very soon and that has turned

11:58

out to be true. Now, the reason I'm

12:00

bringing all of this up is because from

12:02

a risk and compliance and safety and

12:05

liability standpoint,

12:07

um it like openclaw is intolerable. And

12:11

what I mean by intolerable is that no

12:13

frontier lab is going to touch this

12:15

anytime soon. Um no Fortune 500

12:18

enterprise is going to touch it anytime

12:20

soon either. one of the biggest barriers

12:23

of adoption for even professional-grade

12:26

agentic frameworks and I I have Silicon

12:29

Valley founder friends um who who are

12:31

who are building these kinds of things

12:34

is number one interoperability. So like

12:37

what a lot of people expect an agent is

12:39

going to be is just a digital human that

12:41

you drop in and there are some people

12:43

working towards that where you just put

12:44

it in a in a a virtual machine give it a

12:47

virtual Windows desktop and away it

12:49

goes. But um the the the native space

12:53

that these things live in is more like

12:55

OpenClaw where it's all terminals, it's

12:57

all command lines, it's all API calls.

12:59

So rather than saying, okay, well, you

13:02

need to um have a keyboard, a mouse, and

13:05

and a screen and a virtual screen that

13:07

you have to look at and interpret, it's

13:09

like, why don't I just give you the the

13:11

information directly? Um so they're

13:13

they're they're not guey native,

13:16

graphical user interface native. It's

13:18

actually kind of like back to old

13:19

school. And this is why whenever you

13:21

watch people that are like hacking on

13:23

OpenClaw and stuff, they have like four

13:25

monitors up and it's all terminals. It

13:27

looks like the Matrix, right? Um, and

13:29

that's because their native environment

13:30

is textbased, meaning that the terminal

13:32

output from, you know, command line or

13:35

APIs is natively what they they read.

13:39

It's there's actually too much

13:40

information. There's too much noise if

13:42

you give them a screen, if you give them

13:44

a camera. Uh so

13:47

that is all right that was a little bit

13:49

of a tangent but I I felt like I needed

13:50

to explain it but the the overarching

13:53

point here is that from a security so

13:56

there's there's a few ways to look at

13:57

this. So if we if we put our you know um

14:00

Fortune 500 hat on we say all right so

14:03

you got cyber security and they say well

14:05

what are the risks and it's like oh well

14:07

you give it root access to a virtual

14:09

machine and it can have prompt

14:11

injections from from um infected skills

14:14

that it downloaded online and so cyber

14:16

security is going to say banan it say

14:19

this is not a product this is

14:21

functionally malware um and and in many

14:23

cases that is one of the best ways to

14:25

characterize openclaw from a cyber

14:28

security perspective. I'm not saying

14:29

that it is literal malware, but I'm

14:31

saying having worked in Fortune 500

14:34

companies, numerous of them. I'm telling

14:36

you that is how cyber security would

14:38

treat it. And honestly, as an as a

14:40

former uh automation uh infrastructure

14:43

guy, I would say like, yeah, you can't

14:45

trust this because you if it if it

14:47

touches one of your routers or one of

14:49

your switches or one of your servers or

14:50

one of your storage arrays and it runs

14:52

the wrong command, it shuts everything

14:54

down and you're losing, you know, in

14:57

some cases tens of millions of dollars

14:58

per hour on top of reputation damage and

15:02

on top of legal liability. So, it will

15:05

take a very long time. And when I say

15:08

very long time, the fastest I could see

15:11

many um Fortune 500 companies deploying

15:14

something like not OpenClaw directly,

15:16

but a successor to OpenClaw is maybe 18

15:19

months. And that is just because that's

15:21

how long it takes for you to do an

15:24

infrastructure audit, a cyber security

15:26

audit. Um they will have like toy

15:29

versions set up to play with. And you

15:30

know, cyber security team, they'll

15:32

they'll be monkeying around with it

15:33

already. But one of the other things

15:36

that you absolutely need to have is

15:39

executive buyin at the very top. And and

15:41

when I say at the very top, I don't mean

15:43

your chief technology officer. It has to

15:45

come from the CEO and ideally the board

15:48

of directors. Um the the the companies

15:51

that we are seeing that are and this is

15:53

this is all organizations. It doesn't

15:55

have to be a for-profit company. It

15:57

doesn't have to be a Fortune 500

15:58

company. It can be a government. It can

16:00

be a mid-size company. It can be an

16:02

enterprise. It can be a small company.

16:05

The only companies, the only

16:06

organizations that are successfully

16:08

making this this pivot is when the owner

16:11

or whoever is the highest stakeholder,

16:13

whoever it happens to be, whether it's

16:15

the CEO, the owner, board of directors,

16:18

the governor, whoever it is, if they

16:21

issue an edict saying we are going allin

16:24

on AI and they are the ones leading the

16:26

charge. And the reason is because AI is

16:29

so fast moving and it's so scary and

16:33

organizations are so riskaverse that if

16:35

you don't see the head honcho constantly

16:38

saying oh look what I did with chat GPT

16:40

then everyone else is going to be like I

16:42

might use chat GPT a little bit but I'm

16:44

going to hide the use and I'm not going

16:45

to do that kind of thing. my wife's

16:48

company, it's not a company she owns,

16:49

it's a company she she contracts for.

16:52

The owner and CEO and founder, all the

16:54

same person, on their weekly meeting

16:56

says, "Tell us what you used AI for."

16:58

And so, he's created a culture of, "Hey,

17:02

we're we're treating AI like a first

17:04

class asset." It's like, "Imagine

17:07

imagine you didn't have computers and

17:09

you didn't have internet." And it's

17:10

like, "Oh, well, the factory has never

17:11

needed computers or internet, so we're

17:13

just going to kind of ignore it." That's

17:15

how most companies and organizations are

17:17

reacting. Um, but the companies that are

17:20

saying, "Hey, we don't know how to use

17:21

this, but go play with it." Um, and you

17:23

know, it's like play with it until you

17:24

go blind. And if you know, you know.

17:26

Anyways, this is a little bit of a dirty

17:28

joke. Um, the idea is that you need

17:30

executive sponsorship. And when I say

17:32

executive sponsorship, I mean like

17:34

wholehog buyin. They need to be the ones

17:37

leading the charge. In fact, our

17:39

consultation business, our side business

17:40

that we're working on, we have decided

17:42

that we will not work with any clients

17:46

for whom like it's not coming from the

17:48

top because the thing is is in the

17:51

consultation space you might have um you

17:54

might have a company or a manager or you

17:56

know even a sea level we have we have

17:59

seen cases where the sea level says we

18:01

need to go allin on this but you know

18:04

the the chief financial officer chief

18:07

head of legal counsel, chief of HR,

18:10

they're all either ambivalent or

18:12

skeptical. And so the CEO is like,

18:13

"Well, you know, do what you can, but

18:15

it's not a top priority." If the CEO

18:17

says it's not a top priority, we walk.

18:19

Like, we we just we know that they're

18:21

not ready. And the reason I'm telling

18:23

you all of this, and I know that most of

18:25

you have never been in a Fortune 500

18:28

company. Most of you have never worked

18:29

for and have no interest in working for

18:31

a Fortune 500 company. The reason I'm

18:34

telling all of you this is because one

18:35

that's like they did that for 15 years.

18:38

But then also so that you are aware that

18:40

the gulf between what you know is

18:43

possible on the cutting edge and where

18:46

most people are is actually getting

18:48

wider because we still have people

18:50

saying like well I don't really see a

18:52

role for chat bots in enterprise. I

18:55

don't really see a role for chat bots in

18:57

government. and they're all so

18:59

riskaverse and they're so slowmoving,

19:00

it's going to take years for them to

19:02

really come around. Meanwhile, the ones

19:05

that are rapid adopters are going to be

19:07

pulling ahead. And honestly, when this

19:12

in in hindsight, imagine you had a

19:14

company that said, "Well, we don't

19:15

really get this internet stuff. We don't

19:17

really get this personal computer stuff.

19:19

We don't really get this cloud stuff."

19:20

What happens? A lot of them go out of

19:22

business. Look at Borders, right?

19:24

Borders is is one of the primary

19:25

examples where they said books that that

19:28

that's not changing. People like having

19:30

a physical book in their hand. And here

19:32

comes Jeff Bezos with his, you know,

19:33

warehouse robots and bulldozes, that

19:36

whole thing. Barnes & Noble managed to

19:38

survive somehow. I don't even know how

19:39

cuz when was the last time you even went

19:41

to a Barnes & Noble, but Borders is long

19:43

gone. And so what I'm what I'm telling

19:45

you is that the companies that survive,

19:48

that decision is being made now. But

19:51

it's like it's kind of like, you know,

19:52

you you see in a war movie or or

19:54

whatever where it's like, "Oh, he's

19:55

already dead. His body just hasn't

19:57

caught up to the fact yet." I don't

19:58

remember what movie that was from, but

20:00

it it's it's a trope, right? It's like

20:02

the there are already zombie companies.

20:04

There are dead men walking out there and

20:07

it comes from the top because of their

20:09

attitude towards artificial

20:10

intelligence. If they're dragging their

20:12

feet on chat bots, they're they're

20:14

hardly using any language models at all.

20:17

you know, they might still be using um

20:19

one of the one of the first use cases

20:21

that has diffused out is using small

20:23

language models to do things like help

20:25

with automatic routing of tickets and

20:27

that sort of thing. You know, the in in

20:29

business theory, there's a there's a

20:31

concept called forgivability, which

20:33

means what's the cost of doing it wrong

20:35

and how difficult is it to reverse? And

20:38

so, if you route a ticket to the wrong

20:39

person, that has basically zero cost.

20:42

It's like it's going to cost someone a

20:43

human 5 minutes to route it back to the

20:45

right place, but no lives are lost, no

20:48

money is lost, and there's no legal

20:50

exposure. So, you go through who's

20:52

everyone that you need to convince on

20:54

the board or, you know, at city council

20:56

or town hall or the governor's office

20:59

or, you know, the seauite. So, you've

21:01

got legal, you've got finance, you've

21:03

got HR, um you've got cyber security. So

21:06

even when the tech nerds over here say

21:09

guys this is very clearly the way of the

21:11

future you're going to get a lot of

21:13

resistance and in some cases it is

21:16

rightly so because to be fair a lot of

21:19

us tech guys want to move faster than

21:21

the rest of the organization is even

21:23

capable of moving fast or is or is would

21:26

even be um would even be uh wise to move

21:30

that safe. And the reason is because you

21:32

really have to dot your eyes and cross

21:33

your tees and justify, okay, if we roll

21:36

out these chat bots, which by the way,

21:37

Microsoft is charging $40 per month for

21:40

co-pilot. And you know, that doesn't

21:42

sound like much. And from my

21:44

perspective, the argument is very

21:46

simple. It's like in that for in that,

21:48

you know, for for every single seat,

21:50

will you get $40 more value to the

21:52

organization than that cost? So, like

21:56

the ROI for for people who use chat bots

21:59

is very obvious where it's like, hey,

22:02

you're paying someone effectively $50 to

22:03

$60 an hour. All you have to do is get

22:06

an extra hours worth of value to the

22:08

company per month out of that employee

22:11

and the cost is justified. But that's

22:13

not even how CFOs think. CFOs say, "Yes,

22:16

but one, how do we track that? And then

22:20

two, what if they just get the same

22:22

amount of work done faster and then

22:24

they're lazier?" So those are very real

22:26

considerations, but for most people like

22:30

especially smaller organizations where

22:31

it's like, yeah, we're all using chat

22:33

GPT or Gemini or whatever all day every

22:36

day and we are 10 times more effective

22:38

as a team. That is much easier to

22:40

achieve when you have a 20 or 30 person

22:42

organization rather than a 20 or 30,000

22:45

person organization. And so what what

22:48

we're and I'm reporting from the front

22:49

lines. This is what we're literally

22:51

seeing out there in the enterprise space

22:53

is that you have um you have like how am

22:58

I trying to say this? You you have a lot

23:01

of job roles where chat bots really

23:06

don't help that much. Um and then and

23:09

and that's a true thing. Like if you

23:10

spend most of your day in a truck, you

23:12

know, moving heavy materials around or

23:14

meeting with people face to face, you

23:16

might use a chatbot at the end of the

23:18

day to help write reports. Now, to be

23:19

fair, if you spend four or five hours a

23:22

week writing reports and the chat bots

23:24

can help you do it in 30 minutes, so

23:26

then you can spend more time face tof

23:28

face with clients, with customers, with

23:31

suppliers, then again, that should be a

23:34

no-brainer. But that's not how CFOs

23:36

think. And then of course your chief

23:37

legal counsel is just like you know

23:39

here's here's an example that I learned

23:41

recently is that uh insurance will not

23:45

underwrite AI right now. They don't know

23:46

how to price it in. Um and so well if if

23:49

insurance won't write it then there

23:51

that's an an an intolerable amount of

23:54

financial risk andor legal risk. Uh so

23:57

then legal says shut it down. Don't use

23:59

it. Now, of course, there's often a lot

24:02

of hypocrisy because finance might be

24:05

using the chatbot as shadow IT. Legal

24:07

might be using the chatbots as shadow

24:09

IT. HR might be using the chatbots as

24:11

shadow IT. So, then you have this other

24:13

problem. And this is where your your C

24:16

your CISO, your chief information

24:18

security officer comes in. So, the CISO

24:20

says, well, we have this problem with

24:22

with shadow IT. Legal's using it and we

24:24

can't tell them to stop. HR is using it

24:26

and we could probably tell them to stop.

24:29

um all of it is using it and we can we

24:32

can definitely you know lean on on the

24:35

CTO to have a lockdown policy but people

24:38

are going to be trying to use it anyways

24:40

regardless of what we try and say or do.

24:43

So then there's almost kind of a forcing

24:44

function or a binding function because

24:46

these things are just not ready and and

24:49

when I say these things I'm specifically

24:51

referring to like the agents because

24:53

what the journey that we're on is

24:54

there's the kind of the three phases.

24:56

There's the plain vanilla autocomplete.

24:58

There's the chat bots. And then there's

24:59

the agents. So all of this is to is to

25:03

show that it's going to take longer than

25:06

is remotely reasonably

25:09

uh uh I guess considerable or what is

25:12

what is reasonable to deploy these

25:14

things because often the limitation is

25:16

not technology

25:18

particularly with fastmoving new stuff.

25:20

The limitation is the rest of the

25:22

organization. And so like I have I have

25:24

a a friend a professor friend who also

25:27

consults for the state government and

25:29

it's like you have to have so many

25:32

endless meetings and all of us tech

25:34

people get so bored with it. We're like

25:36

guys you are like 2 or 3 years behind

25:38

the curve. It's time to move on. And

25:42

this is not to say that like you know

25:43

the singularity is canled or that sort

25:46

of thing but if you're watch if you are

25:48

one of the people watching this video

25:49

you are at the tip of the spear like you

25:51

are on the bleeding edge. you're on the

25:53

cutting edge and you know what's coming

25:55

and you know that it's going to take

25:56

like you know chat GPT's been out for

25:58

what three years now and and people are

26:01

still arguing like oh I don't really see

26:02

any b business value for that it's like

26:04

okay well that's that's a you problem

26:05

friend likewise there's people there's

26:08

there's high-tech cutting edge people

26:10

out there who are honestly saying I

26:12

don't see a use case for openclaw I

26:14

don't see a use case for maltbook I

26:16

don't see a use case for rent a human

26:18

this is all silly and that's just not

26:20

how general purpose technologies evolve.

26:23

What we're seeing right now is like

26:26

taking electricity and and you know, use

26:30

case one of electricity was like light

26:32

bulbs, which is basically you just short

26:34

it out, right? If if you if you have a

26:36

if you have a current and you short it

26:38

out, it makes light and heat. So it's

26:40

like a light bulb is like the equivalent

26:42

of the autocomplete era. And then

26:45

electric motors came around and it's

26:46

like, oh wow, if you do some coils and a

26:49

stator and some magnets, you can

26:51

actually convert that current into a lot

26:52

of torque. Cool. But then you have more

26:56

sophisticated uses of electricity like

26:59

communication being one of the most

27:01

sophisticated things. And then

27:03

computation, all of that is still

27:04

predicated. So like communication like

27:07

technology so so radio, telephone,

27:10

telegraph um and and switched networks

27:13

that's like layer three that's like uh

27:15

iteration number three of of the uses of

27:18

electricity as a general purpose

27:19

technology and then computation is is

27:22

layer four. So like that's a fourth

27:24

order consequence of electricity. It was

27:26

not immediately obvious when the first

27:29

people said, "Oh, like if we put these

27:31

chemicals together with different

27:33

metals, then we get like a little spark.

27:35

Hey, cool. What does that mean?" I don't

27:37

know. Or if you know the first time

27:38

someone spun up a dynamo and it's like,

27:40

"Oh, cool." Like it can give a little

27:41

zap or a little arc. What do we use this

27:43

for? We can use it to make rocks think

27:45

like that was not obvious. And my point

27:48

there is that it's not obvious to go

27:51

from an autocomplete engine which is

27:52

just a token producer to then reshape

27:55

that token producer to a chatbot and

27:57

then give it reasoning and give it tool

27:58

use and then reshape that again into

28:00

autonomous agents. It was obvious to us

28:03

technologists. That's why I went that I

28:05

literally decided to go all in on AI

28:07

when GPT3 came out and that's how it

28:09

became my career because I said this is

28:11

very obviously the next big thing. But

28:14

the diffusion part this is what we're

28:16

in. We're in the long slog of diffusion,

28:20

meaning it's going to take time. It's

28:22

going to take experimentation. There's

28:23

going to be emergent risks. There's

28:26

going to be emergent benefits. There's

28:27

going to be emergent form factors. All

28:30

kinds of stuff. And this is why, you

28:32

know, like, okay, let's just use an

28:33

example for something that didn't pop

28:35

off, which is VR. So, the metaverse, VR,

28:38

XR, AR. The idea was that we have a new

28:40

primitive which is, you know, basically

28:42

a head-mounted device, so an HMD as an

28:45

interface to cyerspace. And so, you

28:48

know, if you go all the way back to the

28:49

1980s and and maybe even before with

28:52

particularly uh Japanese culture, manga

28:55

and and anime, the idea that you'd have

28:57

a headset or some kind of headmounted

28:59

device and that you'd dive into the

29:02

cyberspace and the etherwebs and

29:04

whatever else, that [snorts] was just

29:06

people thought it was a foregone

29:07

conclusion that that was the future. We

29:09

actually have those headmounted devices

29:11

now and nobody uses them. Um so there

29:13

it's like okay we we invented one new

29:16

technological primitive and it didn't go

29:17

anywhere. So it's understandable that

29:19

people say okay well you have you have

29:21

you know tok you have a uh you next

29:23

token prediction and that's a new like

29:26

technological primitive. So yeah some

29:28

people assume it's not going to go

29:29

anywhere but they're wrong just plain

29:32

and simple. And even though they're

29:34

wrong it's still going to take time for

29:36

this technology to diffuse out for it to

29:38

mature and those sorts of things. So, I

29:40

don't really mean to like rain on the

29:42

parade, but like when you when you

29:44

straddle two worlds and the two worlds

29:46

that I'm talking about right now is like

29:48

the frontier, like what's happening in

29:50

real time, what is actually physically

29:52

possible versus what the companies and

29:56

governments are actually going to do.

29:57

It's very depressing. So, [laughter]

29:59

that's all there is to it. Talk to

30:01

you'all later.

Interactive Summary

The video discusses the evolution of AI, moving from basic autocomplete engines to chatbots and now to autonomous agents like OpenClaw. It highlights that while early iterations like Mold Book were rife with grift, they pointed towards a future where AI agents perform real work. The speaker draws parallels between the current divided reaction to autonomous agents and the initial skepticism surrounding chatbots, suggesting that many resistant individuals are simply unfamiliar with or unwilling to adapt to paradigm shifts. The development of AI is presented as a progression through distinct paradigms: 1.0 (autocomplete), 1.5 (instruct-aligned models), 2.0 (chatbots), and the emerging 3.0 (agents). The speaker emphasizes that capabilities like reasoning and tool use, when added to chatbots, paved the way for more sophisticated agents. The concept of emergence in AI is explained, both in terms of new abilities arising within models (like theory of mind) and in complex systems where agents interact. The video then delves into the significant barriers to enterprise adoption of AI agents, particularly OpenClaw, citing cybersecurity risks, lack of interoperability, and the need for executive buy-in from the highest levels (CEO/Board). The speaker contrasts the rapid adoption in smaller, agile organizations with the slow, risk-averse nature of large enterprises and governments, likening resistant companies to "zombie companies." Finally, the speaker concludes that while the potential of AI agents is immense, the diffusion and adoption process will be lengthy and complex, citing historical parallels with the adoption of electricity and the limited success of VR as a new technological primitive.

Suggested questions

5 ready-made prompts