HomeVideos

This is the WAY OF THE FUTURE

Now Playing

This is the WAY OF THE FUTURE

Transcript

585 segments

0:00

So, the entire internet seems like it's

0:01

losing its composure over Claudebot. And

0:04

what I wanted to do was just kind of

0:05

explore what's going on. Now, I spent a

0:08

little time kind of reading up on it and

0:10

figuring out what is it, how does it

0:12

work, and that sort of thing. So, let's

0:14

just start at the top. Claudebot is a

0:17

semi-autonomous

0:19

personal agent that the primary

0:21

difference is that it is proactive. It

0:24

finds stuff to do rather than just

0:26

waiting for you to give it commands.

0:28

Now, this is a non-trivial

0:30

uh problem and it's one that I had been

0:32

working on for quite a while and I'll

0:34

talk about my previous work um on this

0:37

uh in a little bit. But the long story

0:39

short is that uh people are freaking out

0:42

about it because this is the level of

0:44

autonomy that the a the agentic browsers

0:47

were supposed to promise, but those were

0:49

built in more of a uh let's say a a

0:54

paradigm that was more corporate

0:55

friendly. Whereas Claudebot is open-

0:58

source and it's rogue. It's very

0:59

renegade. And the reason that it emerged

1:02

in the open-source space is because of

1:05

the lower risk. It's saying, "Hey, this

1:07

is open source software. Use at your own

1:09

risk. If it deletes, you know, all of

1:11

your emails or, you know, buys a million

1:13

platton tickets to Tahiti, that's on

1:15

you." Whereas, you know, the Comet

1:17

browser and the Open AI browser, they

1:19

had to make sure that it couldn't do

1:21

those things. So, this is where the

1:24

open- source movement has a very

1:25

distinct advantage over corporations,

1:28

over closed source, because it's just

1:30

like, I'm just going to release this and

1:31

see what happens. So, people are

1:33

freaking out about it. Um, this this

1:35

tweet in particular was hilarious. So,

1:38

Flowers, this is this is the same person

1:39

who was formerly uh Future Flowers um

1:42

said, "We should give Claudebot Minute

1:44

Man 3 access for a fast fun alignment

1:46

test."

1:48

Um, so the idea is like, okay, like what

1:51

kinds of guard rails does this have and

1:53

what, you know, what are the safety

1:55

risks? And on a technical level, what a

1:58

lot of people are concerned about is

1:59

that you have uh an AI that's constantly

2:02

running. It's got ports open, so it's

2:04

hackable, so it's a security nightmare.

2:06

Uh, so that's one thing that people are

2:08

worried about. But you can run it on a

2:10

local PC, you can run it on a Mac Mini,

2:12

you can run it on um a micro PC and then

2:15

it just becomes your just little shotgun

2:17

ride along buddy. You could put it in a

2:19

container. Um if you containerize it,

2:21

you can run it on anything including

2:23

your phone. So this is very obviously

2:25

the direction that things were going to

2:26

go. Uh and the reason that like all the

2:30

all the primitives have been worked on.

2:31

So there's a few technological

2:34

primitives and and what I mean by

2:35

primitive is like the basic building

2:37

blocks. So the technical primitives that

2:39

we've been working on over the last few

2:40

years and I say the royal we like all of

2:42

us um is uh so number one models that

2:46

were capable of agency so you know

2:48

taking tasks or uh uh taking taking

2:51

commands solving tasks solving problems

2:54

um and that sort of thing. Tool use was

2:56

one of the next big ones. So tool use

2:58

the ability to use JSON the ability to

3:01

use APIs the ability to even say I don't

3:03

know how to use that API. Let me let me

3:05

find the documentation. So a lot of

3:07

those autonomous like those th those

3:09

fully autonomous tasks that were done in

3:12

service to userdirected tasks uh were

3:15

like kind of some of the building

3:16

blocks. Uh memory management so

3:18

recursive language models that

3:20

innovation has been a big contributor

3:22

because one of the biggest problems is

3:24

not you know uh what like what can you

3:27

do it's remembering what should you do

3:30

what does this person need. So things

3:32

like retrieval augmented generation was

3:34

kind of the first uh version of

3:36

long-term memory, but it was really

3:38

unstructured. It was basically just a

3:40

soup of memory um and and had very very

3:43

little structure. So recursive language

3:45

models are a better more structured uh

3:48

memory management system. And so this is

3:51

this is similar to uh what I what I

3:53

worked on a few years ago called Remo.

3:55

Um which was what was it? uh it was

3:58

recursive emergent memory optimization I

4:01

think is what it stood for. I I worked

4:03

on I worked on agentic memory systems

4:05

for a while. Um so for for those that

4:08

have been around for a while you

4:09

remember when I was working on that and

4:10

when I was working on things like Nala

4:12

and the ACE framework and I'll talk

4:13

about that in just a minute. So um the

4:17

where we're going with this is this was

4:19

very obviously the path forward. So over

4:22

the last couple years, I've made a few

4:24

videos talking about how uh just due to

4:27

efficiency, this is what the market

4:29

would demand because as soon as the AI

4:31

was capable of being fully autonomous,

4:34

people would build things that were

4:36

fully autonomous and then the market

4:37

would demand things that were fully

4:38

autonomous. So this is and I started

4:41

saying that because you know 2 three

4:43

years ago when open AI and and anthropic

4:46

and everyone else is saying humans are

4:48

going to be in the loop and it's just

4:49

going to empower humans and it's going

4:50

to be an empowerment tool and I called

4:52

BS on that because that's not how

4:55

technology works. You don't get to

4:57

decide what a technology does ahead of

5:00

time. there's emergent capabilities and

5:02

when when you when you take a step back

5:04

and you say we are clearly building

5:06

something that is a thinking engine and

5:09

what's the difference between being able

5:10

to carry out a task that you tell it to

5:13

carry out and autonomously carrying out

5:14

that task? Well, you just need something

5:16

to specify the task. So this is what

5:19

like that that's really what an agentic

5:21

framework is is you have one module or

5:24

one motor or one loop that's saying okay

5:26

what's the most important task to do

5:28

next? So that's one loop and then you

5:30

have the outer loop which is go do the

5:32

task and so then you have other services

5:35

that manage the memory and that sort of

5:36

thing. I remember when when I had built

5:38

a small team and we were working on the

5:40

ACE framework. Um one of the team

5:42

members implemented it in such a way

5:45

that it started doing things

5:46

autonomously and he's like he's he shut

5:48

that off. He's like well I didn't like I

5:50

didn't like that it was coming up with

5:51

its own ideas. I want it to only do what

5:53

I tell it to. I'm like I think you're

5:54

fundamentally missing the purpose of

5:57

autonomy. And he's like, "Well, no, but

5:59

we need user stories." And I So I was

6:00

like, "Okay, this this team clearly

6:02

doesn't get it." Um or that that person

6:04

on the team clearly didn't get it. Um so

6:06

I was like, "All right, you guys can do

6:08

whatever you want."

6:10

Anyways, so long story short, I've been

6:13

in the space for more than four years

6:15

now. I wrote uh natural language

6:16

cognitive architecture four years ago.

6:18

And so I'm really excited. Um, it took

6:21

longer. It took both less time and more

6:23

time than I would have hoped for this

6:25

kind of stuff to to be out there. Um,

6:27

and I'm glad that I was there to help

6:29

help it along. So, let's see. Um, so

6:33

this is this is my original work. So, I

6:35

I wrote natural language cognitive

6:37

architecture uh four more than four

6:39

years ago. Um, and this was basically

6:42

with GPT3. I realized that this was the

6:45

agentic framework. So um I'm not saying

6:48

that this is exactly how Claudebot works

6:50

because again this was a while ago and I

6:52

you can't predict everything but this

6:54

was the theory that I had. So you had

6:56

the interloop which would do a search

6:59

space uh create a kernel which is

7:01

basically what do I do then build a

7:03

dossier. So that's a task uh

7:05

specification and then load that into a

7:07

shared database. And this is pretty

7:09

close to how Cloudbot works. Um, where

7:11

there's like a tasks.mmarkdown file and

7:14

then you have um you have a few other

7:16

other shared things. And the the

7:19

innovation here and I was actually

7:21

working towards that was just put

7:22

everything in plain text. You don't need

7:24

a database just put it in plain text

7:26

because that's that's what the language

7:27

models read. And so everyone has settled

7:28

on do it in markdown. And then the outer

7:31

loop is actually the the task execution

7:33

loop. So this was my first idea uh where

7:35

you have you know the the shared

7:37

database. So that's kind of your uh

7:40

recursive language uh model. So so

7:43

that's your that's your context

7:44

management. That's your shared tasks and

7:46

that sort of thing. And so then you

7:48

build a context. Uh so by extracting

7:51

from that you build a corpus which is

7:53

basically recruiting all the information

7:54

that you need uh to execute the task.

7:57

Then you do the task and you output the

8:00

task into the environment. So that's

8:01

your API calls and that sort of thing.

8:03

And then that updates uh so you get

8:05

feedback. So you get the input

8:07

processing and output loop. That's the

8:08

outer loop. And then the inner loop is

8:10

the task manager. And so this was

8:13

actually pretty pretty salient. This is

8:15

pretty close to how how Cloudbot was

8:17

ultimately um implemented. So the next

8:20

uh layer that I was working on was the

8:22

ACE framework. So ACE framework stands

8:23

for autonomous cognitive entity and it's

8:26

a more sophisticated hierarchy. And

8:29

basically doing a side-by-side

8:31

comparison, Claudebot uh actually does

8:34

all but the aspirational layer. So to

8:37

provide a little bit of more context,

8:39

the entire theory of the ACE framework

8:41

was that you'd have um hierarchical

8:44

layers. So different processes that were

8:46

responsible for different aspects of you

8:49

know uh basically making stuff happen.

8:52

So you have the global strategy layer

8:54

which is the environmental context and

8:55

your long long time horizon planning.

8:58

The agent model which is basically a

9:00

list of what the what your agent can and

9:03

cannot do. Um as well as so like it has

9:05

to understand what it is. So like I

9:08

understand that I'm Claudebot and here

9:09

are my tools, here are my hands, here's

9:11

my memory, here's how I work. Uh because

9:13

and this was important at the time

9:15

because uh language models had a lot

9:17

less baked in in terms of what they were

9:19

and what they were capable of. So we had

9:21

to explicitly state you're a language

9:23

model this you're a part of the ACE

9:26

framework and that sort of thing and the

9:28

agent forge team has taken all this and

9:30

run with it. Um so they're still they're

9:32

still chugging away as far as I know. Um

9:34

then the executive function which is

9:36

risks resources and plans. Um as far as

9:38

I know the cloudbot focuses more on

9:40

plans and maybe resources. I don't know

9:43

if it has a risk control layer. Um but

9:45

that would be a really easy thing to

9:46

add. Uh next is cognitive control which

9:49

is task selection and task switching. So

9:51

cognitive control is about saying like

9:53

I'm failing at this task. I need to I

9:55

need to either cancel this task or try a

9:57

different method or that sort of thing.

9:59

Uh the inspiration for this layer was

10:00

actually frustration. Um the point the

10:03

the the neurocognitive point of

10:05

frustration is to tell you that what

10:06

you're doing isn't working. And so you

10:08

get frustrated enough you either quit or

10:11

you try harder or you try something

10:13

else. So that's basically what the

10:14

cognitive control layer does. And then

10:16

you finally have the task prosecution

10:18

layer which is actually executing a

10:20

specific task like call this API, make

10:23

this calculation, write that function,

10:25

that sort of thing. Now many many tasks

10:27

actually require um all of these layers.

10:30

So this was basically kind of what what

10:32

a lot of people in the project and and

10:34

people who were observing it at the

10:35

time, they said you're basically

10:36

describing the like an org chart of a

10:39

company. Um, and so some people actually

10:41

represented it as like like floors of an

10:44

office building. And so then you have

10:46

many many small agents taking on each

10:49

role in that floor and they all talk

10:51

with each other. And so then you have a

10:54

northbound bus and southbound bus. So

10:56

the northbound bus is basically feedback

10:59

from the environment. So this is the

11:01

green bar here is your interface with

11:04

the outside world. So that's APIs,

11:05

that's telemetry, that's anything that

11:07

you have uh control over um or get input

11:11

from from the outside world. So that

11:12

information needs to be disseminated to

11:14

all the agents and layers. And then the

11:16

southbound bus is command and control.

11:19

Now, one of the things that is missing

11:20

from Claudebot is an aspirational layer.

11:22

So this is this is one of the main

11:24

critiques that a lot of people have have

11:26

have created or or said is that Cloudbot

11:29

doesn't have like its own Supreme Court

11:31

to decide like you know does this does

11:33

this abide by our mission values or our

11:36

mission parameters or universal ethics

11:38

or that sort of thing. And so the

11:40

aspirational layer is about morality,

11:43

ethics and overall mission. So that's

11:45

that's this is very similar to what you

11:47

would say is like the constitution. So

11:49

like Claude's constitution serves as an

11:51

aspirational layer and this has been

11:53

really the c the centerpiece of my work

11:56

since I got into AI safety which is the

11:58

heristic imperatives. I'm really glad to

12:00

be talking about the heristic

12:01

imperatives again. So the heristic

12:03

imperatives are what what I came up with

12:06

after studying morality, ethics,

12:07

philosophy, game theory and that sort of

12:09

thing which is okay if you have a fully

12:12

autonomous uh machine what are the

12:15

highest values that you should give it

12:17

so that it stays aligned with humanity

12:20

and pro-life and that sort of thing. And

12:22

so that was um reduce suffering in the

12:25

universe uh increase prosperity in the

12:27

universe and increase understanding in

12:29

the universe. So I came to those three

12:32

values by figuring out like what are the

12:35

what are the deonttological

12:37

like like what what are the most

12:39

universal deonttological values. So

12:41

that's a duty based ethics which is

12:43

saying like from where I'm at today what

12:46

should I try and achieve and so many

12:49

people confuse like the paperclip

12:51

maximizer is an example of a teological

12:53

thing. It's like the the the best

12:55

version of the universe is the one with

12:57

with the most paper clips. So that is

12:59

purely a teological

13:01

version of morality or ethics or mission

13:04

or purpose. Whatever whatever um

13:07

whatever intervates something, whatever

13:09

gets you going uh whereas uh a more

13:12

deontological thing is from where I'm at

13:14

today, what do I have a duty to do? And

13:17

so like a duty to protect. So this is

13:19

this is where Asimov worked with the

13:22

three laws of robotics which is um you

13:24

know robot may not harm a human. So

13:25

under like whatever whatever the outcome

13:28

is, whatever the long-term outcome is,

13:30

do not do any actions that harm humans

13:32

and then do not do any actions that

13:33

allow harm to come to humans and that

13:35

sort of thing. And then of course later

13:36

in the foundation the zeroth principle

13:38

was actually your goal is to preserve

13:40

life. So ultimately something like

13:43

Claudebot will need an aspirational

13:45

layer and I would recommend the heristic

13:47

imperatives. So uh reduce suffering this

13:50

is a this is a very pro-social and

13:52

pro-life heristic which is basically

13:55

most intelligent animals will recognize

13:58

the suffering in other animals and try

14:00

and inter intervene. So you'll see this

14:02

where like um you know uh elephants and

14:06

and other animals if they see another if

14:08

they see another animal in distress

14:10

they'll try and help. Generally speaking

14:12

animals will try and help each other

14:13

because they recognize that that

14:14

distress that that suffering is bad. And

14:17

so I I said suffering specifically

14:20

because there's a difference between

14:21

pain. Pain is instructive. Pain is like

14:24

you need pain to understand what hurts

14:26

you. But suffering is non-addaptive.

14:28

Meaning suffering is just pain that has

14:31

no real purpose. So suffering is

14:34

generally bad. Um now you from an

14:36

artistic perspective some people say ah

14:38

suffering creates art which is you know

14:40

that that's a defensible assertion. You

14:42

look at Vincent Van Gogh and you know he

14:44

suffered a lot and he created great art

14:46

and that sort of thing. Um but that

14:48

doesn't necessarily mean that it is it

14:50

is teologically good. Um and also you're

14:53

never going to eliminate suffering. So

14:56

another thing is what I've established

14:57

is a vector. So reduce suffering not get

15:00

it to zero not eradicate suffering just

15:02

reduce it. Just just control suffering.

15:05

Next is prosperity. Because when you

15:08

have a heristic direction that says

15:09

reduce suffering um the best way to to

15:12

reduce suffering is to reduce life

15:14

because the less life there is the less

15:16

suffering there is. So then it took a

15:18

while to figure out the term prosperity.

15:21

So to counterbalance reduce suffering

15:23

you then have a value of increased

15:25

prosperity which is uh a very univers

15:29

universal word. It comes from Latin

15:30

prosperatoss which means to live well.

15:33

Literally the the root of prosperity

15:35

means you want to live well. You want to

15:38

flourish, you want to thrive. So you you

15:40

reduce suffering, you increase thriving,

15:43

you increase flourishing in the

15:45

universe. And you it's universal because

15:47

all life depends on all other life. Now

15:49

that's not universally true. There are

15:51

parasites. There are things there are

15:53

life forms that are basically just

15:55

harmful. Um but at the same time every

15:58

every life lives in the trophic level or

16:01

in the ecosystem and occupies a

16:03

particular niche. So then the final one

16:06

was because I realized well just those

16:09

two values you could end up with a green

16:11

earth that has no intelligence on it

16:14

that has nothing that no curiosity no

16:16

expansion and so then I realized that

16:18

the the core objective function of

16:20

humans that makes humans different is

16:23

that we are curious. So I said how do we

16:25

encode a curiosity

16:27

uh like algorithm into into an

16:29

autonomous machine whether it's AGI or

16:32

ASI or your personal cloudbot as it

16:34

turns out to be and that was to increase

16:36

understanding. So curiosity is the

16:40

desire to know for its own sake. So the

16:42

desire to understand for its own sake.

16:44

So rather than saying curiosity, which

16:46

is I just want to know things, right?

16:49

Because pure curiosity, unbridled

16:51

curiosity can lead you to do things

16:53

like, you know, section 31 and torturing

16:56

frogs just to see what happens. There

16:58

was actually an episode, this was also

16:59

inspired by an episode of Star Trek

17:01

where there was this galactic entity

17:05

that wanted to experiment on the crew of

17:06

the Enterprise just to see what would

17:08

happen. And so that was that was a an

17:11

ethical uh dilemma showing pure

17:14

curiosity for its own sake can actually

17:16

be destructive and it can be harmful and

17:18

cause suffering. So however you do want

17:21

to understand and that's one of the core

17:23

drives of humanity. It's like hey what's

17:26

over there? What's on the horizon? How

17:28

do I get across this body of water? Why

17:29

does fire happen? Our curiosity is what

17:32

sets us apart. So then I said okay we

17:35

have created this higher layer of

17:39

organization and how do we encode that

17:42

into a machine now plenty of people cuz

17:45

you know if if you followed me for a

17:46

while you know that heristic imperatives

17:48

people are talking about it people have

17:49

put it into agents I'm really excited

17:51

and the reason I'm making this video and

17:53

I didn't know that I was going to make

17:54

the video this way but the reason I'm

17:56

making this video is because I think we

17:57

have a really powerful opportunity to

18:00

now take that pri prior work you know

18:03

the the agent forge team worked on um

18:05

the ACE framework team worked on and

18:08

anyone who's tried to implement natural

18:09

language cognitive architecture the core

18:11

heristic imperatives would be I think

18:13

really great to add to something like

18:16

Cloudbot. Um so yeah, I guess that's

18:19

that's where I'll leave it. I think I

18:20

was working on um a notebook LM to

18:24

understand it, but it looks like it

18:25

failed. Um anyways, so I'll leave it

18:28

there today and um yeah, thanks for

18:30

watching. If I sound different, it's

18:32

because I'm practicing using my voice

18:34

differently. Um, I uh I mentioned on a

18:38

private video for my fans that um that

18:40

my voice usually gets really tired after

18:42

an hour or two. And that had me really

18:45

worried because I'm going to be

18:46

narrating my book. Um and and I was

18:50

like, I'm I don't know if I'm going to

18:51

be able to do this. And they said, oh,

18:53

like you need to practice using your

18:55

voice differently. And so I was doing a

18:57

little bit of research and I realized

18:58

that um I learned to use my voice in two

19:01

primary places. One was singing chorus

19:03

and then two was leading meetings. And

19:05

both of those require a tremendous

19:07

amount of energy and projection which

19:09

you don't need to do when the microphone

19:11

is literally a few inches from your

19:12

face. So if I sound weird, if I sound

19:14

different, it's because I'm practicing

19:16

using my voice differently so that I can

19:18

entrain the narrator voice. And that's

19:20

that's was the inspiration. I was like,

19:22

I need to make a video today. And

19:24

Claudebot is blowing up. So, thank you

19:26

for being here. Um, and I'll end by

19:28

plugging um all of the ways that you can

19:30

support me. Um, and a little bit of

19:33

news. So, we're very close to getting

19:35

the Kickstarter up for the post labor

19:37

economics book. Um, I actually do need

19:39

to do a little bit of recording for

19:41

that. I think um we've got a editor

19:44

chosen. So, this is an editor um who has

19:48

uh been the copy editor for like New

19:49

York Times bestseller. So, we found the

19:51

right guy. Um, the right genre, the

19:53

right the right talent, the right skill

19:54

set. So, the book is going to it's going

19:56

to really shine. So, there's that. And

20:00

then, um, the Kickstarter, the book, the

20:03

editor. Um, oh, yeah. And so, then, uh,

20:06

if you, if you're hanging around, what I

20:08

have what I have figured out for my fan

20:11

base is I've been trying to figure out

20:14

like the bonus content strategy for a

20:15

while. And what I have settled on, what

20:18

I'm trying right now is just fireside

20:20

chats. So just more unstructured

20:22

fireside chats. Um, and I'm posting

20:25

those on every platform that you can

20:27

subscribe on as a as a paid member. So

20:29

that's here on YouTube. That is Patreon,

20:32

that's sub uh Substack, that's Twitter,

20:34

um, and Spotify. So those are the five

20:36

platforms where you can sign up as a

20:38

paid subscriber and get those the bonus

20:40

content. And I'm going to talk about uh

20:43

it's going to be a little bit more

20:44

broader topic. So, some of it's going to

20:45

be personal updates, some of it's going

20:47

to be burnout, some of it's going to be

20:49

more philosophical. I don't know what

20:51

it's going to be, but it's basically

20:52

going to be if you want twice as much,

20:54

Dave, sign up for one of those platforms

20:57

and I will be posting um the insider

20:59

content there. And it's it's basically

21:01

the same format. It's going to be me

21:02

talking to the camera for 20 to 30

21:04

minutes. Um I'm also going to start as

21:06

my health improves, I'm going to start

21:07

hiking again, so you can get some of the

21:09

hiking videos and that sort of thing.

21:11

All right, with all that being said,

21:12

thank you for watching and I hope that

21:14

you guys enjoy and consider integrating

21:17

the heristic imperatives into Claudebot

21:19

and its successors and we'll go from

21:21

there. Cheers.

Interactive Summary

The video discusses Claudebot, an open-source, semi-autonomous personal agent that proactively finds tasks rather than waiting for commands. Unlike corporate-friendly agents, Claudebot is renegade and open-source, meaning users accept the risks. This open-source nature gives it an advantage in rapid development and iteration. The video also touches upon the technical concerns, such as security vulnerabilities due to its constantly running nature, but notes it can be run locally on various devices. The speaker then delves into the technological primitives that have enabled such agents, including models capable of agency, tool use (like APIs and JSON), and advanced memory management through recursive language models. The speaker shares their past work on agentic memory systems, including projects like Remo, Nala, and the ACE framework, highlighting that the current direction of AI development was predictable. The core concept of an agentic framework is explained as having loops for task specification and execution, with services for memory management. The speaker recounts an experience where a team member resisted autonomous behavior, missing the fundamental purpose of autonomy. The video then details the speaker's earlier work on a "natural language cognitive architecture" from over four years ago, which shares similarities with Claudebot's architecture, particularly in its use of plain text for task specification. The ACE framework is presented as a more sophisticated hierarchical model with layers for global strategy, agent modeling, executive function, cognitive control, and task execution. A missing component in Claudebot is identified as the aspirational layer, which deals with morality, ethics, and mission. The speaker introduces their "heuristic imperatives" (reduce suffering, increase prosperity, increase understanding) as a potential framework for this aspirational layer, explaining their philosophical underpinnings and differentiating them from purely teleological or deontological ethics. The video concludes with the speaker discussing personal reasons for practicing their voice for narration and plugging their upcoming book and paid content. They encourage integrating the heuristic imperatives into Claudebot and its successors.

Suggested questions

4 ready-made prompts