HomeVideos

Project Glasswing/Claude Mythos: Anthropic’s $x00 Million Marketing Stunt

Now Playing

Project Glasswing/Claude Mythos: Anthropic’s $x00 Million Marketing Stunt

Transcript

254 segments

0:00

So, the AI commentator community seems to have

0:03

lost its ever-loving mind. Again.

0:05

This time, it's about a new announcement about

0:07

Anthropic's new ""Project Glasswing"" and the

0:10

underlying

0:10

Claude Mythos model that they say they're

0:12

not releasing because they say it's too

0:14

powerful.

0:15

Some people are saying this is incredibly

0:17

dangerous and it's very good that Anthropic

0:19

is holding off on the release, and other people

0:21

are saying this is just a publicity stunt.

0:23

Here's the reality:

0:25

Both of those things are true.

0:26

Well, to an extent. Kind of. We'll get into it.

0:29

One thing is, the model itself is probably the

0:31

least relevant part of the whole thing.

0:34

And the publicity stunt part? We've been seeing

0:35

a lot of this particular attention grabbing

0:37

technique lately.

0:38

I'm sure it's only going to be getting worse as

0:40

the AI companies get more desperate to

0:42

keep up the investment dollars flowing so they

0:45

can be shoveled into the fire.

0:46

How are we supposed to believe this Shhh----?

0:55

This is the Internet of Bugs.

0:56

My name is Carl.

0:57

I've been a software professional since the 1980s and I'm

0:59

trying to do my part to make the Internet

1:00

a safer and less buggy place.

1:02

You can find links where you can get in touch

1:03

with me at InternetofBugs.com if you're so

1:04

inclined.

1:05

I'm not going to spend a ton of time going over

1:08

what Claude Mythos actually is.

1:09

I'll put some links in the description if you

1:11

need to catch up.

1:12

The short version is: Anthropic announced that

1:13

they have this new model they're calling Mythos,

1:16

but they say it's too dangerous to release it

1:18

to the public.

1:18

It is so good at finding and exploiting bugs,

1:21

they say, that they "believe it could reshape

1:24

cybersecurity" so they created something they're

1:26

calling "Project Glasswing" in conjunction

1:29

with a bunch of big tech firms quote: "in an

1:31

effort to secure the world's most critical

1:33

software" unquote.

1:34

Let me give you the short version of what's

1:36

going on and then I'll give you the more

1:37

details.

1:38

So there are three things at play here.

1:40

First off, Anthropic is doing that "We're

1:42

telling you that this thing is so great that

1:43

we're not letting you see it, we're only

1:45

getting access to it to people we want to

1:47

and you're going to have to take our word and

1:49

their word for how great it is."

1:51

The second thing that's going on is a pattern I'm

1:52

starting to see more often where AI companies

1:54

spend a ton of money on something that's

1:56

legitimately useful, then they attribute the

1:58

useful result

1:58

of all that spending to the functionality of

2:00

their software and not the amount of money

2:02

they actually spent on it.

2:03

And the third thing is they're making a big

2:05

deal about the one particular kind of task

2:06

they have a good way of a model doing, in the

2:08

hope that the press and the public will think

2:10

"Well, if that new AI is so good at X, then it

2:12

must also be good at all this other stuff,

2:15

right?"

2:16

So if you want a one- sentence takeaway it's

2:17

this: "The security risks the news are reporting

2:19

are real, but they're not because of how good

2:22

the new AI model is, so much as they're really

2:24

about how much money the company is spending to

2:26

show off this one specific scenario that

2:29

they've invented to make their AI look better

2:31

than it probably actually is."

2:32

All right, so let's get into some details.

2:34

Bragging about an AI product that you haven't

2:36

actually released is a pretty common pattern.

2:38

OpenAI let a select few people have early

2:40

access to ChatGPT-5, and many of them reported

2:43

how fantastic it was! And then they had to walk

2:45

it back after it was released and people

2:47

with no incentive to make OpenAI happy started

2:49

really testing it objectively, and it landed

2:51

with a thud.

2:52

A similar thing happened with DEVIN the so-called

2:55

"first AI software engineer" where they released

2:57

some demo videos and quotes from people that

2:59

they had handpicked to show it to, and then

3:01

the demos turned out to have been greatly

3:03

exaggerated. And Devin has just utterly failed

3:04

to have the impact that they claim.

3:06

When companies do this, you have every reason to

3:08

be skeptical.

3:09

For example, in this case, if it was really as

3:11

dangerous as they're saying and they really

3:13

were as concerned about the safety implications

3:16

as they say, then a responsible company - unlike

3:18

Anthropic - would just shut the hell up about it until

3:20

all the bugs the new AI had found had

3:22

been fixed. That's called "responsible disclosure".

3:25

It happens all the time. Where security experts

3:28

find the vulnerability, they notify the affected

3:30

company about the problem, and then they wait

3:32

for the fix to be released before they announce

3:34

their involvement.

3:35

So, next item up is how AI companies have

3:37

started trying to get you to confuse effort

3:39

and money for model ability.

3:42

I made a video recently about how OpenAI had

3:44

run a custom-built internal model for hours

3:47

to simplify some math equations, got some

3:49

academics try to paper about it, and then wrote

3:51

a press

3:51

release that made it sound like ChatGPT was a

3:53

PhD physicist and had discovered new science.

3:56

This is the same kind of thing. Some security

3:59

folks over at aisle.com did some great research

4:02

and wrote up a great piece about it sub titled

4:05

"why the moat is the system and not the model."

4:08

You should check that out if you want more

4:09

detail - I've linked it below. They took some

4:11

of the more celebrated bugs that Anthropic says

4:13

the new model found and they tested them

4:15

on small, cheap, open-weight models, and got very

4:17

similar results. That's a very strong evidence

4:20

that the new model isn't the real story here. It's

4:22

not that spectacular. What's important

4:24

is the amount of time and money that they spent.

4:26

So here's an excerpt from an Anthropic blog

4:28

post about the denial of service bug they found

4:30

in OpenBSD which is a very secure operating

4:32

system that I've been running on my own servers

4:34

since the 1990s.

4:35

They say that the bug was found after a

4:37

thousand runs at a cost of around twenty

4:39

thousand dollars

4:40

likewise they spent ten thousand dollars to

4:42

find a bug in FFmpeg, and they say they

4:44

found a few thousand other bugs. Now they don't

4:46

tell us the total amount of money they spent

4:48

on the computer time for the project, but if we

4:50

were to extrapolate from the OpenBSD

4:52

and FFmpeg bug costs times a few thousand

4:54

for the other bugs, we could easily be at tens

4:57

or hundreds of millions of dollars. And consider

4:59

too that they almost certainly spent a lot of

5:01

money on computer cycles looking at software

5:03

that they didn't have any bugs that they found.

5:06

They also said they were gifting a hundred

5:07

million dollars of compute to their partner

5:09

companies in "Project Glasswing", and four

5:11

million dollars in grants to open source groups.

5:14

Let

5:14

me put that in perspective for you: The biggest

5:16

bug hunting program in the world, HackerOne,

5:18

spends something like eighty to ninety million

5:20

dollars a year total. "Project Glasswing" is

5:23

spending

5:23

125% of that just

5:25

in compute for their partner companies. They

5:27

could easily have spent an additional multiple

5:29

of that searching for the bugs that were

5:30

announced

5:30

in this press release, to say nothing of however

5:32

much was spent training in the model on

5:34

existing

5:35

bugs while it was being built in the first

5:37

place. Anyone who spends that much money looking

5:39

for bugs is going to find a ton of them, and we

5:41

should all be happy that they were willing

5:42

to spend that, and the end result of this will

5:44

be a much safer Internet for everyone, and

5:46

that's fantastic. And Anthropic should absolutely

5:49

be commended for that. But they're not saying

5:51

"we spent a ton of money making Internet safer

5:53

for you" they're saying "look how dangerous

5:55

and powerful our new model is - you should be in

5:58

AWE of it" and that's just not the reality

5:59

here. Lastly, they're glossing over that there is

6:03

a big difference between discrete tasks

6:05

with well-defined success criteria and the

6:07

ambiguities that we as humans deal with all

6:09

the time. It's very straightforward to set up an

6:12

environment with a checklist of steps

6:14

to use to see if a bug has been found, and then

6:16

let the AI run in that environment for

6:18

hours days or even weeks, until you either find

6:20

something or you decide you've spent enough

6:22

money and it's time to look elsewhere. But this

6:24

is much more like the way that an AI

6:25

is taught to be good at chess than it's

6:27

training for human equivalent general

6:29

intelligence.

6:30

And those are two different things, and getting

6:31

better at one doesn't help you with getting

6:33

better at the other. The AI Security Institute

6:36

in the UK has access to the model and they

6:38

ran it against a number of their "Capture the

6:39

Flag" scenarios that they use in hacking

6:41

competitions. It did pretty well - better than any

6:44

other AI they've tested, but keep in mind

6:46

that doing hacking Capture the Flag problems is

6:48

working with clear rules and well-defined

6:51

victory conditions. These are more like playing

6:53

chess or go than general software building.

6:55

So it's vastly easier to train an AI to do this.

6:58

So, what does that mean for you and for

6:59

the Internet? Well it means that, as has

7:01

happened several times before, when new bug

7:03

hunting

7:03

tools were released or new bug hunting

7:05

techniques were published, we're going to be

7:06

going through

7:07

a bumpy time for a while from a security

7:09

standpoint. Expect a lot of security updates for

7:11

your

7:11

phones, tablets and laptops, a lot of news

7:13

reports about hackers for a while, and once

7:15

we get through this remediation process for

7:17

everything that this initiative is discovering

7:19

we'll all hopefully be in a better place and

7:22

things will calm down. It also means that

7:24

- finally -

7:25

AI has given us something that might help

7:27

counteract all the nightmare security holes

7:29

that vibe coding is creating. Now I doubt that

7:31

most vibe coders are going to spend the time

7:33

and effort using this kind of technique to look

7:35

for the exploits in their code, but they

7:37

could, and that's more than we had before. It's

7:39

not a silver bullet though. Finding and

7:42

preventing

7:42

the customer-facing network and OS-related

7:44

kinds of things that Capture the Flag games

7:46

prioritize, like buffer overruns, or remote access,

7:48

privilege escalation, that kind of stuff - that's

7:51

great, but there are a lot of other kinds of

7:53

bugs that aren't so easy to look for, like

7:55

logic errors, data loss, data corruption,

7:57

synchronization errors, half-committed

7:59

transactions, roach

8:00

motel UIs, whack-a-mole, and that kind of

8:02

stuff. We're still a very long way from

8:04

a "true AI software engineer". So, don't panic, don't

8:07

let the news stress you out, but do try

8:10

to be extra vigilant for a while. Thanks for

8:12

watching. Let's be careful out there.

Interactive Summary

This video analyzes Anthropic's announcement regarding their new 'Mythos' model and 'Project Glasswing.' The speaker argues that while the security research initiative is beneficial, the company is exaggerating the model's inherent intelligence. Instead, the results are largely driven by massive financial and computational investment, similar to patterns seen with other AI hype. The video concludes that while this will lead to a more secure internet, it does not signify the arrival of general software engineering AI.

Suggested questions

3 ready-made prompts