HomeVideos

Azure Update 3rd April 2026

Now Playing

Azure Update 3rd April 2026

Transcript

364 segments

0:00

Hey everyone, welcome to this week's

0:02

Azure update. It's the 3rd of April. As

0:04

always, we have the chapters so you can

0:07

jump to any particular update you care

0:09

about the most.

0:11

New videos this week. So obviously you

0:13

should go to the movies and see Project

0:15

Hail Mary. It was phenomenal. The book

0:18

was great. The movie was great. So it's

0:19

not one of my videos. My videos

0:22

I dove into Work IQ.

0:24

So when we think about how we work, not

0:28

just the knowledge, the artifacts in our

0:31

documents, our chats, and our meetings,

0:34

but a personalization that learns how

0:37

artifacts relate to each other,

0:39

how people relate to them, how people

0:42

relate to people,

0:43

uh

0:44

how we talk, how we function, our rhythm

0:47

of business,

0:49

and then special inferencing engines and

0:52

skills around document creation and deep

0:54

knowledge retrieval, and creating really

0:57

good PowerPoints. That's all part of

1:00

Work IQ that obviously

1:01

the Microsoft Copilots they're grounded

1:04

on, Co-work is grounded on, but you can

1:07

ground your own agents.

1:09

And then what is Copilot Co-work?

1:12

And I don't often say this, but it's a

1:15

really, really cool demo. About 8

1:16

minutes in if you want to skip the video

1:18

but go and look at the demo

1:20

to show how the Co-work capability

1:24

can just be given some outcomes you

1:26

want. I can interrupt it. It's running

1:28

as a cloud agent, but it's using these

1:31

long-running deep reasoning models to go

1:34

and do a bunch of cool stuff including

1:36

creating a web app all with a single

1:38

prompt. So it's kind of cool.

1:41

And then I kind of finished a flow of

1:44

videos I was trying to create around AI

1:47

and the Microsoft AI ecosystem. So I

1:49

just created sample.ai. It's just my

1:51

YouTube stuff. It's about 15, 16 videos,

1:54

but it's a very curated path if that's

1:57

useful trying to learn. Very much the

1:59

Microsoft focused

2:02

way of leveraging AI.

2:04

All right, so onto what's new on the

2:05

computer side.

2:07

So the Azure Red Hat OpenShift offering.

2:10

So that's the jointly created, managed,

2:12

supported offering between Red Hat and

2:15

Microsoft. Well, this is now available

2:17

in Indonesia Central. And so this is

2:20

really useful to have in new regions

2:22

when I think about I need to run a

2:24

workload closer to the customers maybe

2:26

from a latency perspective, but it may

2:28

also be around meeting different

2:29

regulatory requirements.

2:32

And this is kind of an interesting one.

2:34

So for virtual machines and virtual

2:35

machine scale sets, there's now a full

2:38

caching capability for ephemeral OS

2:41

disks. And this is in preview. So

2:43

remember the ephemeral OS disk is where

2:45

it's not using a durable managed disk.

2:49

It's that it's going to create the the

2:51

disk on local host resources. The the

2:54

temp space, the cache space.

2:56

So what that means is ordinarily

2:59

the rights we make to that ephemeral OS

3:01

disk go to the local storage in the node

3:04

that the VM runs on,

3:06

but the actual main OS image it's

3:08

reading from is still remote.

3:11

So what this new feature does is it

3:12

caches the entire OS disk to the local

3:16

storage. That removes any remote storage

3:18

dependency. So that's going to obviously

3:20

increase the resiliency of that running

3:22

VM from any remote storage failures, but

3:25

also improve the performance with the

3:28

super, super low latencies because it's

3:30

just local running on that storage. Now

3:32

the way it's going to work is the

3:33

caching occurs in the background once

3:36

the VM boots. So obviously there's a bit

3:38

of time for it's going to cache that.

3:40

And it's just an enable full caching

3:42

flag that I set for the ephemeral OS

3:44

disk.

3:45

And remember why we use ephemeral OS

3:48

disks. It's where there's nothing in the

3:51

OS disk state that we care about. It's a

3:54

tin soldier. Something's wrong with it,

3:56

we just rebuild another one in its

3:57

place. It's really common with virtual

3:59

machine scale sets where we constantly

4:01

create and delete the virtual machines.

4:03

And so I don't want to pay for a managed

4:05

disk. I want better performance.

4:08

Ephemeral OS disks are fantastic for

4:10

that.

4:12

Networking.

4:14

So App Configuration now has an Azure

4:17

Front Door integration in preview.

4:19

Remember App Configuration enables us to

4:21

have all of those different

4:22

configurations for the applications to

4:23

be centrally managed,

4:25

and then it can

4:27

be used and delivered to client

4:29

applications.

4:31

Well, with the integration with Azure

4:32

Front Door, which remember Azure Front

4:33

Door is that global layer 7 anycast

4:38

split TCP

4:40

delivery layer with caching,

4:42

well, when it integrates with that,

4:45

it now enables the scaling of the

4:47

delivery of those configurations to be

4:49

in terms of millions of clients. And I

4:52

don't have to try and develop my own

4:54

proxy layer anymore. And obviously this

4:55

can work with single page apps, so the

4:57

spa, the mobile apps, a whole bunch of

5:00

other different scenarios. And what's

5:01

going to happen here is

5:03

the App Configuration endpoint is

5:05

established on Azure Front Door,

5:08

and then the Azure App Configuration

5:10

store is set as its origin.

5:12

So you use managed identity to secure

5:14

that communication,

5:16

and then Azure Front Door will simply

5:17

retrieve the selected key values,

5:19

feature flags, um whatever is required.

5:23

It can then cache and respond it. So

5:24

it's it's avoiding you having to worry

5:26

about secrets and other things, but

5:29

gives you a really high scalable secure

5:32

service.

5:33

On the storage side,

5:35

so premium SSD V2 is in a new region in

5:39

GA. Remember

5:41

premium SSD V2 is all about it's a

5:43

sub-millisecond latency,

5:45

but you have separate dials for the

5:47

IOPS, the throughput, and the capacity.

5:49

And I can dynamically change the IOPS

5:51

and throughput while it's being used. So

5:53

now we see these in South India and US

5:56

Gov Arizona.

5:58

So when I think about scenarios where

6:01

I need high IOPS, I need high latency

6:03

databases, big data, analytics, gaming,

6:07

um

6:08

this is really useful and I would use it

6:10

in VMs, but I can also use it in

6:12

containers where I need some kind of

6:14

durable state.

6:17

Uh the user delegation shared access

6:20

signature for tables, files, and queues

6:21

is now GA. So that was already available

6:23

for blob,

6:25

but now it's been pulled to the other

6:26

storage services. And the whole big deal

6:28

here is user delegation SAS is more

6:31

secure than the regular account or

6:33

service SAS because both of those are

6:36

tied to the master storage account key,

6:39

whereas this is tied to an Entra ID

6:42

identity. So whereas the storage account

6:44

key has all powerful, it can do

6:47

anything, this would be a subset of the

6:49

permissions of the identity creating it.

6:52

Can never be more.

6:54

And it's only valid for up to 7 days

6:55

maximum. So there's a lot more control.

6:57

It's a a better secured method.

7:01

Um Azure Data Box can now be imported

7:05

into Azure Files uh provision V2 storage

7:08

accounts. So the whole goal, remember

7:10

Azure Data Box is hey, I need to move a

7:13

massive amount of data.

7:15

Very often it's how I'm migrating from

7:17

on-prem to the cloud. I don't want to do

7:19

it over the network, but it can work

7:21

from the cloud back to on-prem as well.

7:23

And so I'm doing a big data migration.

7:26

Well, now as part of that migration, I

7:28

can use a provision V2 account. The big

7:31

thing about provision V2 was like the

7:33

dynamic

7:34

provision disk V2, the capacity, the

7:37

IOPS, and the throughput can all be

7:39

separately set.

7:41

And so now, hey, I can use those

7:44

more fidelity type storage accounts as

7:48

part of Data Box.

7:50

Uh Azure NetApp Files cool access has

7:53

some enhancements in preview. So the

7:55

whole goal of the cool access is the

7:57

less used data can be moved from the

8:00

Azure NetApp Files storage, which is

8:02

considered the hot tier, to regular

8:04

Azure storage, which is considered the

8:05

cool tier, to help drive cost savings.

8:08

So with the premium and the ultra

8:10

service levels of Azure NetApp Files,

8:14

they've enhanced the quality of service

8:15

algorithms that drive the allocation of

8:18

throughput to really try and minimize

8:20

any performance impact you see as a

8:22

result of that cooling.

8:27

On the database side, so Cosmos DB for

8:29

PostgreSQL is being retired

8:32

uh end of March 2029, so 3 years' time.

8:36

The writing was on the wall for this.

8:38

It's been replaced by PostgreSQL Elastic

8:40

Cluster. All right. PostgreSQL Elastic

8:42

Cluster is

8:44

built on the same Citus extensions

8:48

that we see for that distributed

8:50

sharding up of PostgreSQL databases,

8:53

but the Elastic Cluster has the built-in

8:55

HA, backups, DR, all the future

8:57

engineering investments. So just make

8:59

sure you migrate. I mean, honestly, you

9:01

want to migrate as soon as possible, but

9:03

definitely you need to migrate um

9:06

before then. And there is migration

9:08

tooling available for you to achieve

9:10

this.

9:12

Uh Event Grid has a number of updates.

9:14

So remember the whole point of Event

9:15

Grid is I can build event-driven

9:17

solutions at massive scale.

9:21

And I as the app that wants to trigger

9:24

and do something off of the event don't

9:26

have to hammer poll. I I don't have to

9:28

constantly ask the source, do you have

9:30

something? Do you have something? Do you

9:31

have something?

9:33

Event Grid takes care of that and then

9:34

calls that often serverless

9:38

handler that's going to do something

9:40

with the event data.

9:42

So what I've done is a whole bunch of

9:44

MQTT enhancements. So there's things

9:46

like in-order message delivery within a

9:48

client session.

9:49

It's now has the ability for one

9:52

connection attempt per second per client

9:53

connection limiting. So I'm going to

9:55

throttle the number of requests. It's up

9:57

to 15 MQTT topic segments and there's

10:00

also cross-tenant delivery or when

10:02

they're available in GA. In preview,

10:05

there's MQTT OAuth 2.0 for the

10:09

authentication. There's custom webhook

10:11

authentication and there's static client

10:13

ID identifiers. Now it can also now use

10:17

managed identities for webhooks. There's

10:19

a cross-tenant webhook delivery and it

10:21

also supports network security

10:22

perimeters now. Remember network

10:24

security perimeters are about, hey, I

10:26

have a bunch of different Azure PaaS

10:28

services. If I put them in the same

10:30

network security perimeter, they're

10:32

allowed to all talk to each other.

10:34

And then I can also control

10:37

communication

10:39

for inbound and outbound to them as a

10:41

group of services. So it's a really nice

10:43

set of rule controls I can do there.

10:48

So Copilot Co-work is now available in

10:51

front here. I did a whole separate video

10:52

on this. It is crazy, crazy good. And

10:55

again, I don't often push to go and

10:57

watch one of my videos, but go and watch

10:59

the Co-work video. Even if you skip me

11:01

waffling on, go to like minute eight and

11:04

it's a demo.

11:05

It runs in the cloud, so it has no

11:08

use of your local machine. It's not

11:09

using local resources. It doesn't have

11:11

full reign to your local machine.

11:13

I can interact with it while it's doing

11:16

its work. It's grounded in Work IQ and

11:19

Outlook and Teams, all these fantastic

11:21

data sources, and I just tell it the

11:24

outcomes I want.

11:25

And it goes and works out

11:27

all of the plan and then just does it.

11:30

It It is a game-changer. I've done some

11:32

really cool things already just in this

11:34

week.

11:36

But uh yeah, go and go and watch the

11:38

demo. It's crazy good.

11:40

Uh Azure Azure Speech Neural HD 2.5. So

11:44

this is all about giving more choices

11:47

um in the regions you use it, the

11:49

quality, the performance, the

11:51

expressiveness.

11:52

Where I want a really low latency, think

11:54

real-time type interaction.

11:57

There's a whole number of different

11:59

speaking style updates for English

12:01

content.

12:02

Uh I I know it's all things like

12:04

struggling and skeptical styles. I can

12:07

do things like sighing and yawning,

12:09

which again will be important if I'm

12:10

trying to do a voice about my content.

12:13

But uh just a lot of work about trying

12:16

to have those um synthetic voices.

12:19

And then I'm only mentioning this cuz I

12:20

thought the core was really named. There

12:22

There's constantly new models being

12:24

added to foundry. That's one of the

12:26

whole points around I think the

12:28

Microsoft strategy in general is model

12:30

choice. There's no such thing as the

12:31

best model. There might be the best

12:34

model this week for this type of

12:35

requirement,

12:37

but

12:38

nearly every scenario is using multiple

12:40

models and they're going to evolve over

12:41

time. But I love the name of this thing.

12:44

This is a new Nvidia model, NeMo Triton

12:46

3 super 120B

12:49

uh A12B.

12:50

So it's a mixture of experts. So the

12:52

whole point of a mixture of experts is

12:54

there's 120 billion parameters,

12:56

but

12:58

it only activates 10% of them, so 12

13:01

billion for any specific inference. So

13:05

it's actually fairly compact in its

13:08

resource utilization, but based on what

13:11

that inference request is, it can choose

13:14

which expert it has that is the most

13:17

applicable to what it's been asked to

13:18

do. It has a 1 million token context and

13:21

it's really geared towards token

13:23

text generation. Obviously I could then

13:25

pass that to a text-to-speech model. I

13:27

could have a speech-to-text model in

13:29

front of it, but its specific goal is

13:33

around that um text generation model. Um

13:35

but super, super powerful.

13:37

And that is it. As always, I hope this

13:39

is useful.

13:40

Uh amazing, amazing, amazing for all the

13:42

updates this week. And till next video,

13:44

take care.

Interactive Summary

Loading summary...