HomeVideos

Azure Update 27th March 2026

Now Playing

Azure Update 27th March 2026

Transcript

364 segments

0:00

Hey everyone, welcome to this week's

0:02

Azure update. It's the 27th of March. As

0:04

always, we have the chapters so you can

0:07

jump to any particular update you care

0:09

about the most. New videos this week.

0:13

So, I dived into, hey, I want to build

0:16

an agent. Should I use agent builder?

0:20

Should I use copilot studio? Should I

0:22

use Microsoft Foundry? So, I really went

0:25

through what are some of the personas

0:27

that would use the different options and

0:29

what are the capability differences that

0:31

would drive me to use one over another.

0:34

We also had our 400,000 subscriber ask

0:37

me anything session. So, the recording

0:39

is now available to see that. And then I

0:41

did a video on the new entra backup and

0:45

recovery. So, I get these snapshots

0:46

taken daily. There are five of them. So

0:49

I can go back and restore the state of

0:52

objects to a previous point in time. If

0:55

I then combine that with things like

0:56

soft delete and protected actions, I get

0:59

that complete ability to mitigate

1:02

accidental or malicious changes to my

1:06

entra environment. So went through that.

1:10

Okay. On to what's new on the compute

1:12

side. So Azure Kubernetes service now

1:16

has application network in preview. So

1:19

think of this as a new application layer

1:22

abstraction

1:24

for all of the Kubernetes traffic. So it

1:27

enables me to control that service to

1:30

service sets of communications. It gets

1:33

me better observability

1:35

without having to inject a side car into

1:39

each of the various pods. So you can

1:42

think of it as providing mesh

1:44

capabilities

1:46

without all of that mesh overhead and

1:48

management I would have to do. And one

1:50

of the interesting things it's doing and

1:52

we're starting to hear more about this

1:53

is it's using Spiffy

1:56

as a method to identify the various

2:00

parties and then using that as part of

2:03

the network tracking the network

2:05

control.

2:06

This is Linux only today, but the great

2:09

thing is because of this abstraction

2:13

layer, there's no app changes to take

2:15

advantage of it.

2:18

Um, the AKS meshlessto

2:21

app routing is in preview. So basically,

2:24

I can now use the Kubernetes gateway

2:26

APIs for ingress management again

2:29

without having to leverage a sidecar

2:31

architecture.

2:33

and the AKS network logs have gone GA.

2:36

So container network logs provide a

2:38

capture of the network flow metadata. So

2:41

it's not all of the packet data but it's

2:43

the metadata. So IP addresses, ports,

2:46

namespace, the pods, the services, the

2:48

flow direction, um verdicts from

2:50

policies and a bunch of other stuff at

2:53

layer 3, four and seven. Now this can be

2:56

stored as

2:58

writing to local storage and then

3:01

optionally to a log analytics workspace.

3:04

And when I do that right, I can filter

3:06

it to only specific resources of

3:09

interest. And there's also an on demand

3:11

mode where I would use Hubble as part of

3:14

that capturing. So it's all like getting

3:15

really good insight into the traffic.

3:19

There are now manage GPU metrics in

3:22

preview. So if I'm using nodes with

3:25

GPUs, I can see performance and

3:27

utilization data from where I'm using

3:31

Nvidia GPUs. And so this is going to

3:34

hook into manage Prometheus and

3:35

Graphfana. So things like GPU

3:37

utilization, memory utilization,

3:39

streaming, multiprocessor efficiency.

3:42

There are metrics around temperature,

3:44

power, bandwidth, frequency,

3:45

reliability. So basically giving me

3:47

really good insight into the GPU

3:49

characteristics of the nodes.

3:53

AKS fleet manager now has crosscluster

3:55

networking in preview. So if I have

3:58

applications that span AKS clusters,

4:02

what it's actually going to provide is a

4:03

managed psyllium cluster mesh. But

4:07

because it's managed, it's going to make

4:08

it really easy to configure and

4:10

obviously manage the thing. So what it's

4:12

going to do is once I enable this any

4:14

published service from anything within a

4:17

particular cluster can then be used by

4:19

any connected cluster as if it was local

4:21

to the cluster. Um all I have to do is

4:25

for the services there's a global

4:27

annotation I'm going to mark it as but

4:30

then hey I I really get this easy

4:32

crosscluster network communication and I

4:35

also am going to get global

4:36

observability because I'm going to get

4:38

shared metrics and flow logs across all

4:41

those clusters that are part of that

4:44

crosscluster networking.

4:47

Um AKS uh container network metrics have

4:51

filtering in GA. So the normal metrics

4:54

that are part of container network

4:55

observability can generate a massive

4:58

volume of data. And filtering allows me

5:01

to control what data is captured. So I'm

5:04

only going to get the signals I really

5:06

want, which helps reduce storage, which

5:08

helps reduce cost, which also reduces

5:10

the noise when I'm trying to actually

5:12

understand what's going on.

5:15

There is an AKS network AI agent

5:18

available in preview. So what it enables

5:22

me to do is interact using natural

5:24

language. I can give it a problem

5:25

description and it will turn it into

5:27

diagnostics information from all of the

5:31

various data that's captured. So it's

5:32

going to make the cluster

5:33

troubleshooting much easier.

5:38

AKS now has a blue green agent pool

5:41

upgrade option available in preview. So

5:44

when I think of safe deployment

5:46

practices, one approach is kind of we

5:48

have these rings of deployment. We do

5:51

kind of a rolling upgrade approach which

5:52

is the traditional approach. But what

5:55

this lets me now do is I can have a

5:57

parallel node pool that runs the new

5:59

configuration

6:01

and then as I make a change that new

6:04

node pool has the new configuration. I

6:06

can split traffic to start hey seeing is

6:09

it functioning as I would anticipate. If

6:12

all's good I move all the traffic over

6:14

to it and then obviously I can delete

6:16

the old node pool. There's a problem I

6:18

can roll back to the current node pool

6:21

and I can use this for Kubernetes

6:23

upgrades, node image upgrades, config

6:26

changes.

6:28

It does mean you've got double the

6:29

number of resources during the upgrade.

6:31

They're not always existing. When I want

6:33

to use the blue green, it will go and

6:34

create the new node pool with a new

6:36

config.

6:38

During the time of that blue green, I've

6:40

got double the resources, so double the

6:42

cost, double the quotota use. You need

6:44

to make sure you have the quotota, but

6:45

then obviously it gets deleted. So, it's

6:47

only double during the period of the

6:49

upgrade.

6:52

Um, Arc enabled Kubernetes now has

6:55

oneclick enablement of recommended

6:58

Prometheus alerts based off of community

7:00

rules. So that's going to give you

7:02

really good coverage of the cluster, the

7:04

nodes, the pods, and I had this

7:07

previously, but I had to do a template

7:08

based deployment. So this makes it much

7:10

much easier.

7:13

Um, Azure container storage now has

7:17

elastic sand integration in G. So

7:20

remember, Azure container storage is all

7:21

about providing

7:23

very high quality storage for my AKS

7:26

workloads. Previously it was GA for

7:30

local node NVMe storage.

7:33

Now in addition to that I support

7:35

elastic sand which gives me more durable

7:38

storage, more flexible pools for various

7:41

different scenarios of different levels

7:43

of performance. So now I get a a greater

7:45

choice

7:47

database.

7:49

So SQL database now has this automatic

7:52

index compaction in preview. So this is

7:54

SQL DB, SQL MI and SQL in fabric. It is

7:58

a background automatic index compaction.

8:02

So I'm automatically going to reduce the

8:03

amount of storage space I use, therefore

8:05

the cost and also I'm going to get

8:07

improved performance because it's going

8:08

to use less CPU, memory, and dis IO. So

8:11

this removes the need for me to have

8:13

scheduled index jobs and I just enable

8:16

it with a single command.

8:20

um SQL managed instance now has change

8:22

event streaming in preview. So any rowle

8:25

change so an insert an update delete can

8:27

now stream to an event hub with this

8:30

change event streaming and it's

8:32

basically in near real time then

8:34

obviously from event hub I can trigger

8:36

various serverless things to work off of

8:38

that. So it's going to let me build an

8:39

event driven solution use real-time

8:42

analytics and more without having to do

8:45

anything specific in my code.

8:48

Uh, SQL Server has soft delete available

8:50

in preview. So, hey, I can set a soft

8:53

delete retention. So, I can self store

8:55

SQL servers in the event of a deletion.

8:59

Uh, SQL hypers scale has some new SKUs

9:02

in preview. Remember hypers scale um

9:05

enables me to scale to much higher

9:07

performance and capacity because it

9:09

separates the compute from the page

9:11

servers. So, there are new 160 192 vcore

9:15

options for premium series hardware.

9:18

So that gives me a much larger compute,

9:20

much larger memory configuration where I

9:22

have those really really demanding

9:24

workloads. So if I think large scale

9:25

OLTP, HTAP analytics heavy workloads and

9:29

I can use this for both single database

9:31

and elastic pool.

9:34

There have been some disknments

9:37

in preview um across SQL database. So

9:43

vector databases are huge today. When we

9:45

think of generative AI, it's natural

9:47

language interactions. We often want

9:49

these vector databases that store

9:51

embeddings in these high dimensions that

9:55

represent the semantic meaning of data

9:57

and then I go and search for hey I'm

9:59

looking for something. I turn that into

10:00

an embedding and I find the closest

10:02

match. So disk an ANN is a Microsoft

10:05

research created vector search

10:06

capability that is part of SQL database

10:09

part of SQL database in fabric and it's

10:12

been improved so that the tables are no

10:14

longer read only after that index

10:17

creation. There are filters applied

10:19

during vector searches and not after. So

10:22

it's going to be a lot more performant

10:24

use less resource. There's also

10:26

improvements between choosing between

10:28

disk A&N and the regular uh the K

10:31

nearest neighbor algorithms along with

10:34

some other optimization. So basically

10:35

just improving uh those all up

10:38

capabilities related to the vectors.

10:41

Uh Azure monitor OLTP ingestion is in

10:44

preview. So I can bring in open

10:46

telemetry data, metrics, logs, traces

10:50

directly into an Azure monitor workspace

10:52

because it has a native open telemetry

10:55

protocol supported endpoint. Um uses

10:57

entra for the authentication

11:00

and then Postgress SQL has custom time

11:03

zone for the chrom the scheduled jobs.

11:06

So now I can set a time zone to be used

11:08

for those scheduled jobs which is really

11:11

useful to ensure jobs happen based on a

11:13

desired regional time zone. So hey I

11:15

want to make sure this doesn't happen

11:16

during business hours of the place using

11:19

it instead of trying to work around well

11:21

what is the default based on the server

11:26

post SQL has migration updates in G. So

11:29

I can now migrate from EDB Postgress SQL

11:34

and Google alloy alloy DB uh to Azure

11:38

manage Postgress SQL and I can also use

11:40

PG output now for minimal downtime

11:43

online migrations

11:46

and then Microsoft fabric now supports

11:49

my SQL mirroring. So I have my Azure

11:51

database uh for my SQL flexible. It can

11:55

then mirror without me having to create

11:57

data pipelines or anything else into

12:00

fabrics one lake in basically near real

12:03

time. So it makes it immediately

12:05

available for any of the fabric

12:06

workloads like analytics AI uh PowerBI

12:10

you you kind of name it

12:13

and then fabric cosmos DB private

12:15

endpoint enabled databases mirroring has

12:18

gone GA. So I have a Cosmos DB database.

12:21

It's using private endpoints. I can now

12:24

enable the mirroring of it to Microsoft

12:27

Fabric. There's some additional

12:29

networking I have to add during the

12:31

establishment of the mirror. But once

12:33

it's established, I can remove it again.

12:35

So I I'm reducing the connectivity for

12:38

my Cosmos DB to only be those private

12:40

endpoints.

12:42

on the uh miscellaneous side.

12:46

So foundry priority processing has gone

12:48

GA. So there are certain situations

12:51

where the latency for inferencing is

12:54

critical to the AI app agents

12:56

performance. Now one thing we've had in

12:59

the past and still do is to use

13:01

provisioned throughput units PTUs. So a

13:05

guaranteed amount of throughput

13:07

which is a set amount that I provision

13:10

and pay for or it's set in advance

13:13

instead of the regular pay as you go

13:14

usage.

13:16

Well, priority processing gives high

13:20

speed performance on a pay as you go

13:22

basis. So maybe I don't know the exact

13:24

amount I need or maybe I've got a PTU

13:27

but I need some additional at certain

13:29

times. So this lets me get higher

13:32

priority processing. So lower latency,

13:35

higher throughput when I have that time

13:38

critical inferencing need, but I'm not

13:40

doing that commitment to that amount of

13:42

throughput in advance. Now obviously I'm

13:44

going to pay a price premium for this.

13:46

There is a price premium over the

13:48

standard tier pricing. It varies by

13:50

model. Um but it is available for the

13:53

latest models for global and data zone.

13:56

And obviously it's not I only use one of

13:59

them. I could combine this with pay as

14:02

you go with standard I pay as you go

14:05

with PTU with batch to work out what is

14:08

the right solution for what I need.

14:12

Oh, I went backwards somehow. Didn't

14:14

even notice that. Uh, Entra ID external

14:18

MFA has gone G. So that lets me use an

14:20

external MFA solution

14:23

right supports open ID connect as part

14:25

of the entra ID authentication. So that

14:28

includes using conditional access and it

14:30

replaces the old custom controls which

14:32

are being deprecated

14:34

and then uh entertenant governance has

14:37

gone G. So this will help a number of

14:40

different things. So one it will help me

14:42

as an organization detect almost shadow

14:45

tenants being used by my company. So

14:47

based on patterns of external

14:49

identities, multi-tenant apps, even

14:51

billing, it will go and find those other

14:54

tenants. It will then help create

14:56

relationships to help administer those

14:59

other tenants. And then I can also

15:01

enable a secure tenant creation. So any

15:03

new tenants are configured correctly at

15:05

creation time. And there is an API

15:08

available now. And some of the features,

15:10

hey, they're still in preview. But

15:14

that is it. As always, I hope this was

15:16

useful. Until next video, take care.

Interactive Summary

Loading summary...