HomeVideos

Global Network Webinar - Introduction Module of Advancing Responsible AI - 3 June 2025

Now Playing

Global Network Webinar - Introduction Module of Advancing Responsible AI - 3 June 2025

Transcript

1378 segments

0:02

We have eight modules to promote

0:05

understanding of the need for guard

0:09

rails in managing AI

0:12

risks. And the table of contents we look

0:16

first into the diversity of AI. Then we

0:20

try to understand the AI risks and then

0:24

little bit discussion about guard rails.

0:27

What are they they and then

0:30

we are exploring the principles of

0:33

responsible AI and a little bit about

0:36

the AI regulations also and then very

0:39

short introduction to other

0:41

modules but let's first try to

0:44

understand the diversity of

0:48

AI you can all you can always uh take a

0:53

different kind of lenses and and one

0:56

lens is that that you uh study AI from

1:02

the roles of AI. For example, this

1:06

supportive AI has a very has a has a

1:10

their AI has very clear assistive role

1:14

and that's very clear nowadays for

1:16

everyone and it's a tool for tool to

1:20

enhance expert work and it acts there as

1:25

a asparing partner or knowledge

1:29

enhancer and there the expert remains

1:32

always accountable for how AI generated

1:36

information is shared and used. And

1:39

there on the opposite side we have this

1:42

autonomous AI uh full automation where

1:46

the AI performs tasks totally without

1:51

human

1:52

assistance. AI chat spots are like this

1:55

and this where the hype is now very very

1:59

very very high. this agentic AI it it

2:04

the agent it refers to systems that make

2:07

independent decisions they plan actions

2:10

without human

2:12

assistance and this also these

2:15

self-trained AI

2:17

models what we are we have been for

2:19

example at statistics Finland doing for

2:22

years

2:23

now they are trained for specific

2:26

purposes uh and they are also autonomous

2:32

although they are not foundation

2:35

models and they're there in the bit

2:37

there in between there's this

2:39

collaborative AI where the AI

2:43

does the main part of the work but the

2:48

there's still human in the loop who

2:50

decides if the output of AI was correct

2:55

or not and the expert remains

2:59

accountable

3:00

able for ensuring the correctness of AI

3:04

generated

3:06

outputs. But you can also look at

3:10

the AI from totally different

3:13

perspective because AI will arrive and

3:17

arrives in many different packages.

3:21

Uh I borrowed a picture from a slideshow

3:26

from Gardner because I thought this is

3:29

really a good example and it helps to

3:32

understand the diversity of these

3:35

packages. It enters to your office. So

3:40

on the on the right hand side there's

3:43

this green green green box and there are

3:46

these off the shelf applications such as

3:49

co-pilots, powerbi, chat GPT and so on.

3:54

So where AI is embedded into software by

4:00

vendors and then the the red box uh

4:05

bring your own AI box. There are

4:08

solutions de developed where within the

4:11

organization for specific

4:14

needs for example by different

4:16

department

4:18

uh custom GPTs in chat GPT platform are

4:23

such a

4:24

things and that also includes shadow

4:29

AI and and and that's

4:33

also an AI which is totally without

4:38

oversight. And then in the blue blue

4:41

box, we have these in-house trained

4:44

models or adapted foundation models. We

4:49

are probably fine-tuning and so on. And

4:53

in the blue box, there are also models

4:55

embedded by software develops developers

4:59

into the software.

5:02

And as we can see that there in the

5:04

middle there's this circle that says

5:07

that we should be able to coordinate and

5:10

run and secure all this that comes into

5:14

the

5:16

office.

5:18

And then we then there's a another

5:22

perspective a third

5:25

perspective while AI may operate

5:29

autonomously or collaborative the

5:32

significance and the risks of its task

5:36

must be

5:39

assessed

5:41

and not all AIdriven act actions they

5:45

they don't carry equal height or risks

5:47

for the organiza for the organization's

5:51

core

5:52

objectives.

5:54

So task

5:56

significance so how critical is the

6:00

task to the organization core operations

6:04

or goals. So low sign significance is uh

6:08

AI drafting a standard email response.

6:12

But the example of high significance is

6:15

when AI is generating statistical

6:18

outputs used in a national policym and

6:23

there's a big difference within these

6:25

two but we also have to do risk

6:29

assessment uh where we try to identify

6:32

the potential consequences when AI makes

6:36

a failure.

6:38

So what's the potential impact of AI

6:42

failure

6:44

in in in this task? You ask yourself

6:48

this question. So examples of lowrisk

6:51

task. So again automating document

6:55

formatting and the high-risk tasks. What

6:58

happens when AI editing data for

7:01

official statistics makes a failure?

7:04

It's a there's a huge difference.

7:09

So not all AI is created

7:13

equal. So we may actually mean very

7:18

different things varying not only

7:21

function but in significance impact and

7:25

risks for official statistics.

7:28

So it's very important that we recognize

7:32

that AI also enters statistical offices

7:37

very quietly embedded in tools,

7:40

platforms and

7:43

services and we don't really know

7:47

uh about these hidden risks when we are

7:49

not aware that there is an AI. So I

7:55

think

7:55

especially producers of official

7:58

statistics we have to have context

8:01

specific policies and workflows that

8:05

they reflect to the nature of AI task

8:09

its role in production and the

8:11

consequences of

8:14

failure and

8:16

and that would somehow ensure that AI is

8:20

neither overregulated nor left

8:24

unsupervised and you have to find the

8:27

balance

8:28

here

8:30

and adopting AI it is really balancing

8:34

risks and

8:37

benefits and now particular particularly

8:41

uh generative AI at the present is it's

8:44

fundamentally transforming all

8:46

industries not just uh

8:49

statistics it's it's offering

8:52

groundbreaking thinking opportunities

8:54

and and really on the other side

8:58

significant

9:00

challenges.

9:03

So the more organizations utilicize AI

9:07

to enhance their operations, the more

9:10

risks arise. So I I I somehow thought

9:15

that it's it's like a it can be compared

9:19

to preparing puffer fish. If the

9:22

preparation of of fugu puffer fish goes

9:26

wrong, the chef uh has to commit

9:30

hariri because it will poison all the

9:34

all those people who ate that. So when

9:36

we and when producers of official

9:39

statistics when we utilize AI in our

9:42

data production, we bear a full

9:46

responsibility for the accuracy

9:49

uh and accountability of the data we

9:52

have generated. Just the same way as a

9:57

chef is responsible for the preparation

10:00

of fugu.

10:02

So but but we cannot always eliminate

10:05

all the risks but we can minimize risks

10:09

and manage the

10:11

responsibility while we are maximizing

10:14

the

10:16

benefits.

10:19

Uh unfortunately AI is a as a is a

10:24

greatest power of cyber risks at the

10:27

moment already. It's not only in the

10:29

future.

10:32

uh for example an AI powered chatbot

10:35

with uh weak security can be manipulated

10:40

through prompt injection attacks to leak

10:43

sensitive information and many other

10:46

things. So and many identified AI

10:50

risks are also cyber risks but not all

10:54

are cyber risk is especially refers to

10:57

digital threats where attackers

11:01

deliberately exploit system vulner

11:04

vulnerabilities

11:06

uh data breaches uh information leaks

11:09

and so on.

11:15

So I think always that understanding AI

11:19

before using it it is a question of

11:23

responsibility also.

11:25

So when adopting it we must understand

11:29

what we are working with. So we just

11:33

cannot take a black box such as

11:36

generative AI is now and ask it to

11:39

produce statistics for us. So, so the

11:44

the hype around generative AI has been

11:46

immense immense and as a as a result

11:50

it's not always clear

11:53

what generative AI is really useful for

11:58

and where it's simply not mature enough

12:02

yet.

12:04

uh but one way to begin to understanding

12:07

the opac parts of AI is to look at

12:11

through the lens of risk and of course

12:14

as a producers of official statistic

12:17

what would what could be more natural

12:19

than try to classify the

12:23

risks and why is AI risk categorization

12:29

so important

12:31

because

12:33

Yeah, thousands of AI related risks they

12:36

have been already

12:39

identified but it is impractical to

12:42

assess each one individually. So instead

12:45

categorizing these risks at a higher

12:48

level it helps us to understand it

12:50

better. Uh there is already such a

12:54

living database called MIT AI risk

12:58

repository. It's been generated by MIT

13:03

researchers and there are own over at

13:06

the moment over

13:08

1,600 AI risks categorized by their

13:12

cause and risk

13:14

domain and the causal classification

13:17

tells that tells you also how when and

13:20

why these risks occur. So I really

13:24

encourage the statistical society to

13:27

explore and consider utilizing this AI

13:30

risk

13:32

repository. Really really it's a really

13:35

good uh tool for identifying

13:39

risks. Uh I have picked up couple of

13:45

uh specific uh

13:47

subclasses from this MIT classification.

13:53

For example, uh this governance failure,

13:58

uh it's it it tells that there are very

14:01

weak rules and oversight that can't keep

14:05

up with AI progress and that leads to

14:09

poor risk management and then this

14:13

competitive dynamics. It's it means

14:16

that rapid AI development is is driven

14:20

by competition and and that is

14:23

increasing the risks of unsafe and

14:27

faulty

14:28

systems. But uh unfortunately I have had

14:31

not had enough time to go through so so

14:36

in

14:38

so in detail this um taxonomy of MIT

14:43

that I still used my own classification

14:48

uh in this presentation because there

14:51

are a couple of things that are lacking

14:53

there and and that perhaps I should have

14:56

somehow made a combination of these too

14:59

because for example these ethical and

15:02

societal risks and societal impact risks

15:05

which I call actually there in another

15:08

slide microlevel risks they are

15:10

something that are little bit lacking

15:12

from that other other classification and

15:16

and many other re researchers have also

15:21

attempted to classify all these uh AI

15:24

related risks

15:26

and they can be really classified from

15:30

multiple perspectives.

15:36

I go through now the

15:39

uh a main classes of the AI

15:43

classification I have made but there

15:46

there are some there's a they they at

15:50

the risk side they

15:53

use such a vocabulary that may be

15:57

sometimes a little bit strange for

15:59

example this prompt prompt injection

16:01

it's a malicious

16:04

where man malicious actor manipulates an

16:06

AI system prompts uh and that orders the

16:10

the model's behavior uh and

16:13

jailbreaking. It's it's where AI refers

16:17

to bypassing safeguards to make an AI

16:20

system generate proh prohibited and

16:24

unintented outputs and so on. And uh but

16:29

perhaps if if somebody wants to check

16:33

more about this vocabulary, this

16:36

presentation will be

16:40

shared.

16:42

Uh AI systems, they often handle

16:46

sensitive data and that makes them

16:48

really vulnerable to privacy violations,

16:52

all kind of data breaches and cyber

16:55

attacks.

16:56

So and um if there's a poor data

17:00

protection, it increases these risks

17:04

posing threats to organizations and

17:09

individuals. And I have some exo

17:12

examples here

17:15

uh

17:18

uh like like this generative

17:21

uh AI tools they have unintentionally

17:25

leaked user data and that happened also

17:28

to chat GPT

17:32

2023 where data bridge was where there

17:36

was a data bridge where some users could

17:39

see other users

17:41

uh chat histories and payment

17:45

information due to a buck. And this was

17:48

just an short

17:51

uh example what can

17:56

happen, what kind of privacy and

17:58

security risks can really

18:01

occur. And then these operational risks

18:05

they pertain challenges in in

18:08

maintaining the efficiency and cost

18:12

effectiveness and the reliability of AI

18:17

systems. Uh it's about ensuring that the

18:21

AI infrastructure remains controlled and

18:26

predictable and maintains the efficiency

18:29

and reliability.

18:32

Uh these operational failures they are

18:36

uh can be such failures where the system

18:39

starts to degradate and reliance on

18:42

third party services can then lead to

18:45

failures and generative AI systems

18:49

may be vulnerable to these denial of ser

18:54

service attacks

18:56

uh which can make them temporarily

18:59

unusable

19:00

And and there are also resource

19:04

dependency

19:05

risks. For example, high cost and

19:09

limited expertise can create

19:11

dependencies on on just few large

19:15

technology

19:21

companies. And then we have uh ethical

19:25

and societal risks.

19:31

uh and

19:34

examples. This I think these are most

19:37

familiar to all statisticians. For

19:40

example, this bias bias in in training

19:43

data. It can lead to unfair and unfair

19:49

outcomes affecting groups on race,

19:51

gender or societ soc economic status.

19:58

uh manipulation of societies is can be

20:03

something like AI algorithms on social

20:06

media platforms can amplify political

20:14

polarization and then this macrolevel

20:18

impacts soci societal impact

20:22

risks. So it's it's not

20:24

only the individual users AI can affect

20:29

it can also shape society as a whole and

20:33

these micro level risks include economic

20:36

environment or even existential

20:39

concerns. So example of economic

20:45

risks that AI and automation can replace

20:49

jobs especially low skilled ones and in

20:53

and that can cause inequal

20:56

inequality and at the same time for

20:59

example new AIdriven jobs they require

21:03

reskilling and retraining and perhaps

21:07

not everyone has access to

21:10

and environmental

21:13

risks where large AI models consume a

21:17

lot of energy and resources.

21:20

My daughter always says to me when I

21:23

answer I check that everything from

21:25

chach GPT that don't use that chachpd it

21:28

will

21:29

destroy our in whole environment and so

21:33

I think they the younger generation is

21:36

already more aware of that

21:39

uh and then these existential risks

21:43

uh some experts warn about

21:48

uh uncontrollable super super

21:51

intelligent AI that could pursue goals

21:55

that conflict with human values and and

21:58

that can cause serious long-term risks

22:01

to humanity. But these kind of risks

22:05

have not occurred yet, but this is

22:08

something that some researchers see that

22:12

might happen in the future.

22:17

And then uh explainability and

22:21

transparency risk although

22:24

other responsible AI principles

22:28

explainability and transparency they are

22:31

in in totally separate things but but in

22:35

this risk classification I've added them

22:38

into one and same risk class. So they

22:43

can make these uh lack of explanability

22:48

and

22:49

transparency can make it difficult to

22:52

stay for stakeholders to understand the

22:55

whole decision making process and and

22:58

that really undermines the trust. So if

23:02

stakeholders can't follow how AI works,

23:05

trust and accountability suffer.

23:10

So many system are like black boxes. The

23:15

logic is hidden or it's too complex to

23:19

explain. So if we can't explain AI's

23:23

role and decision making process,

23:26

uh I think people are less likely to

23:29

accept and oversee its use. So output

23:35

clarity is important and generative AI

23:38

models may produce

23:41

fabricated or culturally inappropri and

23:46

inappropriate content.

23:50

So complicating the the whole

23:53

explanability process. Uh for example is

23:58

uh hallucinations and gen AI outputs.

24:02

They require still human re review as

24:07

their origins are often opac. So that

24:11

makes trust and validation really

24:14

difficult.

24:16

So decisions made by AI can be hard to

24:20

understand and a

24:25

explain. And

24:27

then functional and technical

24:31

risks. This focuses on vulnerabilities

24:35

and exploits within AI models and

24:38

systems. And

24:40

uh yeah the risk category we had early

24:44

operational reliability risks focuses on

24:48

practical management and resource

24:51

dependencies. So

24:56

uh system exploits as a example where

25:00

attackers can bypass safety

25:03

measures using these prompt injections

25:06

and jailbreaking methods, should I call

25:10

them methods, even though they are quite

25:12

criminal methods and they they allow

25:15

malicious users to make the system

25:18

produce

25:19

really content.

25:22

uh and model integrity risks

25:26

uh are where over time the AI systems

25:31

they become less

25:34

reliable. And the example of of

25:37

course quite familiar for for those data

25:41

scientists is is model drift system

25:45

accuracy cleans and the real world

25:47

condition when real world conditions

25:53

change. And other examples I date the

25:56

poisoning and and other inconsistent

26:00

behaviors.

26:04

uh legal risk they arise from the reg

26:08

regulatory and liability challenges that

26:11

are associated with AI. Uh there are

26:14

copyright copyright problems

26:17

and and and with regulatory compliance.

26:22

AI evolves faster than many legal

26:24

systems. So that creates a confusion

26:28

about who is responsible and what rules

26:31

apply

26:33

and regulator regulatory non-compliance

26:37

uh failure to add to AI specific

26:41

regulations such as what we have now in

26:43

EU AU's AI act or GDPR which is quite

26:47

important here on the AI

26:51

uh branch. So it can lead to legal

26:54

penalties and

26:56

restrictions. Uh organizations must

27:00

actively follow legal

27:06

updates and then we come to these guard

27:09

rates in the context of artificial

27:12

intelligence. So AI guard rules they are

27:16

safeguards that ensure AI systems

27:20

operate safely and ethically and and and

27:24

operate so as they are intended to

27:27

operate. So the term now covers a broad

27:30

set of protections that help minimize

27:34

and and mitigate these risks. So they

27:38

can be technical guard rails, they can

27:42

be organizational guardrails, they can

27:44

be ethical guardrails. And why these

27:47

guard rails? Why do they matter? Because

27:49

without these

27:51

guardrails, AI can

27:55

behave unpredictably.

27:58

So guardrails ensure that AI systems

28:01

operate safely, ethically and minimizing

28:04

these potential

28:06

risks. Examples, MLOps and LLMOPS. They

28:10

are practic practical guard rail

28:13

frameworks. They include ethical uh uh

28:17

uh and

28:18

technical guard rails. So they provide

28:22

tools and processes for managing AI

28:25

through its entire life cycle. Uh

28:29

responsible AI principles. They are more

28:32

ethical guard rails. They guide AI to

28:37

align with public values like fairness,

28:39

explainability and so on. Uh and then we

28:44

have AI regulations. Um they are legal

28:48

guard rails.

28:52

And then again let's put these all

28:55

things together. So there are these many

28:59

kind of guardrails. And the goal of this

29:03

whole

29:05

whole project or whatever I should call

29:07

this advancing responsibility AI is to

29:10

build a network of guard rails to

29:13

support the safe use of AI especially of

29:17

course in official statistics.

29:19

So what does it mean in practice? So

29:23

that you put the right guard rails into

29:26

the right places. So strict oversight

29:30

there where it's needed and no

29:32

unnecessary restrictions where it's not

29:35

needed. So guardra protect core

29:39

principles of responsible

29:44

AI. And how do we turn the principles

29:48

into guard rails? So that you identify

29:52

risk is just the beginning. So these

29:57

principles, responsible

29:59

AI principles, they act as a guard raise

30:03

only when they are not just words but

30:07

goals we actively commit to. So we need

30:12

organizationwide

30:13

commitment that we can put these

30:16

principles into

30:18

practice. And once the principles are

30:20

set, the next step is to turn each one

30:24

of one into concrete

30:27

actions. For example, when we talk about

30:30

transparency, MLOps can support it. But

30:34

transparency, it requires governance

30:37

structures and clear communication to

30:40

stakeholders. And for example, when we

30:44

talk about data protection and privacy

30:47

techniques, we know like anonymization

30:50

and or sedonmization they help but they

30:54

are not enough without proper procedures

30:56

and oversight mechanisms. So in other

31:00

words, principle responsible AI

31:03

principles, they become guard rails only

31:06

when supported by practical tools,

31:09

processes and

31:14

accountability. And then we have the

31:16

principles of responsible AI. Uh the

31:21

naming and the division of responsible

31:24

AI principles, it's it's not

31:26

standardized. However, the overall set

31:30

of aspects they encompass has largely

31:34

become established standard. So I have

31:38

here

31:40

uh uh the principles I have used here is

31:45

ethics, data privacy, data security,

31:48

transparency, explanability,

31:50

inclusivity, fairness and

31:52

accountability.

31:55

uh transparency

31:57

is one of what I think it's one of the

32:01

most important principles. It's not only

32:04

a important principle in AI but in the

32:07

whole statistical production. So it

32:10

means that the processes decisions

32:12

decisions and the use of AI system is

32:15

open and visible to all stakeholders.

32:20

And it's not only about technical

32:22

transparency ML ops can provide but it's

32:28

also the organizational

32:30

transparency. It's clearly communicating

32:33

how AI is governed by whom and under

32:36

what

32:38

rules. It includes also stakeholders

32:41

access access to relevant document

32:45

documentation how the statistical office

32:47

has produc used AI in producing a data

32:53

and why this matters it enables trust

32:57

and

32:58

fairness and again I'm repeating now I

33:01

recognize now technical transparency

33:04

alone is not enough the stakeholders

33:07

they also must understand the context

33:10

and scope of AI use. So real

33:13

transparency then includes both

33:16

technical clarity and organizational

33:19

openness.

33:24

uh AI should also make fair and unbiased

33:31

decisions and fairness means that

33:35

predictions and decisions must be

33:38

equitable and free from

33:40

discrimination and fairness applies

33:43

across the whole AI life cycle from

33:47

training the model to output.

33:52

And how do we achieve this fairness? It

33:56

requires really careful attention to

33:59

first to the training data. We all know

34:02

that biased data produces biased

34:05

results. Tools for bias

34:09

detection. Fairness is isn't one time.

34:13

It requires continuous technical checks

34:16

and monitoring and also ethical

34:20

reflection. And why transparency and

34:23

explainability matter? Because fairness

34:27

loses meaning if experts and

34:30

stakeholders can understand AI

34:33

decisions. Explaining how decisions

34:36

decisions are made builds trust and

34:41

accountability.

34:42

And again generative AI brings really

34:46

big extra

34:49

challenges outputs generative AI makes

34:52

they are not always very traceable

34:55

uh and not traceable to specific

34:58

training data although they are open

35:00

access models but it's impossible to go

35:04

through all that data it has used that

35:07

has been used in training that model and

35:11

that makes bias detection and and

35:13

explanability really hard and I would

35:16

say it's uh quite often even impossible

35:19

with generative

35:23

AI. Uh

35:25

accountability means that we have clear

35:29

responsibility for decisions AI

35:33

makes and who is

35:36

responsible. We have to define

35:39

ownership. Who ensures the quality and

35:42

reliability AI assisted data at each

35:46

phase of statistical

35:51

production? Can we trace and verify?

35:55

Yes, AI models and data set must be

35:57

documented and auditable. It must be

36:00

clear how the data was formed or how the

36:04

decisions were

36:06

made and data production must follow

36:11

regulations and also align with

36:15

principles ethical principles of

36:18

official statistics.

36:23

But the final

36:26

responsibility is is a big question and

36:30

NSOS must ensure that the data is

36:33

suitable for decision making and

36:37

research and does not carry hidden risks

36:41

or

36:43

misinterpretation. So NSO is

36:46

responsible.

36:48

Now at the beginning I had these three

36:51

roles there and it's quite clear

36:55

that with with the supportive and

36:58

collaborative AI there's this human in

37:02

the loop and human oversees the result

37:05

and the impact of AI and is

37:09

responsible and the responsibility lies

37:12

with that user. But then with autonomous

37:16

AI it gets a little bit more

37:19

difficult. The person using or

37:22

triggering it. It is not re responsible.

37:26

So rather the responsibility lies

37:29

somewhere with those who have generated

37:31

the service and authorize authorized its

37:35

deployment.

37:37

So this is not that simple anymore.

37:44

uh ethical considerations with AI they

37:47

they must be core part of AI development

37:50

and use

37:52

uh and yes not an

37:56

afterthought and what ethical AI means

38:00

it means that AI must respect societal

38:04

values and avoid

38:06

harm and it should not unfairly benefit

38:10

or disadvantage manage specific

38:14

groups and how do we ensure that our AI

38:19

is

38:20

ethical. It regard requires really clear

38:25

governance who is responsible for

38:28

ethical

38:29

compliance and requires accountability

38:32

across the whole AI life cycle and

38:36

requires ethical impact assessment. Not

38:40

just that technical validation and

38:44

guiding ethical principles, doing good

38:47

and not doing harm and respecting human

38:51

agency and promoting fairness and making

38:56

things understandable.

39:02

And one of

39:04

the perhaps well-known

39:08

uh principles is data privacy and data

39:12

security and that's really protecting

39:15

personal

39:17

data and

39:19

it's for statistical offices it's quite

39:23

important

39:24

thing and and also it's essential for

39:28

maintaining public trust.

39:32

and AI systems. They must comply with

39:35

laws like GDPR, but they also must

39:39

protect sensitive data using strong

39:42

technical and organizational

39:46

safeguards. Communicating clearly how

39:49

data is handled

39:51

helps earn and keep that public trust.

39:57

Uh it

39:58

requires regulatory

40:01

compliance, technical safeguards like

40:05

usage of encryption, secure storage

40:08

access control, certainization,

40:10

synthetic data and so on and ethical

40:15

principles in practice. It's conducting

40:18

bias audits, ensuring transparency in

40:21

how the data is processed and used.

40:25

So privacy and security they are not

40:27

just legal obligations. They are

40:31

fundamental to trust for the and

40:33

responsible AI and again they are

40:38

fundamental to statistical offices for

40:42

to remain public

40:45

trust and then

40:49

explainability. Uh how do we achieve it?

40:53

We have tools and methods that help

40:56

technical and

40:59

non-technical people understand the

41:01

model's reasoning. Uh we have common

41:04

methods for explainable AI like lime and

41:08

shep. Uh but

41:11

uh deep learning and large language

41:13

models are much harder to explain

41:17

decisions in in large language models.

41:20

They rely on such a complex patterns

41:22

that are really hard to trace and

41:25

traditional tools

41:28

like lime and shop. They often fall

41:32

short and that makes

41:34

explanability and therefore trust also

41:38

harder to achieve in high stakes uses.

41:46

inclusivity it's it's

41:49

not as a principle in all the

41:53

uh responsible AI classifications but I

41:58

always want to have it as a separate one

42:00

because I think that that it ensures

42:04

somehow equal repres

42:07

representation in in the whole AI

42:09

development and helps prevent unintended

42:12

exclusion or harm and build systems that

42:16

reflect the needs of broad and diverse

42:20

population

42:22

and but uh why it's difficult and

42:26

especially in generative

42:29

AI because generative AI models often

42:32

learn from biased training data. So we

42:37

have then also quite often biased

42:40

output.

42:41

This can lead to unequal treatment or

42:44

exclusion of certain

42:47

groups. And inclusive public sector AI

42:51

is is really important because we have

42:54

to design services usable by all

42:58

citizens regardless of

43:01

technical skills.

43:06

So uh now I've got gone through go

43:10

checked all the uh

43:13

principles uh of responsible AI but then

43:17

couple of slides about implementing

43:19

responsible AI and checking a little bit

43:22

about the AI regulation side. So

43:25

implementing responsible AI is not a

43:28

theory it's a really a commitment to a

43:31

to action. It ensures the AI supported

43:37

statistics that they are transparent,

43:39

fair, accountable, secure, ethical and

43:41

so on. By actively committing to these

43:45

principles, statistical producers can

43:49

strengthen the public

43:51

trust and increase the reliability and

43:55

credibility of

43:56

outputs and ensure AI enhances rather

44:01

than compromises statistical integrity.

44:04

So without responsible AI, I think

44:08

there's a higher risk of bias, misuse

44:11

and loss of credibility and it would

44:15

will undermine evidence-based policym

44:19

and public trust in official

44:24

statistics. And then uh another piece in

44:27

these AI puzzle uh AI regulations, they

44:33

embed key responsible AI principles like

44:38

transparency, fairness and data privacy

44:41

into formal legal

44:43

requirements and when when responsible

44:47

AI are voluntary and internally defined

44:51

while the regulations bring external

44:54

enforcement. ment and industrywide

44:57

consistency. So most AI laws are

45:00

motivated by responsible AI concerns

45:03

such as avoiding bias and ensuring

45:06

explanability together. I I think they

45:09

complement each other. So responsible AI

45:12

builds culture and values and

45:15

regulations ensure compliance and and

45:18

and accountability.

45:20

And examples from practice all big

45:23

companies, tech companies like and for

45:27

example Microsoft and Google, they apply

45:29

also

45:31

responsible AI principles

45:36

uh even beyond the

45:38

laws required uh just to strengthen the

45:42

public trust and ensure uh ethical AI

45:46

development.

45:49

uh about the AI regulations that's our

45:53

module two actually uh just two slides

45:58

about this we have two broad approaches

46:01

we have regulated AI some specific AI

46:06

laws such as uh this European Union EU

46:10

act and then we have in United States

46:13

federal and state level AI legislations

46:16

and in China We have nationally

46:19

developed AI regulations and then we

46:22

have existing laws applied to AI and

46:26

most of the countries rely on general

46:28

laws uh to regulate AI and including

46:33

Australia, Canada, New Zealand and so

46:35

on. But but

46:38

but definition of AI varies a lot. We

46:43

have for example OECD definition EU's AI

46:46

act uh SNA and balance of payments

46:51

uh US census bureau and this definition

46:55

they vary a lot and it's it makes things

46:58

a little bit difficult sometimes

47:02

uh EU's AI act it's the systems are

47:06

classified into different risk

47:09

categories based on how they likely the

47:13

how likely the risk is and what the AI

47:17

systems is in intended to be used for.

47:21

So problem it I see there's a problem

47:25

related to the definition of an operator

47:29

the AI act. So the use for example

47:32

statist Finland is likely the use of AI

47:36

is is likely to fall into the low or

47:38

minimal risk category because Statistics

47:42

Finland itself does not make decisions

47:44

or take actions based on the data it

47:47

produces. The actual decision maker who

47:52

uses AI generate data from statistics

47:55

Finland as a basis for decisions may not

47:58

be using AI themselves and therefore

48:03

activities may not fall under AI acts

48:06

definitions of an AI operator user. So

48:11

this creates a a regulatory blind spot

48:15

where the influence of AI is significant

48:19

but the no single actor formally meets

48:22

the definition of AI user operator I've

48:26

forgotten the term in in English under

48:29

the regulation. So just because and just

48:34

because something is legally permitted

48:37

for statistical offices it does not mean

48:40

that it's ethically acceptable

48:43

especially yes from the perspective of

48:45

official statistics. So even though uh

48:49

these regulations don't force us to do

48:53

things or avoid doing things I think we

48:57

have to be more

49:00

uh look at this more from the ethical

49:04

perspective. So once more a picture

49:07

perhaps there should be one more picture

49:10

before the rise risks identified that at

49:15

AI better understood and then AI's risks

49:19

identified guardrails built and

49:22

responsible AI adapted that's the

49:25

process it should

49:28

go and

49:30

then short introduction to other modules

49:34

of this advanced responsible AI. So we

49:38

have total of eight modules uh with this

49:42

introduction introduction serving as the

49:45

first one. Uh the different modules they

49:48

cover various aspects of responsible AI

49:51

and explore strategies and guardrails

49:54

for more effective risk management. Uh,

49:59

additionally, advancing responsible AI

50:02

seeks to consider the role of AI

50:05

regulations in ensuring responsible

50:08

statistical production. The other models

50:11

are ethical principles. Uh, this module

50:15

will likely be feature

50:18

uh a lecturer from academia. Uh the next

50:22

one which will be quite soon is actually

50:24

AI operationalization. MLOps and Llops.

50:29

uh it's uh where ML ops provides guard

50:33

rails in various forms and then we have

50:36

the

50:37

explanability and data privacy and

50:40

security and then we have case

50:44

studies and then we have this continuous

50:47

learning and

50:48

adoption that covers

50:51

uh uh provides practical advice and tips

50:56

for some kind of ongoing develop

51:00

velopment. The dates for the other

51:04

modules will be announced as the web

51:07

femininas are finalized. So after

51:10

holiday season which in Finland is quite

51:13

long but but uh somewhere in the autumn

51:18

the third module will be

51:22

ready. Thank you. That was the last

51:26

slide.

51:29

Thank you Rita for this very great um

51:34

presentation.

51:36

So uh in the meantime while you were

51:38

speaking there were already a couple of

51:41

questions coming in in the chat.

51:45

Um maybe be while I try to uh switch on

51:51

the cameras and the microphone. Maybe Pa

51:54

you can try to do that if I don't manage

51:57

to do it. Um so maybe in as a starter

52:02

let me ask you one question first and

52:06

then maybe somebody can come in live.

52:09

There is a question from

52:12

Shahu Ibraim Sharif. Sorry if I'm

52:16

mispronounced the name. Um, can you give

52:19

a real world example how you used AI in

52:24

a statistical field in in official

52:27

statistics and what was the reason for

52:30

using

52:31

AI and was this uh feature accessible to

52:35

the public such such as journalists or

52:38

other users? Yeah. Uh yeah. uh at least

52:43

at Statistics Finland we have used uh AI

52:47

in in classification. We have

52:52

uh generated our own AI models in quite

52:57

many classification

52:59

cases and why we what was the reason for

53:03

using uh I

53:05

think uh when we use AI based

53:09

classification for example we don't need

53:12

that much human hands anymore because

53:17

going classification

53:20

do with text based classification done

53:24

uh manually. It's really hard work and

53:26

it's a lot of human hours, work hours.

53:34

Yeah. Yes. Thank you. So, exactly. So,

53:39

um AI is often

53:42

used to reduce basically the human

53:45

interaction or the human uh decision,

53:48

right? to to make

53:51

the processing of uh statistical

53:54

information more efficient. Um so I have

53:59

now enabled the microphones and the

54:02

camera. So if you have questions and you

54:05

would like to ask them live uh so maybe

54:08

raise your hand and then we can give you

54:12

the floor.

54:16

Um, and I saw that there was a question

54:20

by

54:22

Tutu again. Sorry if I mispronounced the

54:25

name. Uh, do you want to come in or

54:28

should I read out the

54:36

question?

54:40

Yes. And this is not

54:43

Can you keep

54:45

uh no so the question was do you have an

54:49

example use case for a technical or

54:52

organizational guard rail? Uh maybe we

54:56

can continue with that question. Thank

54:58

you. Yeah. Well, that's a good question

55:01

because that that's the thing I love

55:04

because we have been building uh our

55:07

infrastructure technical infrastructure

55:09

for for machine learning for some years

55:12

and and we have managed to somehow

55:15

finalize that and when we have the

55:18

infrastructure there in

55:20

cloud we we we also use such a such a

55:25

components that we uh version

55:29

everything. So everything is

55:32

reproducible and and and I think this is

55:35

this is really one of the

55:37

technical technical

55:40

uh guard rails, one of the best but of

55:44

course I always have to remind that that

55:47

when we come to generative AI we are not

55:50

that far. No one is that far that we

55:52

could say that that all the technical

55:55

guard rails are

55:59

there.

56:00

Yeah.

56:02

Um there there's another question by

56:05

Tracy Cougler. Tracy, not sure if you

56:09

want to come in yourself otherwise I can

56:12

read out the question for you.

56:16

Hi. Yeah. Um so my question is around

56:20

the kind of tension

56:21

between transparency and how hard it is

56:26

to really understand where the output is

56:29

coming from in a lot of AI systems

56:33

and thinking about how you kind of

56:36

manage that tension or draw that line

56:39

between when and how it's responsible to

56:42

use systems where you can't always fully

56:45

understand where the output is coming

56:47

from and if there are circumstances

56:50

where you should just avoid using those

56:52

systems entirely.

56:56

But do do you actually mean more now

56:59

explanability than transparency?

57:02

Uh both I think. Yeah. Yeah. Yeah.

57:06

transparency for example with the models

57:09

the self-trained models it's it's quite

57:12

quite easy actually when you have the

57:15

infrastructure and the processes but the

57:18

explanability I think it's always much

57:21

much more

57:24

difficult so

57:26

uh

57:29

yeah is

57:32

is is there actually a way to for let's

57:37

say if you use for instance a large

57:39

language model to to identify and

57:43

classify uh things right um and you

57:48

haven't trained that system yourself you

57:51

use something that you basically buy

57:55

from somewhere uh yeah you through or

57:58

which you use through an

58:00

API and you don't really know what's

58:03

going on in in that black box. Um how do

58:07

you deal with that? How do you explain

58:09

that to the users of official

58:12

statistics?

58:14

Yeah. And at the moment uh there there

58:19

are some for example some evaluation

58:21

agents. We are not using them. We are

58:24

just starting to use them. But to

58:27

evaluate the the output of large

58:30

language models that's really

58:32

challenging and I think no one at this

58:35

this earth has really solved the problem

58:39

but but to use large uh

58:42

products where large language model is

58:45

used in that way that there's still this

58:47

human oversight that's that's I that I

58:51

think it's it's still somehow acceptable

58:55

and and can be

58:59

somehow let's say there's this human

59:02

that's evaluating the result but as a

59:06

autonomous

59:08

AI I would not use large language

59:12

models and that's what the academia says

59:15

all the time and and there's this other

59:18

other hype that goes goes there in the

59:22

private private sector that you can use

59:26

large language was all over but when you

59:29

know it and understand it more you say

59:33

not yet

59:38

okay makes sense thank you

59:41

I see in hand up that's Amra

59:45

uh from UNC please come in

59:50

thank you Alex and thank you Rita for

59:52

the present

59:55

uh your presentation as a whole and the

59:59

modules uh as you presented them the

60:01

upcoming ones I they are more focused on

60:04

having a more a solid obviously

60:07

governance structure and governance

60:09

framework for the AI to mitigate it

60:12

risks and so on and this is absolutely

60:15

obviously needed uh I'm just thinking

60:17

also of the opportunities of the AI

60:20

especially for countries in south who

60:22

have very limited

60:24

And that would be mostly in the support

60:26

AI and the kind of at best collaborative

60:29

AI fields maybe. All right. And my

60:32

question do you have like a breakdown or

60:35

did you make like a breakdown of the

60:37

tasks in the production cycle of

60:39

statistics where either supportive AI or

60:42

collaborative AI could be used?

60:49

Yes. Um me and my colleagues we have

60:53

worked through the GSBBM

60:56

uh quite many times where the AI could

61:00

be used and in which role and and yeah

61:04

but it's something that really has to be

61:06

with last time when we did it that was

61:09

nearly a year ago and it's and it's it's

61:12

not it has to be updated because the

61:17

the AI is changing

61:20

so quickly and new things arriving that

61:24

it's yeah but I I think it should be

61:27

really produced such a GSBB and AI uh

61:33

map where you can use it and in in which

61:36

role.

61:41

Thank you. Um I see there's another hand

61:44

up by Zuzu.

61:47

uh please come

61:56

in. I think we cannot hear you. Maybe

62:00

your microphone is muted on your side.

62:02

Can you try

62:06

again? No, we cannot hear you. Sorry for

62:10

that.

62:13

Um yeah, let

62:16

maybe

62:18

uh write the the question into the

62:21

meeting chat. Sorry for that. Uh I will

62:24

try to enable your microphone in the

62:29

meantime. Yeah, there there

62:33

were

62:36

also some questions uh from the

62:39

registration form that we maybe can take

62:42

in in the meantime. Um so the we we

62:47

talked

62:49

already about the use of LLMs so large

62:53

language models and uh there there's a

62:56

question whether you have considered the

62:59

practical use of local large language

63:02

models so I guess large language models

63:04

that are running on your local machine

63:07

in uh and if they have particular

63:10

advantages that might have they might

63:14

have in terms of

63:17

uh energy impact on the one hand but

63:19

also in terms of data conf

63:23

confidentiality. Maybe you can elaborate

63:26

a little bit about that. Thank you.

63:32

uh I I easily start thinking about

63:36

uh deepseek and R2 when we are talking

63:40

about uh locally running

63:44

uh models large language models locally

63:47

and

63:48

and I just read an article this morning

63:52

about warnings of this. So yeah, I I'm

63:58

not a very big fan of this running these

64:02

models locally. So I still think

64:06

that common infrastructure for this is

64:09

the best one. Although it's

64:12

environmentally sometimes a little

64:15

bit not so positive thing.

64:21

Yeah. Thanks. Uh, Tuzu, you want to try

64:24

to come in once

64:29

more? Let's try once

64:32

more. No, sorry, not not working. Okay,

64:37

so that

64:41

um so there there was a question about

64:44

the use case for technical and

64:46

organizational uh guardrails and Tuzu

64:50

wanted to ask whether you have

64:54

recommendations for real time monitoring

64:56

and accountability structures that an

64:59

organization can implement for

65:01

guardrails in

65:02

practice. Um so maybe you can answer

65:06

that.

65:07

Yeah, we have been testing several

65:12

several AI monitors. Uh the problem is

65:17

that some of them are really good and

65:20

but they are also very expensive. So we

65:24

have now last week decided uh that we

65:28

start uh uh picking up uh proper

65:33

algorithms from from GitHub and and

65:36

start building the monitor by ourselves

65:40

because they quite often they start they

65:43

are six figure

65:45

numbers you have to pay for the license

65:47

for per year. So it's it's it's really

65:50

expensive.

65:53

So and with the and with the

65:55

accountability it's it's

65:58

a I' been discussing with the top

66:02

management and and we are trying to

66:05

somehow build uh some kind of uh

66:09

accountability

66:11

uh structure that that everyone would

66:16

understand where where their

66:18

responsibility lies. that we are not

66:20

ready

66:22

yet. But I think it's it's really

66:25

something that has to be done quite

66:27

quickly before there's a there here and

66:30

there running and we are not at all

66:32

aware who's responsible for the results

66:35

and and perhaps when the when we start

66:39

putting out totally biased biased

66:43

figures. Oh yeah,

66:48

thanks. There was also a question by

66:54

um Caroline Wood who was saying okay so

66:57

recently we were told that open AI

67:00

smartest AI model disobey direct

67:02

instructions to turn off and even

67:05

sabotage shutdown mechanisms in order to

67:08

keep working. So uh the the question

67:11

there was can guardrails really be

67:13

built? Is it safe?

67:19

Uh uh can guard rails be built and is it

67:24

safe? Is that the question? Yeah. Yeah.

67:27

So can can guard rails really be built?

67:29

Is this something that that works? So

67:33

that was kind of the question there.

67:36

Yeah. I I I can think about for example

67:40

data privacy and security

67:43

principles. Uh I'm I'm responsible for

67:47

that module also but of presenting that

67:50

module also and it's it's been really

67:53

hard to find anything new. I mean we we

67:56

have already this anonymized

68:00

anonymization and and and and so on and

68:04

and it's it's not that easy. I mean,

68:08

yeah.

68:12

Yeah. Some Yeah. It's the infrastructure

68:15

and the organizational uh regulations

68:19

and all these very familiar old things

68:23

that when you combine these I think then

68:26

you can build some kind of guard rails.

68:31

But I recognize that it's it's it's

68:34

difficult to al already talk with our IT

68:38

people because they don't although they

68:40

are gurus in in cloud things for example

68:46

but they don't understand the what the

68:50

the AI brings when you bring a put on a

68:53

large language model into a cloud what

68:56

what's the what's the combination then

68:59

it's it's difficult

69:03

Yeah. Um, there's also a question by

69:06

Kristoff Bon. Uh, Kristoff, you want to

69:09

come in?

69:13

Hello. Good evening. Thank you for this

69:16

very uh stimulating uh presentation. I I

69:20

I it gave me a lot of question. I have

69:24

now more question I think after this

69:26

presentation than before. Also I think

69:27

this is this is a sign that it was

69:29

really really good. Um I I have some

69:32

sort of pro provocative question. We we

69:35

always talk about explanability,

69:37

transparency and reproducibility in

69:40

black and white terms while it seems

69:42

that there are lots of gray zone already

69:45

uh in a sense that

69:47

um I mean it's difficult already to

69:50

explain and it depends once again to wh

69:54

um some models right uh we we are using

69:57

imputation we're using some machine

69:59

learning models inside the NSO um I'm

70:02

not sure we completely understand And

70:04

then we can un fully explain already

70:07

what is going on when you classify

70:09

something as you know a different uh

70:13

different classes. So um the question uh

70:17

I have

70:18

is how we decide the level of gray is

70:22

acceptable and and and who decide that

70:25

is it does should it come from the law

70:29

where somebody has the right to know why

70:32

he has been classified in as

70:36

um in a category or should it come from

70:39

a technical side and and what is the

70:40

role of NSO uh here how how to define

70:44

this level. It's it's it's a big

70:46

question I imagine. Yeah. Yeah. Yeah.

70:48

Yeah. Yeah. It's really is.

70:51

Yes. Yeah. I think we are

70:55

mainly all the time nearly at that gray

70:58

zone. I

71:00

think because I I for for me personally

71:04

explainability and all these methods

71:06

they are they are really difficult to

71:09

understand. I'm a computer science. I

71:11

say that every time when when we start

71:13

talk about those methods that who should

71:15

decide what's the right level. I think

71:19

this is something that has to do with

71:21

the AI governance as a

71:23

whole which is uh much more than just

71:28

this implementing responsible

71:31

AI and it I think it comes there we are

71:35

at the statistics fin the beginning of

71:38

that journey just trying to identify

71:42

what kind of governance structures we

71:46

need. But what what I think what I think

71:49

it's important that our stakeholders we

71:52

can tell something if they ask we we

71:55

just we cannot answer that it's a AI

71:58

that produces these figures. We don't

72:01

know how and

72:07

why. Yeah. Yeah. You should trust your

72:09

models, right? And and this is probably

72:11

the best answer you can say. You're

72:13

trusting because there's human in the

72:15

loop. Yeah.

72:19

Yeah, thanks. Um, so I wanted to give

72:23

another chance uh for people to ask

72:26

questions. So please if you want to come

72:28

in either raise your hand or type it

72:32

into the chat, that's also fine if you

72:35

cannot speak.

72:38

Um, and I see one more question here

72:42

from Carolene Wood. Carolene, you want

72:44

to come in?

72:49

I can just read my question. Yeah,

72:52

that's fine. It's a question I'm

72:54

starting to to ask myself in the past

72:57

few months because it's maybe ethical,

73:00

but if we combine the risk of AI to

73:02

disobey human orders and with its

73:04

environmental impact and we see more and

73:07

more that AI is diminishing the need for

73:10

human work. You said it yourself read

73:13

less words less hands involved in the

73:16

statistical work. So is it worth this

73:20

risk on a large scale to to to use AI at

73:23

all? Why is humanity embarking on AI? I

73:27

know it's a bit of an ethical large

73:29

question.

73:33

Uh I think we just have to at

73:36

least it's it's very clear at Statistics

73:39

Finland. we we have to produce

73:42

statistics meet with much less

73:47

people.

73:49

So there are other things that we should

73:51

be thinking also but that's so clear at

73:54

the moment. So it's it's the tool we can

73:57

reduce the the amount of people

74:02

producing

74:03

statistics. It's not it's not very it

74:06

does not sound very nice.

74:11

I guess I mean uh in addition

74:15

to well of course you you have

74:17

efficiency gains

74:20

uh through the use of AI but I guess you

74:23

would also have um maybe a better

74:27

quality of your statistical output in in

74:30

principle. Right?

74:33

In principle, yes. But but the the the

74:37

monitoring and all this

74:39

infrastructure, it has to be really good

74:42

that you more you use it more you have

74:45

to evaluate and monitor the results and

74:49

with large language models these things

74:51

are not ready. So the expectations are a

74:54

little bit too high at the moment.

74:58

So I said in some meet in some meeting

75:02

uh last year that when do I expect for

75:05

example this aentic AI is is there in in

75:09

statistical production I asset after 10

75:12

years. So

75:15

it's not there. Yeah it's

75:19

all

75:20

right. Okay. There was just a comment

75:22

from Christina goodness. I'm so glad

75:25

someone just asked that question. So

75:28

confirms this. So last chance to ask a

75:33

question now. Um

75:36

otherwise if there is none so really

75:39

last chance raise your

75:41

hand. Okay, I don't see any. Well, in

75:46

that case, I would like to ask everyone

75:49

if you can please put your camera on for

75:52

a few seconds and uh also maybe your

75:55

microphone and give Rita a round of

75:58

applause. Please join me in doing

76:02

that. Thank you, Rita. Thanks so much.

76:05

Thank you so much, Rita. Appreciate it.

76:07

Thank you, Rita.

76:11

Thank you.

76:13

Thanks so much. Thank you. Uh I just

76:16

want to say one more thing uh before you

76:19

all go. So we will have our next webinar

76:23

also related to AI. It's uh on the topic

76:27

of AI operation operationalization

76:31

MLOPS and LLM ops on the 17th of June.

76:36

Uh so please sign up to for that

76:41

and uh join us next time. I will put the

76:46

registration link in the chat. Thank

76:49

you.

76:51

Thank you. Bye. Thanks Rita. Thanks

76:54

everybody. Yeah, please. Bye.

76:56

Thank you. Thank you. Thank you so much.

77:08

Bye-bye everyone.

Interactive Summary

This presentation outlines a framework for managing artificial intelligence risks through the implementation of guardrails and responsible AI principles, specifically tailored for official statistics producers. It explores the diversity of AI roles—supportive, collaborative, and autonomous—and identifies various risk categories, including technical, ethical, and societal. The speaker emphasizes that adopting AI requires a delicate balance between maximizing benefits and minimizing risks, highlighting the necessity of transparency, accountability, and fairness to maintain public trust and comply with evolving legal frameworks like the EU AI Act.

Suggested questions

5 ready-made prompts