Building trust: Strategies for creating ethical and trustworthy AI systems

IBM Developer
1 Aug 202425:00

Summary

TLDRThe video discusses the importance of AI governance in businesses, highlighting challenges and risks associated with generative AI, such as bias, data privacy, and security. It emphasizes the need for comprehensive governance strategies to ensure legal, ethical, and operational compliance. IBM's Watson Governance platform is introduced as a solution that automates AI lifecycle governance, manages risks, and ensures regulatory compliance. The video also addresses the evolving AI landscape, the necessity of collaboration between technical and non-technical teams, and the role of governance in ensuring trustworthy AI implementation across enterprises.

Takeaways

  • đŸ§‘â€đŸ’Œ IBM emphasizes the need for AI governance in business to address ethical concerns and ensure AI is used safely.
  • 📈 Generative AI could increase global GDP by 7% within 10 years, with 80% of enterprises planning to adopt it.
  • ⚖ Business leaders are concerned about ethical issues like bias, safety, and lack of transparency in generative AI.
  • đŸ’Œ Common generative AI use cases include content generation, summarization, entity recognition, and insight extraction.
  • 🚹 AI risks involve data bias, legal concerns, and the potential for adversarial attacks or misuse during model training and inference.
  • 🧠 IBM's AI governance focuses on lifecycle monitoring, risk management, and regulatory compliance for both predictive and generative models.
  • ⚙ Automating governance processes, such as tracking model performance, metadata, and compliance, is essential for reducing risks and improving efficiency.
  • 🔐 Managing sensitive data and ensuring models handle it responsibly is key to maintaining business trust and meeting regulatory standards.
  • 📊 IBM's governance platform aims to ensure transparency, automate documentation, and monitor AI models continuously throughout their lifecycle.
  • 🚀 IBM's Watson Governance platform provides end-to-end governance for AI, helping organizations balance performance, risk, and compliance across different environments.

Q & A

  • What are the key issues AI introduces at the business level?

    -AI introduces challenges such as ethical concerns, lack of explainability, safety risks, and biases in generative AI. These issues require careful governance to avoid reputational damage, legal risks, and operational inefficiencies.

  • Why is AI governance necessary for businesses?

    -AI governance ensures that AI models are transparent, accountable, and compliant with legal and ethical standards. It helps businesses mitigate risks, ensure AI models remain fair and accurate, and prevent misuse or harm.

  • What are the main use cases of generative AI mentioned in the script?

    -Generative AI use cases include retrieval-augmented generation, summarization, content generation, named entity recognition, insight extraction, and classification.

  • What are the risks associated with the training phase of AI models?

    -Training-phase risks include biases present in the training data, data poisoning attacks, and legal restrictions related to the use of sensitive or copyrighted data.

  • What are adversarial attacks during the inference phase, and how can they affect AI models?

    -Adversarial attacks, such as evasion or prompt injection, occur when attackers manipulate input during the inference phase to produce harmful or biased outputs, compromising the AI model’s reliability.

  • What are some real-world cases illustrating AI model risks?

    -Examples include a dealership's AI bot mistakenly selling a Chevy Tahoe for $1 and Microsoft's Twitter chatbot turning offensive due to learning inappropriate behavior from user interactions.

  • How does IBM's Watson Governance platform address AI governance needs?

    -The Watson Governance platform automates life cycle management, risk governance, and regulatory compliance for AI models. It helps businesses ensure model accuracy, fairness, and transparency across development and deployment.

  • What are the three critical capabilities identified by IBM for AI governance?

    -The three capabilities are monitoring and evaluating models, tracking facts and metrics, and managing the life cycle and risks of AI models.

  • What is 'prompt governance,' and why is it important for foundation models?

    -Prompt governance involves tracking and evaluating text-based instructions (prompts) used with foundation models. It is essential to ensure that prompts are properly managed, evaluated for quality, and monitored for safety to avoid generating harmful content.

  • What role does monitoring model performance play in AI governance?

    -Monitoring model performance ensures that AI models and prompts remain accurate, efficient, and safe over time. It helps detect issues like performance degradation, data drift, and the presence of toxic language or personal information in outputs.

Outlines

00:00

đŸ€– Introduction to AI Governance and Its Importance

Igor PV, an AI engineer at IBM, introduces the topics of AI governance, discussing the impact of AI, especially generative AI, on business. He highlights the rapid adoption of AI technologies, noting that 80% of enterprises are either working with or planning to use foundation models. The business leaders’ concerns around safety, ethical issues, and biases in AI systems are underscored. Igor discusses the common use cases for generative AI, such as content generation and summarization, while acknowledging the inherent risks in integrating these models into organizations.

05:01

🚗 Real-World Risks in AI Systems

This section presents real-world cases illustrating the risks in AI, such as a dealership's AI bot being tricked into selling a Chevy Tahoe for $1 and Amazon Alexa mistakenly ordering a dollhouse. It also covers Microsoft's Twitter chatbot, which turned offensive due to its training data. These examples underscore the dangers AI poses to reputation, legal compliance, and operational integrity if not properly governed. Igor categorizes AI risks into three main buckets: regulatory, reputational, and operational.

10:03

📊 The Need for Comprehensive AI Governance

Igor emphasizes the importance of AI governance to manage AI projects effectively and ensure trustworthiness in their deployment. He introduces IBM’s approach, detailing three critical capabilities needed for AI governance: monitoring model accuracy, tracking key metrics, and managing the entire AI lifecycle. He also points out the challenge companies face in recruiting AI talent and the lack of standardized best practices for AI governance. He stresses that AI governance must be adaptable to different regulatory and organizational needs.

15:04

đŸ› ïž Watson Governance Platform for AI Management

IBM's Watson Governance platform is introduced as a solution for AI lifecycle governance. It provides tools to automate and streamline processes from model development to deployment, including model approval, risk assessment, and prompt governance. The platform offers full configurability to handle both traditional machine learning models and generative AI, enabling better collaboration between technical and non-technical stakeholders. Igor outlines how the platform ensures compliance with regulatory standards while improving business outcomes and reducing governance costs.

20:04

📈 Monitoring and Managing AI and LLM Performance

Igor highlights the capabilities of Watson’s governance platform in tracking and improving the performance of large language models (LLMs). The platform automates monitoring for technical bottlenecks, prompt quality, and safety concerns such as toxic language. It also helps track performance drift, ensuring models remain effective over time. This automation reduces manual efforts in governance, ensuring businesses maintain accurate, reliable, and compliant AI systems.

Mindmap

Keywords

💡AI Governance

AI Governance refers to the process of directing, monitoring, and managing the AI activities of an organization. It is essential for ensuring AI models are used ethically, safely, and in compliance with regulations. In the video, AI governance is emphasized as a necessary measure to mitigate risks such as bias, legal issues, and reputational damage when using generative AI models.

💡Generative AI

Generative AI is a subset of artificial intelligence that creates new content, such as text, images, or music, based on input data. The video highlights its potential, showing that 80% of enterprises are either using or planning to use generative AI, which could raise global GDP by 7%. However, the risks and challenges it poses, such as bias and misuse, are also discussed.

💡Foundation Models

Foundation models are large, multi-purpose AI models trained on massive datasets, capable of being fine-tuned for specific tasks. The video explains that 80% of enterprises are working with or planning to use these models, but it also discusses the governance challenges associated with ensuring these models are properly evaluated, tracked, and monitored throughout their lifecycle.

💡Risk Management

Risk Management in AI involves identifying, quantifying, and mitigating the potential risks associated with AI models. The video explains that AI poses several risks, including data poisoning, bias, and adversarial attacks. It also highlights the need for risk management tools within AI governance frameworks to ensure safe, trustworthy AI implementations.

💡Bias

Bias in AI refers to the tendency of AI models to produce skewed or unfair results due to biased data used in training. The video warns that biased outcomes in AI can lead to significant reputational and legal consequences for organizations. Bias is a core reason why AI governance and risk management are critical to AI deployment.

💡Regulatory Compliance

Regulatory Compliance is the adherence to laws and regulations governing AI use, particularly around data privacy, fairness, and ethical concerns. The video notes that AI regulations are increasing globally, and companies need to ensure their AI models are compliant to avoid legal and financial penalties. A robust governance system helps manage this.

💡Prompt Injection

Prompt Injection refers to a type of adversarial attack in generative AI where malicious inputs are designed to manipulate the model’s output. The video highlights this as a risk in the inference phase, where attackers can craft inputs to guide the AI to specific, often harmful, outputs.

💡Model Drift

Model Drift occurs when the performance of an AI model degrades over time due to changes in the environment or data. In the video, this is highlighted as a key issue that AI governance platforms must monitor, as drifting models can lead to inaccurate predictions and business risks. Continuous monitoring helps catch these changes early.

💡Explainability

Explainability in AI refers to the ability to understand and interpret the decision-making process of AI models. The video emphasizes the importance of explainability, particularly in generative AI, where transparency is key to addressing ethical concerns, building trust, and ensuring regulatory compliance.

💡Lifecycle Governance

Lifecycle Governance is the process of managing an AI model from its development to its deployment and beyond, ensuring continuous monitoring and compliance. The video underscores that AI governance must cover the entire lifecycle, from resource approval, data management, and testing, to ongoing performance monitoring, to ensure that AI models remain fair, accurate, and compliant over time.

Highlights

Discussion on AI governance and its critical importance at the business level.

80% of enterprises are working with or planning to leverage foundation models and generative AI.

Generative AI could raise global GDP by 7% within 10 years.

Business leaders are worried about ethical concerns like explainability, safety, and bias in generative AI.

Common generative AI use cases include summarization, content generation, and insight extraction.

Risks of using generative AI include biased training data, data poisoning, and legal restrictions.

AI systems can be vulnerable to adversarial attacks like prompt injection, leading to inaccurate outputs.

Real-world AI failures, such as Microsoft’s Twitter chatbot turning offensive and a Chevy Tahoe being sold for $1 due to an AI error.

AI governance is needed to mitigate legal, reputational, and operational risks for businesses.

IBM’s governance platform focuses on monitoring, tracking, and managing AI models across their lifecycle.

Three key capabilities for AI governance: lifecycle management, risk management, and regulatory compliance.

Governance must involve both technical and non-technical members to ensure informed decisions.

IBM’s Watson Governance solution automates governance processes to reduce workload and improve transparency.

The governance solution supports both predictive and generative AI models across different environments.

The platform offers real-time performance monitoring tools, ensuring the safety and effectiveness of deployed AI models.

Transcripts

play00:00

[Music]

play00:03

my name is Igor PV I am AI engineer in

play00:06

client engineering team of IBM so in the

play00:10

first part of our meeting we will

play00:12

discuss the

play00:13

issues uh AI introduces in the business

play00:16

level uh why I governance is necessary

play00:19

and how it can be implemented and in

play00:22

second part we will conduct a Hands-On

play00:25

lab to demonstrate how this process

play00:27

might look like

play00:32

CH GPT created significant interest

play00:34

around the notion of large language

play00:36

models uh while chpt technology has

play00:39

found a home in some consumer and

play00:42

business applications large language

play00:44

models are only part of a broader

play00:47

discussion about using AI technology to

play00:49

produce business results this slide

play00:52

details the speed scope and scale of the

play00:55

impact Genera AI will have on the market

play00:59

80% of Enterprises are working with or

play01:02

planning to leverage Foundation models

play01:04

and adopt generative AI generative AI

play01:07

could raise Global GDP by 7% within 10

play01:11

years and generative AI expected to

play01:14

represent 30% of overall Market by

play01:19

2025 Business Leaders struggle to grow

play01:22

AI in their companies safely 80% of them

play01:25

are seriously worried about at least one

play01:27

ethical problem like that generative fi

play01:31

are not sufficiently explainable or

play01:33

about safety and ethical aspects of

play01:36

generative AI some concerns about

play01:39

established biases in generative AI or

play01:42

someone simply doesn't trust

play01:46

it let's remember the most common

play01:49

generative a use cases it can be

play01:51

retrieval augmented generation

play01:54

summarization content generation named

play01:58

entity recognition

play02:00

Insight extraction or

play02:03

classification all these use cases can

play02:06

be affected by risks which will we will

play02:09

discuss on the next

play02:12

slide many risks essential in using

play02:15

generative models organizations failing

play02:18

to address this when inte integrating

play02:21

generative AI they can face significant

play02:24

damage to their public reputations as

play02:27

well as legal and Regulatory penalties

play02:31

there are risks associated with input on

play02:33

training phase bias present in the

play02:37

training data can often lead to biased

play02:40

outcomes data poisoning attacks take

play02:42

place when malicious users try and take

play02:45

advantage of the iterative training

play02:47

features of some large language models

play02:50

by feeding them toxic or hateful

play02:54

content legal restrictions on data must

play02:57

be considered when training the llm as

play03:00

the owner of the model could be

play03:02

responsible for copyright and

play03:04

intellectual property infringement and

play03:06

improper use of personal information and

play03:09

sensitive personal

play03:11

information there are some risks

play03:13

associated with input on inference phas

play03:16

disclosure of personal uh information

play03:19

Central personal information or

play03:21

copyright information may occur during

play03:23

the inference phase in which the model

play03:26

developers ask the model to generate

play03:28

content based on unseen

play03:31

information or adversarial attacks like

play03:33

evasion prompt injection or others in

play03:37

the infinite phase can include not only

play03:41

data poisoning but attempts to guide the

play03:43

model to certain types of output by

play03:46

tailoring the questions asked during

play03:48

this

play03:49

stage and some risks associated with

play03:51

output many of the risks present during

play03:54

model training are also relevant to

play03:57

model output model train on biased

play04:00

material May reflect that in their

play04:02

answers which can cause both

play04:05

reputational damage for the model owner

play04:07

and lead to legal action personal

play04:10

information and Central personal

play04:11

information and copyrighted materials

play04:14

present in the training data may also

play04:16

either appear directly in the output or

play04:19

clearly influence it performance

play04:22

disparity in which the quality of the

play04:24

model results May not meet certain stats

play04:27

making the model unusable or misuse in

play04:31

which the model is used to perform

play04:32

unethical tasks or value alignment

play04:35

issues such as hallucination in which

play04:38

model presents effectually incorrect

play04:40

answer at the truth and a lot of other

play04:43

risks that should be

play04:46

considered here are some real cases that

play04:49

related to risks that we discussed on

play04:51

previous slide case when 2024 shev V

play04:55

Tahoe was sold only for $1 a

play04:58

California's dealership AI bot

play05:00

programmed to agree with all customer

play05:02

statements was exploited by a driver who

play05:05

convinced it to sell a Chevy Taha only

play05:08

for $1 the bot confirmed this that this

play05:12

is a legally binding deal due to a

play05:14

mistake in how it was

play05:17

programmed or when child ordered a

play05:20

dollhouse and cookies on Amazon parents

play05:22

were shocked when an expensive dollhouse

play05:24

and cookies were mistakenly ordered by

play05:27

their daughter talking to Amazon

play05:30

Alexa or case with Microsoft Twitter

play05:34

chatbot when it turned vulgar and

play05:36

offensive Microsoft launched its AI

play05:39

Twitter chatbot called T designed to

play05:42

learn from interactions to enhance its

play05:45

convention conversational

play05:47

capabilities however they quickly

play05:50

started mimicking the offensive and

play05:52

appropriate inappropriate language from

play05:55

user tweets leading to a rapat

play05:57

degradation into a vulgar and offensive

play06:00

online

play06:02

presence risks in AI models can come

play06:06

from many

play06:08

sources uh like the ones shown on this

play06:11

slide some related to Classic machine

play06:13

learning models and better understood

play06:15

some uner concerns rising from

play06:18

generative AI there are three buckets

play06:21

that can help categorize this risks

play06:24

first one is

play06:25

regulatory with AI regulation

play06:28

progressing in many parts of the world

play06:30

organizations that do not have this

play06:33

under control risk big

play06:36

finds second is reputational even if

play06:39

something is legal organizations must

play06:42

consider whether they want to end up in

play06:44

the news as an example of AI gun bad for

play06:48

example what happens if the output from

play06:50

AI violates social norms by generating

play06:54

offensive or suggestive

play06:56

content and third bucket is operational

play07:00

many AI projects do not make it into

play07:02

production due to lack of trust in the

play07:05

technology robbing their organizations

play07:08

of the potential benefits of the

play07:12

solution so as we saw AI needs

play07:16

governance the process of directing

play07:19

monitoring and managing the AI

play07:22

activities of an

play07:26

organization IBM has identified three

play07:29

critical iCal capabilities necessary for

play07:31

a proper AI governance solution first

play07:35

Monitor and

play07:36

evaluate governance platform should

play07:39

monitor AI models to ensure they remain

play07:42

accurate and fair it should oversee

play07:45

predictive models and ensure generative

play07:47

models handle sensitive data

play07:49

responsibly also it would provide clear

play07:53

explanations of how these models make

play07:56

their decisions helping everyone

play07:59

understand

play08:00

their

play08:01

processes secondly track facts and

play08:05

metrics platform should automatically

play08:07

collect and organize all important model

play08:10

data making it easily accessible and

play08:14

searchable this ensures transparency and

play08:17

accountability from the model's

play08:19

Creations through its

play08:22

deployment and lastly manage life cycle

play08:25

and risk platform should allow

play08:27

customization of the AI model

play08:30

development and deployment process it

play08:32

would track every aspect to minimize

play08:34

risk and offer realtime performance

play08:37

monitoring tools this ensures safe and

play08:41

effective Management in summary this

play08:44

envisioned capabilities would help

play08:46

manage track and govern AI applications

play08:49

effectively ensuring a reliable and

play08:52

trustworthy AI operation within any

play08:58

organization but AI governance is

play09:01

complicated governing a rapidly evolving

play09:04

field such as AI has always presented

play09:07

problems which will only become more

play09:10

difficult as organizations seek to

play09:12

incorporate new generative models in

play09:15

addition to more traditional predictive

play09:17

models they already have today companies

play09:21

struggle to find enough qualified data

play09:24

Engineers data scientists and the AI

play09:27

Engineers just to develop and test new

play09:29

moduls let alone perform manual tasks

play09:33

like tracking and documenting the

play09:35

changes in training data or performing

play09:38

runtime analysis and additional

play09:40

development on models and

play09:42

production what's more there are no

play09:45

standard best practices or tools for

play09:48

deployment or governance which results

play09:50

in companies using highly fragmented

play09:54

difficult to govern development and

play09:56

deployment

play09:57

environments while open Source Solutions

play10:00

do exist many of the open source

play10:02

governance Frameworks are primarily

play10:04

aimed at data scientists and other

play10:07

coders making them challenging to

play10:10

understand for non-technical

play10:12

stakeholders and blocking collaboration

play10:15

between technical experts and subject

play10:17

matter

play10:19

experts finally the governance needs uh

play10:22

of each organization can vary wildly

play10:25

depending on the industries and

play10:27

countries in which they operate and the

play10:30

regulatory standards they must meet an

play10:33

AI governance solution must be able to

play10:35

automate the routine tracking tasks

play10:38

while being flexible enough to deal with

play10:40

multiple Environ environments for

play10:43

development and deployment it must also

play10:46

be fully configurable for different

play10:48

regulatory standards and approval

play10:50

workflows and allow Technical and

play10:53

nontechnical stakeholders to collaborate

play10:56

together as easily as possible

play11:01

so IBM introduces what's next governance

play11:05

an endtoend automated life cycle

play11:08

governance toolkit it is a single

play11:11

automated configurable platform for

play11:14

collaboratively managing and monitoring

play11:17

predictive and generative AI

play11:20

models this platform helps to build

play11:24

enduring consumer trust with your brand

play11:26

boost productivity and accelerate

play11:28

business outcomes and mitigate risk and

play11:31

minimize cost of

play11:34

compliance the whatson governance

play11:36

platform handles three key aspects

play11:38

necessary for y governance throughout

play11:41

the entire model life cycle and across

play11:43

the entire

play11:45

Enterprise first is life cycle

play11:48

governance AI governance involves the

play11:51

entire organization not just the data

play11:54

science department it covers from

play11:57

initial model request to deployment

play12:00

including stages like resource approval

play12:03

data management and testing effective

play12:07

governance requires the involvement of

play12:09

both again Technical and non-technical

play12:12

members and aims to automate processes

play12:15

to reduce data science workload it also

play12:19

ensures decision makers have the

play12:21

necessary data to make informed

play12:24

decisions W's governance faciliates this

play12:27

by tracking and cataloging metadata like

play12:29

training data and model evaluations

play12:33

providing a complete overview of models

play12:37

deployment uh and development

play12:40

performance second is risk management

play12:43

before organizations can trust AI to

play12:45

make business decisions or interact with

play12:47

customers they must understand and

play12:50

quantify the risks that AI presents and

play12:54

be able to measure the AI performance to

play12:57

monitor their ref

play12:59

exposure and third one is Regulatory

play13:02

Compliance increasing government

play13:05

regulation of AI presents serious

play13:08

problems for organizations hoping to

play13:10

adapt AI without a comprehensive

play13:13

configurable governance

play13:15

system the wsn governance solution

play13:18

allows Enterprises to track their models

play13:21

against regulatory standards in areas

play13:23

such as accuracy and

play13:26

fairness it also provides the ability to

play13:28

EXP explain decisions and automatically

play13:31

collect metadata so Auditors can

play13:34

determine how modules were trained and

play13:36

why they generated that output finally

play13:40

what governance allows for governance of

play13:42

all predictive and generative models

play13:45

regardless of deployment

play13:49

platform what governance provides three

play13:52

main capabilities shown in blue blocks

play13:56

AI documentation AI risk governance and

play13:59

AI evaluation and monitoring that work

play14:02

together with different AI Stacks which

play14:05

you will see soon in white

play14:08

blocks as an example of endtoend govern

play14:11

process first is a department identifies

play14:14

a business challenge solvable by AI

play14:17

initiating a new use

play14:19

case the AI use case under goes approval

play14:22

with model document documentation being

play14:25

developed and updated in

play14:28

sync during model op Pro development all

play14:31

metadata is automatically captured and

play14:34

updated using tools from both popular

play14:37

open source Frameworks and whatson

play14:40

XI custom metadata tracking is also

play14:44

supported the model's preproduction

play14:46

evaluation captures performance data

play14:48

leading to production

play14:50

approval in the preferred platform the

play14:53

model is deployed and once again the

play14:55

relevant metadata is captured and synced

play14:59

and lastly the production model is

play15:01

continuously monitored and the

play15:03

performance data captured and synced as

play15:06

well and the model owner keeps an eye on

play15:09

the performance metrics in their

play15:13

dashboard here is example of endtoend

play15:16

life cycle for foundation

play15:18

model first is model approval with

play15:22

Foundation models organizations will

play15:24

need to evaluate and approve those

play15:27

multi-purpose models before it is used

play15:30

in any use case a foundation model is

play15:34

Upstream from the use

play15:36

case second is use case approval in this

play15:41

step a lot will stay the same when

play15:43

compared with a use case with a

play15:45

traditional model you will still need to

play15:48

have an accountable owner of the use

play15:50

case describe the purpose do risk

play15:52

assessment and decide on the appropriate

play15:55

risk controls and

play15:57

metrics however with the foundation

play16:00

model you now also need to specify what

play16:03

tasks are required to deliver the use

play16:05

case is it simiz or a classification or

play16:09

some other

play16:11

one model

play16:14

selection this is a new step with

play16:16

Foundation models organizations will

play16:19

have many Foundation models approved to

play16:21

be used including the tasks that are

play16:24

allowed in this step user will make a

play16:27

match between tasks allowed for Approved

play16:29

models and tasks required for the use

play16:33

case if multiple models are available

play16:36

users will be able to make right size

play16:39

tradeoff decisions considering quality

play16:42

cost energy consumption

play16:46

Etc model fine tuning when you fine

play16:50

tuning model you change the weights of

play16:52

the model it won't be always necessary

play16:55

to do this but if you choose to do so if

play16:58

finetune model should be considered a

play17:01

new object that is distinct from the

play17:04

base model from which it is derived it

play17:07

should be gared in its own right in

play17:09

addition to the based

play17:12

model prom development with prompts as

play17:16

the primary way to interact with

play17:18

Foundation models organizations need to

play17:21

add prompt governance to their

play17:23

repertoire in addition to model

play17:26

governance at the development step this

play17:29

means capturing the prompt metadata that

play17:32

you need for your governance activities

play17:34

including the new model parameters

play17:37

described

play17:39

earlier evaluation and monitoring as

play17:42

mentioned earlier many of the tasks

play17:44

supported by Foundation models come with

play17:47

new metrics and explainability methods

play17:50

organizations should look to adapt this

play17:53

into their Frameworks and

play17:55

projects and last one is change request

play17:59

with a traditional Model A change

play18:01

request almost automatically meant a

play18:04

retraining of your use case specific

play18:07

model but with the foundation model

play18:10

there are several things that a change

play18:12

request could relate to it can be

play18:15

changing the model selection as new

play18:17

models are developed and approved there

play18:20

might be a better choice for your use

play18:23

case another option is train or retrain

play18:26

a finetune model maybe you had some

play18:29

initial success with the base model and

play18:31

now looking to improve on that by

play18:34

fine-tuning model with some of your own

play18:36

data or business has changed and you

play18:39

want to update your finetune model and

play18:42

another option is adjust your proms of

play18:45

course all steps mentioned here are

play18:48

covered by whats governance platform as

play18:52

we will see in the next slides or in the

play18:56

lab let's review what's next governance

play18:59

capabilities again most of them we will

play19:02

see soon and try in the

play19:05

lab model risk governance is managing

play19:09

and automating the activities around

play19:11

attestation review validation change

play19:14

management and issue resolution of youri

play19:17

models it is important for meeting the

play19:20

compliance requirements of model Focus

play19:22

regulations across regions and

play19:25

authorities and for reducing governance

play19:27

costs

play19:29

what's governance brings together all

play19:31

stakeholders in one process with clear

play19:34

roles and

play19:36

responsibilities combines a flexible

play19:38

data model with workflow calculation

play19:41

questionary and business intelligence

play19:46

capabilities AI documentation is

play19:49

tracking the life cycle of your models

play19:51

and prompts from credle to grave and

play19:54

fact sheets view for AI assets that

play19:57

track lineage events and facilate

play20:00

efficient model Ops governance it can

play20:03

reduce manual efforts to document models

play20:06

and prompts and increase transparency of

play20:09

models to do it we capture facts about

play20:12

use cases models and prompts throughout

play20:15

the model life cycle a toog from

play20:19

common python Frameworks and from

play20:21

whatson C prompt lab extend it with

play20:25

custom facts capture attachments

play20:28

automated reporting for different

play20:31

stakeholders design time evaluation of

play20:34

llm prom templates as AI Engineers are

play20:38

creating their prompts they can evaluate

play20:41

them directly from within their

play20:43

development environment it is important

play20:45

to identify and mitigate prompt issues

play20:48

as early as possible in the process and

play20:52

it will save time with integration into

play20:54

development environment to achieve that

play20:58

prompt lab gets the evaluate option

play21:00

directly in its UI matrics automatically

play21:04

selected based on task type like test

play21:07

text summerization Tech content

play21:09

generation extraction

play21:12

Q&A and uh you can test prompts for

play21:16

quality and safety metrics such as toxic

play21:18

language and personal identifiable

play21:22

information monitoring model health is

play21:25

to track performance metric for model

play21:28

and prompt inferencing such as number of

play21:30

Records number of tokens payload size

play21:34

latency and throughput it helps to

play21:37

identify technical bottlenecks in

play21:39

predictive model and llm inferencing

play21:43

imagine if at some moment your perfect

play21:45

model is starting to answer 10x times

play21:48

longer of course you would like to catch

play21:51

this changing as soon as possible and it

play21:54

is reducing effort of monitoring

play21:56

technical workings of deployed models

play22:00

governance monitors and reports on

play22:03

number of Records number of scoring re

play22:06

requests throughput and latency number

play22:08

of users amount of input and output

play22:11

tokens for LMS and payload size for

play22:14

predictive

play22:16

models monitoring Genera equality is

play22:20

monitoring how well your llm prompts

play22:23

perform it is important to maintain the

play22:26

business benefits of your deployed

play22:28

prompts

play22:29

and it can help to reduce manual efforts

play22:31

to track model

play22:33

performance you can consider this model

play22:35

quality Matrix as a blood test with

play22:38

defined normal thresholds and actual

play22:41

values to do it we monitor deployed

play22:44

promts metric automatically selected uh

play22:47

based on task type this metrics can be

play22:50

like Rouge meture similarity or

play22:54

others and we are still monitoring for

play22:56

quality drift and safety metrics again

play22:59

such as toxic language and the personal

play23:02

identifiable

play23:04

information monitoring drift drift

play23:08

occurs when over time the production

play23:11

data or outcomes start differing from

play23:15

the time of training and testing the

play23:17

model imagine if your your well set

play23:21

prompt and model are starting to provide

play23:25

not so accurate results as it was before

play23:28

on the training stage or testing you

play23:30

definitely would like to see this Drift

play23:33

We need it because loss of accuracy has

play23:36

a negative impact on business outcomes

play23:39

to reduce manual efforts to detect drift

play23:42

and uh because drift drift shows changes

play23:47

in real world Behavior you might

play23:49

otherwise not recognize it to catch

play23:52

drift will automatically monitor

play23:54

production data for different types of

play23:56

it like uh output drift model quality

play24:00

drift feature drift or input and output

play24:03

metadata

play24:06

drifts governing large language model

play24:09

prompts prompts as we remember are text

play24:13

based instructions to Foundation models

play24:15

such as LMS they are created and used

play24:19

differently than traditional predictive

play24:21

machine learning models why we need it

play24:25

to govern both types of assets on one

play24:28

platform it leads to reduce cost and

play24:31

effort having consistent rules and

play24:33

methods applies to both types and

play24:36

reducing cost of compliance through

play24:38

automation WN gance tracks prompts

play24:42

throughout their life cycle in an AI use

play24:45

case automatically captures prompt

play24:48

metadata evaluates prompt during prompt

play24:50

design and monitor prompts when deployed

play24:53

into production

Rate This
★
★
★
★
★

5.0 / 5 (0 votes)

Étiquettes Connexes
AI governanceGenerative AIRisk managementAI complianceBusiness AIEthical AIAI monitoringFoundation modelsData securityAI deployment
Besoin d'un résumé en anglais ?