Humans vs. AI: Who should make the decision?

IBM Technology
15 Feb 202208:57

Summary

TLDRThe video script explores the dynamic between human and artificial intelligence in decision-making, particularly in fraud detection. It illustrates how AI excels with high confidence predictions, while humans shine when AI is uncertain. The script advocates for augmented intelligence, combining human judgment with AI insights, but cautions about human biases like automation bias and the reluctance to trust AI when it admits fallibility. The key takeaway is that the most effective decision-making often lies in the collaboration between humans and AI, when carefully managed to minimize cognitive biases.

Takeaways

  • 🤖 The decision-making process can be a combination of human intuition and AI analysis, each with their own strengths and weaknesses.
  • 📊 AI excels in high-confidence predictions, providing high success rates when it is certain about an outcome, such as in fraud detection systems.
  • 🧐 Humans tend to outperform AI when the AI's confidence is low, often due to their ability to bring in additional context and information.
  • 📉 Performance curves for AI and humans differ, with AI showing a steeper curve correlating high confidence with high accuracy.
  • 🤝 Augmented intelligence, which combines human and AI decision-making, can offer the highest success rates for certain confidence levels.
  • 💡 The effectiveness of augmented intelligence is influenced by how AI recommendations are presented to human decision-makers.
  • 👀 Forced display of AI recommendations can lead to automation bias, where humans may overly rely on AI suggestions.
  • 🙌 Optional display, where AI recommendations are only shown upon request, can help mitigate automation bias and encourage independent human judgment.
  • 🔢 Providing an accuracy percentage with AI recommendations can affect human trust and acceptance of the AI's advice.
  • 🧐 Humans may be less likely to incorporate AI recommendations if they are explicitly told there is a chance of being wrong.
  • 🔑 Understanding the strengths of both AI and human decision-making can lead to more effective outcomes when combined in an augmented intelligence approach.
  • 📚 The script emphasizes the importance of considering human cognitive biases in the design of AI-assisted decision-making systems.

Q & A

  • What is the main topic of the video script?

    -The main topic of the video script is the decision-making process, particularly the comparison between human decision-making and artificial intelligence (AI), and how they can be combined for optimal results in tasks such as fraud detection.

  • Why are financial analysts overwhelmed with alerts in a fraud detection system?

    -Financial analysts are overwhelmed because 90 percent of the thousands of alerts generated each day are false positives, making it difficult to focus on the actual fraudulent transactions.

  • How does the video script describe the typical AI performance curve in terms of success rate and confidence score?

    -The script describes the AI performance curve as having high success rates at very low and very high confidence scores, indicating the AI is certain about its predictions. However, at moderate confidence levels, the success rate drops, showing the AI is unsure.

  • How does human performance compare to AI performance in the script?

    -Human performance curves are typically flatter than AI, meaning humans may not be as accurate as a confident AI but can outperform AI when the AI is unsure, especially in complex or statistically rare cases.

  • What is the term used to describe the combination of human decision-making aided by AI?

    -The term used to describe the combination of human decision-making aided by AI is 'Augmented Intelligence'.

  • Why is augmented intelligence considered to have the highest success rate for some confidence scores?

    -Augmented intelligence has the highest success rate for some confidence scores because it leverages both human judgment and AI analysis, particularly in cases where the AI's confidence is not very high or very low.

  • What cognitive bias is mentioned in the script that can affect the effectiveness of AI recommendations?

    -The script mentions 'automation bias', which is the tendency for humans to favor suggestions from automated systems and ignore contradictory information.

  • What are the two display methods for AI recommendations mentioned in the script, and how do they differ?

    -The two display methods are 'forced display', which shows the AI recommendation simultaneously with the decision case, and 'optional display', which only shows the AI recommendation when requested by the human decision maker.

  • How does the accuracy percentage of an AI recommendation affect human decision-making?

    -When an AI recommendation is accompanied by an accuracy percentage, humans are less likely to incorporate the recommendation into their decision, as they may not trust or like the idea that the AI might be wrong.

  • What does the script suggest as the best approach to decision-making in complex tasks?

    -The script suggests that the best approach to decision-making in complex tasks is a combination of AI and human input, known as augmented intelligence, while being mindful of human cognitive biases.

  • What is the final message of the video script regarding the collaboration between humans and AI?

    -The final message is that humans and AI algorithms can form a powerful team to improve decision-making outcomes, provided that we understand and leverage their respective strengths and account for potential biases.

Outlines

00:00

🤖 AI vs. Human Decision Making in Fraud Detection

The script discusses the dilemma of whether a decision should be made by a human or an artificial intelligence (AI). It uses the example of a fraud detection system to illustrate the strengths and weaknesses of both. The AI's performance is typically high when it is confident but lower when unsure, while humans may outperform AI when the AI's confidence is at a 50 percent level, especially in complex or rare cases. The script introduces the concept of 'augmented intelligence,' which combines human decision-making with AI assistance, potentially leading to the highest success rate for certain confidence scores.

05:05

🧠 Overcoming Cognitive Bias in Augmented Intelligence

This paragraph delves into the importance of considering human cognitive bias when implementing augmented intelligence. It contrasts 'forced display' and 'optional display' of AI recommendations and explains how they influence human decision-making. Forced display can lead to automation bias, where humans may overly rely on AI suggestions, while optional display allows humans to form their own impressions before considering AI input. The paragraph also touches on the impact of trust and accuracy percentages on human acceptance of AI recommendations, emphasizing the need to present AI augmentation effectively to enhance decision-making outcomes.

Mindmap

Keywords

💡Decision Making

Decision making is the cognitive process of selecting a course of action from among multiple alternatives. In the video's context, it refers to the choice between a human or an AI making a decision. The script discusses how different tasks may be better suited for either a human or an AI, highlighting the importance of understanding the strengths and limitations of each in the decision-making process.

💡Artificial Intelligence (AI)

Artificial Intelligence refers to the simulation of human intelligence in machines that are programmed to think like humans and mimic their actions. The video script discusses AI's role in decision-making tasks, particularly in the example of a fraud detection system where AI can handle a high volume of alerts more efficiently than humans.

💡Fraud Detection

Fraud detection involves the identification of potentially fraudulent activities, often using algorithms and data analysis. The script uses fraud detection as a practical example where AI can assist in reducing the workload of financial analysts by identifying false positives among numerous alerts.

💡Confidence Score

A confidence score in the context of AI refers to the level of certainty the system has in its predictions or decisions. The video explains how different confidence scores on a scale from 0 to 100 percent can affect the success rate of identifying real alerts versus false positives in a fraud detection system.

💡Human Bias

Human bias refers to the systematic errors in human judgment and decision-making caused by our cognitive inclinations. The script mentions that the way AI recommendations are presented to humans can influence their decisions, potentially leading to automation bias where humans overly trust AI suggestions.

💡Performance Curve

A performance curve in the video represents the relationship between the AI's confidence score and its success rate in predicting the correctness of an alert. It illustrates how AI performance varies with different levels of confidence, showing higher success rates at very high or very low confidence scores.

💡Augmented Intelligence

Augmented intelligence is a concept where human intelligence is enhanced by AI. The video suggests that combining human decision-making with AI assistance can lead to the highest success rates, especially in cases where the AI's confidence score is moderate.

💡Forced Display

Forced display is a method where AI recommendations are automatically shown alongside decision cases for immediate consideration. The script warns that this can lead to automation bias, where human decision-makers might favor the AI's suggestions over their own judgment.

💡Optional Display

Optional display is an approach where AI recommendations are only provided upon request by the human decision-maker. The video script explains that this can help overcome automation bias by allowing humans to form their own initial impressions before considering AI input.

💡Accuracy Percentage

Accuracy percentage in the context of AI refers to the estimated likelihood that a prediction will be correct. The script notes that when AI recommendations include an accuracy percentage, humans may be less likely to incorporate the AI's advice into their decisions, reflecting a distrust of imperfect recommendations.

💡Cognitive Bias

Cognitive bias is a pattern of deviation in judgment that occurs as a result of the way our minds process information. The video discusses how human cognitive biases, such as automation bias and distrust of imperfect recommendations, can affect the effectiveness of augmented intelligence in decision-making.

Highlights

The debate on whether a human or AI should make a decision is explored, emphasizing the strengths and limitations of both.

AI outperforms humans in tasks with high statistical certainty, while humans excel in complex or rare cases.

Fraud detection is used as a case study to illustrate the decision-making process involving AI and human analysts.

The concept of a performance curve is introduced to visualize AI and human success rates in decision-making.

AI algorithms are highly performant when confident but less so when uncertain, unlike humans who may outperform AI in unsure situations.

Humans can bring additional context and information to decisions, unlike AI which sticks to its decision logic.

Augmented intelligence, a combination of human and AI decision-making, is proposed as optimal for certain scenarios.

The success rate of augmented intelligence is highest for moderate confidence scores in predictions.

Human cognitive bias, such as automation bias, can affect the effectiveness of AI-assisted decision-making.

Forced display of AI recommendations can lead to automation bias, where humans favor AI suggestions over their own judgment.

Optional display of AI recommendations allows humans to form their own impressions before considering AI input.

The presentation of AI recommendations, including accuracy percentages, influences human trust and decision-making.

The importance of minimizing human cognitive bias in the decision-making process when using augmented intelligence is emphasized.

AI and human collaboration can lead to improved decision-making outcomes when the right balance is achieved.

The transcript concludes by suggesting that understanding who to ask is key to leveraging the strengths of both AI and humans.

The video invites viewers to engage with the content through questions and subscriptions for more informative content.

Transcripts

play00:00

A decision needs to be made.

play00:05

But who should make it?

play00:08

Me, a human, ... or an artificial intelligence, an AI?

play00:21

We've discussed before that humans can outperform AI at some tasks,

play00:26

but that, statistically, AI will make a better job of deciding for other tasks.

play00:30

So for one single decision, who should decide?

play00:35

Well, the answer is a fascinating combination of holistic curves and human bias.

play00:40

Let's get into it.

play00:42

So, consider a fraud detection system.

play00:46

Fraud detection.

play00:52

The system generates the alerts of potentially fraudulent transactions.

play00:57

Financial analysts review each alert.

play01:01

Now, there's thousands of events generated each day,

play01:03

and the analysts are overwhelmed with 90 percent of those alerts being false positives.

play01:09

An AI system could help alleviate the workload.

play01:13

But which alerts should the AI handle, and which should be processed by a skilled financial analyst?

play01:21

Well, let's draw a graph to answer the question, "Is this a real alert?"

play01:39

So, let's draw a graph with an X and Y axis.

play01:45

The Y axis tracks the success rate.

play01:52

So an alert comes in, we make a prediction as to if it is real or not,

play01:58

and we track if that prediction turned out to be right.

play02:03

Along the X axis is the confidence score.

play02:10

So a confidence score of zero percent

play02:14

says a prediction thinks that this is definitely not a real alert, it's a false positive.

play02:21

A confidence score of 100 percent

play02:25

means that a prediction is certain that it is a real alert.

play02:31

Now a typical AI performance curve will look something like this.

play02:43

So we've got very low confidence scores, this is not a real alert,

play02:48

and very high confidence scores, this is a real alert.

play02:53

They're correlated to a high success rate.

play02:57

That's these areas up here.

play02:59

When the AI is not sure about a given prediction, then it's not such a case.

play03:06

Lower success rate when the AI is not sure.

play03:10

And so effectively the AI algorithm is saying, "I don't know".

play03:15

Now, human performance curves are typically a little bit flatter than that.

play03:20

So the human's performance curve might look something like this.

play03:27

Often not quite as accurate as a very confident AI algorithm,

play03:31

but a little better at making the right decision when the AI is unsure.

play03:36

At a 50 percent confidence level, a human is likely to do a better job than an AI.

play03:45

Now why is that?

play03:46

Well, when an AI is certain of itself,

play03:49

it's highly performant and beats out humans who can lose consistency and focus and attention.

play03:54

AIs, they don't get distracted.

play03:57

But on the other hand, when an AI is unsure,

play04:00

often for cases that are complex or statistically rare,

play04:04

humans can outperform an AI prediction by bringing in additional information and context.

play04:10

They can look stuff up or ask a colleague,

play04:12

whereas the AI sticks to its same old decision logic and information.

play04:18

So when a new alert comes in, if the AI assigns a high or low confidence level,

play04:24

then chances are that statistically speaking, it will do a better job of deriving if that alert is real

play04:32

or a false positive, than a given financial analyst.

play04:35

But this is not a zero sum game.

play04:39

It doesn't have to be AI or human.

play04:43

We have one more option.

play04:46

Augmented.

play04:51

Augmented intelligence combines both a human decision, aided by AI,

play04:57

and this performance curve falls somewhere between the two.

play05:04

And for somewhat low and for somewhat high confidence scores,

play05:09

which make up a significant number of predictions,

play05:11

it's augmented intelligence that will have the highest success rate.

play05:19

Except ...

play05:20

... for augmented intelligence to be most effective, we need to account

play05:25

for the messy business of human cognitive bias.

play05:35

We're not always great at doing what we're told.

play05:40

It turns out that how we present information from an AI algorithm to a human decision maker

play05:47

has a significant influence on how effectively that information is used.

play05:52

So, to illustrate that, let's consider forced display vs. optional display.

play06:10

A forced display simultaneously displays an AI recommendation along with a given decision case.

play06:18

So, for every fraud decision alert that I need to make a decision about,

play06:22

I, as the analyst, also see the AI's recommendation.

play06:27

And this can lead to something called automation bias,

play06:30

which is the propensity for humans to favor suggestions from automated decision making systems

play06:36

and to ignore contradictory information.

play06:40

Effectively, the human decision maker is saying the AI knows best

play06:44

and going with the AI prediction at the expense of their own judgment.

play06:49

Optional display means the AI recommendation is only shown to the human decision maker when they request it.

play06:58

So, a person sees a decision case and can then ask the AI to reveal its recommendation.

play07:05

This overcomes automation bias

play07:07

by giving a person time to consider the case for themselves before consulting an AI recommendation.

play07:15

The human is not overwhelmingly influenced by what the AI thinks

play07:19

because they've had a chance to make up their own first impression.

play07:25

And then there's the whole issue of trust, too.

play07:30

When an AI recommendation is accompanied by an accuracy percentage,

play07:37

which indicates how likely this prediction is to be correct,

play07:39

humans are less likely to incorporate the AI recommendation into their decision,

play07:45

regardless of the accuracy percentage being displayed.

play07:49

Basically, we don't like recommendations that openly tell us that they might be wrong.

play07:55

So, we've seen that who should make a decision, a human, an AI,

play08:00

or a human assisted by an AI recommendation, is something that we can derive.

play08:06

We can move from subjective decisions to the quantifiable.

play08:11

That for a given decision who the most effective decision maker is likely to be.

play08:16

And when the most effective decision maker is a combination of AI and human, that's augmented intelligence,

play08:23

we must consider a presentation of that augmentation to minimize human cognitive bias in the decision making process.

play08:31

Brought together, us humans and AI algorithms make a pretty powerful team.

play08:38

We can improve decision making outcomes - if we just know who to ask.

play08:45

If you have any questions, please drop us a line below,

play08:48

and if you want to see more videos like this in the future, please like and subscribe.

play08:53

Thanks for watching.

Rate This

5.0 / 5 (0 votes)

Ähnliche Tags
AI DecisionHuman BiasFraud DetectionPerformance CurveAugmented IntelligenceCognitive BiasAutomation BiasAccuracy PercentageDecision MakingAnalyst Assistance
Benötigen Sie eine Zusammenfassung auf Englisch?