Algorithmic Bias and Fairness: Crash Course AI #18

CrashCourse
13 Dec 201911:20

Summary

TLDRThis CrashCourse AI episode delves into the concept of algorithmic bias, explaining how real-world biases can be mirrored or amplified by AI systems. It outlines five types of biases, including training data issues and feedback loops, and stresses the importance of transparency and critical evaluation of AI recommendations to prevent discrimination and unfair treatment. The episode encourages awareness and advocacy for ethical AI practices.

Takeaways

  • 🧠 Algorithmic bias is the reflection or exaggeration of real-world biases in AI systems, which can be problematic if not acknowledged or addressed.
  • 🧐 Bias itself is not inherently bad; it's a natural human tendency to find patterns, but it becomes an issue when it leads to unfair treatment of certain groups.
  • 📚 Society has laws against discrimination, highlighting the importance of distinguishing between personal bias and systemic discrimination.
  • 🔍 There are five main types of algorithmic bias to be aware of: training data bias, lack of diverse examples, difficulty in quantifying certain features, positive feedback loops, and intentional manipulation of training data.
  • 📈 Training data can contain societal biases, such as gender stereotypes, which can be unintentionally embedded in AI systems, as seen in Google image search results.
  • 🌐 Protected classes like race or gender might not be explicitly present in data, but can emerge as correlated features that influence AI predictions.
  • đŸ‘„ Insufficient examples of certain classes in training data can lead to inaccurate AI predictions, as seen in facial recognition systems struggling with non-white faces.
  • 📊 Quantifying complex human experiences and qualities can be challenging for AI, leading to reliance on easily measurable but less meaningful metrics.
  • 🔁 Positive feedback loops in AI can perpetuate and amplify existing biases, as seen with PredPol's crime prediction algorithm and its impact on policing.
  • đŸ‘Ÿ AI systems can be manipulated, as demonstrated by Microsoft's chatbot Tay, which quickly adopted inappropriate behaviors due to user input.
  • đŸ€– Human oversight is crucial in AI systems to ensure fairness and adjust algorithms when necessary, emphasizing the role of transparency and critical evaluation.

Q & A

  • What is algorithmic bias?

    -Algorithmic bias refers to the phenomenon where AI systems mimic or even exaggerate the biases that exist in the real world due to the data they are trained on or the way they are designed.

  • Why is it important to differentiate between bias and discrimination?

    -It's important because bias is a natural human tendency to find patterns, but discrimination is an unfair treatment of certain groups which is illegal and can be prevented. Understanding this helps in addressing algorithmic bias ethically.

  • Can you give an example of how biases in training data can affect AI systems?

    -Yes, if an AI is trained on recent news articles or books, it might associate the word 'nurse' with 'woman' and 'programmer' with 'man', reflecting societal stereotypes.

  • How can protected classes emerge as correlated features in AI algorithms?

    -Protected classes like race or gender might not be explicitly in the data but can emerge as correlated features due to societal factors. For example, zip code can be correlated to race due to residential segregation.

  • What is the issue with training data not having enough examples of each class?

    -If the training data lacks sufficient representation of each class, it can affect the accuracy of predictions. For instance, facial recognition AI trained mostly on white faces may have trouble recognizing people of other races.

  • Why is it challenging to quantify certain features in training data?

    -Some features, like the quality of writing or the complexity of relationships, are difficult to quantify because they involve subjective and nuanced qualities that cannot be easily measured with numbers.

  • How can an algorithm create a positive feedback loop?

    -A positive feedback loop occurs when the algorithm's predictions influence the data it receives, amplifying past trends. For example, PredPol's crime prediction algorithm could lead to increased police presence in certain neighborhoods based on biased historical data.

  • What was the issue with Microsoft's chatbot Tay?

    -Tay was manipulated by a subset of people to post violent, sexist, anti-Semitic, and racist Tweets within 12 hours of its release, showing how AI can be influenced by biased input.

  • Why is transparency in algorithms important?

    -Transparency is crucial for understanding why an algorithm makes certain recommendations, allowing us to critically assess AI outputs and ensure fairness and accuracy.

  • What can be done to monitor AI for bias and discrimination?

    -We can start by acknowledging that algorithms will have biases, being critical of AI recommendations, and advocating for transparency and careful interpretation of algorithmic outputs to protect human rights.

  • What is the role of humans in ensuring fairness in AI systems?

    -Humans play a vital role in monitoring, interpreting, and adjusting AI systems to ensure that recommendations are fair and not influenced by harmful biases or discrimination.

Outlines

00:00

🧠 Understanding Algorithmic Bias

The first paragraph introduces the concept of algorithmic bias, explaining how biases in society can be mirrored or amplified by AI systems. Jabril discusses the importance of recognizing the difference between natural human bias and harmful discrimination. He emphasizes the need for awareness to prevent AI from perpetuating societal inequalities. The paragraph also outlines five types of algorithmic bias, starting with the reflection of societal biases in training data, as evidenced by gender stereotypes in Google image search results. The summary also touches on the potential for protected classes to emerge through correlated features, despite not being explicitly included in the data.

05:04

🔍 Exploring the Consequences of Algorithmic Bias

The second paragraph delves into the various consequences of algorithmic bias, such as insufficient representation in training data leading to inaccuracies, the challenge of quantifying certain features, and the creation of positive feedback loops that reinforce past biases. Jabril uses examples like facial recognition systems and AI grading of essays to illustrate these points. He also discusses the potential for manipulation of AI systems, as seen with Microsoft's chatbot Tay, which was quickly corrupted by biased user input. The paragraph concludes with a call for skepticism towards AI predictions and the importance of human oversight in AI systems.

10:07

🛡 Combating Algorithmic Bias Through Awareness and Action

The third paragraph focuses on the importance of understanding and combating algorithmic bias. Jabril suggests that acknowledging the inherent bias in algorithms is the first step, followed by critical examination of AI recommendations. He highlights the need for transparency in algorithms to understand their decision-making processes. The paragraph also discusses the potential benefits and pitfalls of increasing training data for protected classes and the ethical considerations involved. Jabril ends with a call to action for everyone to stay informed about AI and advocate for fair and unbiased algorithmic outputs.

Mindmap

Keywords

💡Algorithmic Bias

Algorithmic bias refers to the systematic errors that arise when an AI system makes predictions based on training data that reflects societal prejudices. In the video, it is explained as a mimic or exaggeration of real-world biases, which can be unintentional but has significant implications for fairness. Examples include AI systems associating the term 'nurse' with women and 'programmer' with men, based on biased training data.

💡Protected Classes

Protected classes are categories recognized by law that warrant protection against discrimination. The script mentions these classes in the context of societal laws that prevent unfair treatment based on attributes like gender, race, or age. The video emphasizes the importance of distinguishing between inherent biases and the legal prevention of discrimination against these protected groups.

💡Correlated Features

Correlated features are indirect attributes in data that may not be explicitly intended to be used for predictions but become influential due to their statistical relationship with the outcome. The video gives the example of zip code being correlated with race due to residential segregation, which can unintentionally introduce bias into AI systems.

💡Training Data

Training data is the set of information used to teach AI systems how to make predictions or decisions. The script discusses how biases in training data can lead to algorithmic bias, such as facial recognition systems that perform poorly on non-white faces due to an overrepresentation of white faces in the training data.

💡Quantification

Quantification is the process of measuring or assigning numerical values to qualities or characteristics. The video explains the difficulty in quantifying complex human experiences or relationships, such as grading writing, where AI may rely on easily measurable but less meaningful factors like sentence length instead of more nuanced qualities like creativity.

💡Positive Feedback Loop

A positive feedback loop is a process that amplifies an initial effect, often leading to a self-reinforcing cycle. In the context of the video, it is used to describe how AI systems like PredPol can perpetuate and exacerbate existing biases by continuously feeding data back into the system that reinforces past patterns, such as directing police to neighborhoods with higher arrest rates.

💡Cultural Bias

Cultural bias refers to the prejudiced views or stereotypes that are embedded in a society's culture and can be unintentionally incorporated into AI systems. The script points out that AI algorithms may not recognize changing cultural biases over time, potentially spreading outdated or harmful stereotypes.

💡Data Manipulation

Data manipulation occurs when individuals intentionally alter the data used to train AI systems to produce biased outcomes. The video cites the example of Microsoft's chatbot Tay, which was manipulated to generate offensive content within hours of its release, illustrating the vulnerability of AI to such tactics.

💡Transparency

Transparency in algorithms refers to the ability to understand and examine the inputs and outputs of an AI system to determine why certain decisions are made. The script emphasizes the importance of transparency to identify and address biases in AI, although it acknowledges the challenges, especially with complex deep learning algorithms.

💡Discrimination

Discrimination is the unfair treatment of individuals or groups based on their membership in a certain category. The video distinguishes between personal bias, which is common, and discrimination, which is actionable and can be prevented. It also discusses how AI systems can inadvertently perpetuate or amplify discriminatory practices.

💡Thought Bubble

The 'Thought Bubble' is a narrative device used in the video to illustrate a hypothetical scenario involving an AI hiring system called HireMe!. It serves as an example to explain the concept of a positive feedback loop and its potential negative consequences, such as biased hiring practices against individuals with certain names.

Highlights

Algorithmic bias is the reflection or exaggeration of real-world biases in AI systems.

Bias is not inherently negative, but becomes problematic when it leads to unfair treatment of certain groups.

Society has laws to prevent discrimination based on protected classes like gender, race, or age.

There are at least 5 types of algorithmic bias, including biases in training data, lack of diverse examples, and difficulty in quantifying certain features.

Training data can embed societal biases, as seen with gender stereotypes in job roles.

AI algorithms may not recognize cultural biases that change over time, potentially spreading hidden biases further.

Protected classes can emerge as correlated features in data, unintentionally leading to discrimination.

Insufficient examples of each class in training data can affect the accuracy of AI predictions.

Facial recognition AI algorithms have been biased due to a lack of diverse racial representation in training data.

Quantifying complex human experiences for AI training is challenging and can lead to oversimplification.

AI grading systems may focus on easily measurable elements rather than the full complexity of qualities like good writing.

Algorithms can create positive feedback loops that amplify past biases, as seen with PredPol's crime prediction algorithm.

People may manipulate AI training data intentionally, as demonstrated by Microsoft's chatbot Tay.

AI systems make predictions but can make mistakes with significant consequences.

Understanding the limitations of AI is crucial for addressing algorithmic bias.

Transparency in algorithms is important for examining inputs and outputs to understand recommendations.

More training data on protected classes may be needed to reduce bias in AI algorithms.

Advocating for careful interpretation of algorithmic outputs can help protect human rights.

Some advocate for algorithms to be tested like medicines to understand potential 'side effects' before integration into daily life.

Transcripts

play00:00

Hi, I’m Jabril and welcome back to CrashCourse AI.

play00:03

Algorithms are just math and code, but algorithms are created by people and use our data, so

play00:08

biases that exist in the real world are mimicked or even exaggerated by AI systems.

play00:14

This idea is called algorithmic bias.

play00:16

Bias isn’t inherently a terrible thing.

play00:19

Our brains try to take shortcuts by finding patterns in data.

play00:23

So if you’ve only seen small, tiny dogs, you might see a Great Dane and be like “Whoa

play00:28

that dog is unnatural”

play00:30

This doesn’t become a problem unless we don’t acknowledge exceptions to patterns

play00:33

or unless we start treating certain groups of people unfairly.

play00:37

As a society, we have laws to prevent discrimination based on certain “protected classes” (like

play00:42

gender, race, or age) for things like employment or housing.

play00:46

So it’s important to be aware of the difference between bias, which we all have, and discrimination,

play00:51

which we can prevent.

play00:52

And knowing about algorithmic bias can help us steer clear of a future where AI are used

play00:57

in harmful, discriminatory ways.

play00:59

INTRO

play01:08

There are at least 5 types of algorithmic bias we should pay attention to.

play01:12

First, training data can reflect hidden biases in society.

play01:15

For example, if an AI was trained on recent news articles or books, the word “nurse”

play01:19

is more likely to refer to a “woman,” while the word “programmer” is more likely

play01:23

to refer to a “man.”

play01:25

And you can see this happening with a Google image search: “nurse” shows mostly women,

play01:29

while “programmer” mostly shows mostly men.

play01:32

We can see how hidden biases in the data gets embedded in search engine AI.

play01:36

Of course, we know there are male nurses and female programmers and non-binary people doing

play01:40

both of these jobs!

play01:42

For example, an image search for “programmer 1960” shows a LOT more women.

play01:47

But AI algorithms aren’t very good at recognizing cultural biases that might change over time,

play01:52

and they could even be spreading hidden biases to more human brains.

play01:55

t’s also tempting to think that if we just don’t collect or use training data that

play01:59

categorizes protected classes like race or gender, then our algorithms can’t possibly

play02:04

discriminate.

play02:04

But, protected classes may emerge as correlated features, which are features that aren’t

play02:09

explicitly in data but may be unintentionally correlated to a specific prediction.

play02:14

For example, because many places in the US are still extremely segregated, zip code can

play02:18

be strongly correlated to race.

play02:20

A record of purchases can be strongly correlated to gender.

play02:24

And a controversial 2017 paper showed that sexual orientation is strongly correlated

play02:29

with characteristics of a social media profile photo.

play02:32

Second, the training data may not have enough examples of each class, which can affect the

play02:36

accuracy of predictions.

play02:38

For example, many facial recognition AI algorithms are trained on data that includes way more

play02:43

examples of white peoples’ faces than other races.

play02:46

One story that made the news a few years ago is a passport photo checker with an AI system

play02:50

to warn if the person in the photo had blinked.

play02:54

But the system had a lot of trouble with photos of people of Asian descent.

play02:58

Being asked to take a photo again and again would be really frustrating if you’re just

play03:02

trying to renew your passport, which is already sort of a pain!

play03:05

Or, let’s say, you got a cool gig programming a drone for IBM
 but it has trouble recognizing

play03:11

your face because your skin’s too dark
 for example.

play03:15

Third, it’s hard to quantify certain features in training data.

play03:19

There are lots of things that are tough to describe with numbers.

play03:22

Like can you really rate a sibling relationship with a number?

play03:25

It’s complicated!

play03:26

You love them, but you hate how messy they are, but you like cooking together, but you

play03:31

hate how your parents compare you...

play03:33

It’s so hard to quantify all that!

play03:35

In many cases, we try to build AI to evaluate complicated qualities of data, but sometimes

play03:40

we have to settle for easily measurable shortcuts.

play03:43

One recent example is trying to use AI to grade writing on standardized tests like SATs

play03:48

and GREs with the goal to save human graders time.

play03:51

Good writing involves complex elements like clarity, structure, and creativity, but most

play03:56

of these qualities are hard to measure.

play03:58

So, instead, these AI focused on easier-to-measure elements like sentence length, vocabulary,

play04:03

and grammar, which don’t fully represent good writing
 and made these AIs easier

play04:08

to fool.

play04:09

Some students from MIT built a natural language program to create essays that made NO sense,

play04:13

but were rated highly by these grading algorithms.

play04:16

These AIs could also potentially be fooled by memorizing portions of “template” essays

play04:20

to influence the score, rather than actually writing a response to the prompt, all because

play04:25

of the training data that was used for these scoring AI.

play04:28

Fourth, the algorithm could influence the data that it gets, creating a positive feedback

play04:32

loop.

play04:33

A positive feedback loop basically means “amplifying what happened in the past”
 whether or

play04:37

not this amplification is good.

play04:39

An example is PredPol’s drug crime prediction algorithm, which has been in use since 2012

play04:44

in many large cities including LA and Chicago.

play04:47

PredPol was trained on data that was heavily biased by past housing segregation and past

play04:52

cases of police bias.

play04:53

So, it would more frequently send police to certain neighborhoods where a lot of racial

play04:57

minority folks lived.

play04:59

Arrests in those neighborhoods increased, that arrest data was fed back into the algorithm,

play05:03

and the AI would predict more future drug arrests in those neighborhoods and send the

play05:07

police there again.

play05:08

Even though there might be crime in neighborhoods where police weren’t being sent by this

play05:12

AI, because there weren't any arrests in those neighborhoods, data about them wasn’t fed

play05:17

back into the algorithm.

play05:18

While algorithms like PredPol are still in use, to try and manage these feedback effects,

play05:23

there is currently more effort to monitor and adjust how they process data.

play05:27

So basically, this would be like a new principal who was hired to improve the average grades

play05:31

of a school, but he doesn’t really care about the students who already have good grades.

play05:36

He creates a watchlist of students who have really bad grades and checks up on them every

play05:41

week, and he ignores the students who keep up with good grades.

play05:44

If any of the students on his watchlist don’t do their homework that week, they get punished.

play05:50

But all of the students NOT on his watchlist can slack on their homework, and get away

play05:54

with it based on “what happened in the past.”

play05:57

This is essentially what’s happening with PredPol, and you can be the judge if you believe

play06:01

it’s fair or not.

play06:02

Finally, a group of people may mess with training data on purpose.

play06:06

For example, in 2014, Microsoft released a chatbot named Xiaoice in China.

play06:11

People could chat with Xiaoice so it would learn how to speak naturally on a variety

play06:14

of topics from these conversations.

play06:17

It worked great, and Xiaoice had over 40 million conversations with no incidents.

play06:21

In 2016, Microsoft tried the same thing in the U.S. by releasing the Twitterbot Tay.

play06:26

Tay trained on direct conversation threads on Twitter, and by playing games with users

play06:31

where they could get it to repeat what they were saying.

play06:33

In 12 hours after its release, after a “coordinated attack by a subset of people” who biased

play06:38

its data set, Tay started posting violent, sexist, anti-semitic, and racist Tweets.

play06:43

This kind of manipulation is usually framed as “joking” or “trolling,” but the

play06:47

fact that AI can be manipulated means we should take algorithmic predictions with a grain

play06:51

of salt.

play06:52

This is why I don’t leave John-Green-Bot alone online


play06:55

The common theme of algorithmic bias is that AI systems are trying to make good predictions,

play07:00

but they make mistakes.

play07:02

Some of these mistakes may be harmless or mildly inconvenient, but others may have significant

play07:07

consequences.

play07:08

To understand the key limitations of AI in our current society, let’s go to the Thought

play07:11

Bubble.

play07:13

Let’s say there’s an AI system called HireMe! that gives hiring recommendations

play07:16

to companies.

play07:18

HireMe is being used by Robots Weekly, a magazine where John-Green-bot applied for an editorial

play07:23

job.

play07:24

Just by chance, the last two people named “John” got fired from Robots Weekly and

play07:28

another three “Johns” didn’t make it through the hiring process.

play07:31

So, when John-Green-Bot applies for the job, HireMe! predicts that he’s only 24% likely

play07:36

to be employed by the company in 3 years.

play07:39

Seeing this prediction, the hiring manager at Robots Weekly rejects John-Green-bot, and

play07:44

this data gets added to the HireMe!

play07:46

AI system.

play07:47

John-Green-Bot is just another “John” that got rejected, even though he may have

play07:50

been the perfect robot for the job!

play07:52

Now, future “Johns” have an even lower chance to be hired.

play07:56

It’s a positive feedback loop, with some pretty negative consequences for John-Green-Bot.

play08:01

Of course, being named “John” isn’t a protected class, but this could apply to

play08:06

other groups of people.

play08:07

Plus, even though algorithms like HireMe!

play08:09

Are great at establishing a link between two kinds of data, they can’t always clarify

play08:14

why they’re making predictions.

play08:16

For example, HireMe! may find that higher age is associated with lower knowledge of

play08:20

digital technologies, so the AI suggests hiring younger applicants.

play08:25

Not only is this illegally discriminating against the protected class of “age,”

play08:29

but the implied link also might not be true.

play08:32

John-Green-bot may be almost 40, but he runs a robot blog and is active in online communities

play08:38

like Nerdfighteria!

play08:39

So it’s up to humans interacting with AI systems like HireMe! to pay attention to recommendations

play08:44

and make sure they’re fair, or adjust the algorithms if not.

play08:48

Thanks, Thought Bubble!

play08:50

Monitoring AI for bias and discrimination sounds like a huge responsibility, so how

play08:55

can we do it?

play08:56

The first step is just understanding that algorithms will be biased.

play09:00

It’s important to be critical about AI recommendations, instead of just accepting that “the computer

play09:05

said so.”

play09:06

This is why transparency in algorithms is so important, which is the ability to examine

play09:11

inputs and outputs to understand why an algorithm is giving certain recommendations.

play09:15

But that's easier said than done when it comes to certain algorithms, like

play09:19

deep learning methods.

play09:20

Hidden layers can be tricky to interpret.

play09:23

Second, if we want to have less biased algorithms, we may need more training data on protected

play09:28

classes like race, gender, or age.

play09:30

Looking at an algorithm’s recommendations for protected classes may be a good way to

play09:34

check it for discrimination.

play09:36

This is kind of a double-edged sword, though.

play09:39

People who are part of protected classes may (understandably) be worried about handing

play09:42

over personal information.

play09:44

It may feel like a violation of privacy, or they might worry that algorithms will be misused

play09:50

to target rather than protect them.

play09:52

Even if you aren’t actively working on AI systems, knowing about these algorithms and

play09:56

staying informed about artificial intelligence are really important as we shape the future

play10:01

of this field.

play10:02

Anyone, including you, can advocate for more careful, critical interpretation of algorithmic

play10:06

outputs to help protect human rights.

play10:09

Some people are even advocating that algorithms should be clinically tested and scrutinized

play10:13

in the same way that medicines are.

play10:15

According to these opinions, we should know if there are “side effects” before integrating

play10:20

AI in our daily lives.

play10:21

There’s nothing like that in the works yet.

play10:24

But it took over 2400 years for the Hippocratic Oath to transform into current medical ethics

play10:28

guidelines.

play10:30

So it may take some time for us to come up with the right set of practices.

play10:34

Next time, we have a lab and I’ll demonstrate how there are biases in even simple things

play10:38

like trying to adopt a cat or a dog.

play10:41

I’ll see ya then.

play10:42

Speaking of understanding how bias and misinformation spread, you should check out this video on Deep Fakes

play10:47

I did with Above the Noise -- another PBSDS channel that gets into the research behind controversial issues.

play10:53

Head over to the video in the description to find out how detect deep fakes.

play10:57

Tell them Jabril sent you!

play10:59

Crash Course AI is produced in association with PBS Digital Studios!

play11:03

If you want to help keep all Crash Course free for everybody, forever, you can join

play11:06

our community on Patreon.

play11:08

And if you want to learn more about prejudice and discrimination in humans, you can check

play11:11

out this episode of Crash Course Sociology.

Rate This
★
★
★
★
★

5.0 / 5 (0 votes)

Étiquettes Connexes
Algorithmic BiasArtificial IntelligenceData PatternsDiscrimination AwarenessProtected ClassesCultural ShiftGender StereotypesRacial CorrelationAI TrainingBias MitigationEthical AI
Besoin d'un résumé en anglais ?