When AI Can Fake Reality, Who Can You Trust? | Sam Gregory | TED

TED
26 Dec 202312:05

Summary

TLDRThe speaker from WITNESS, a human rights group, discusses the escalating challenge of distinguishing real from AI-generated content, highlighting the societal impacts of deepfakes. They share insights from a deepfakes rapid-response task force and emphasize the need for detection tools, content provenance, and a pipeline of responsibility to fortify truth and combat misinformation in an increasingly AI-infused media landscape.

Takeaways

  • 🧩 The distinction between real and fake is becoming increasingly blurred with advances in generative AI and deepfakes.
  • 🚀 The speaker began working on deepfakes in 2017, highlighting the evolution from a hype to a serious threat, particularly in the creation of falsified sexual images.
  • 🌐 The impact of deepfakes is global, affecting women and girls and now expanding to include the potential to dismiss real events as faked.
  • 🛡 WITNESS, the human-rights group led by the speaker, aids people in using technology to defend their rights and has coordinated a global effort to combat deepfakes.
  • 🔍 A deepfakes rapid-response task force has been established, consisting of media-forensics experts and companies that debunk deepfakes and claims of deepfakes.
  • 🗣️ The task force has dealt with cases from Sudan, West Africa, and India, demonstrating the complexity and challenges in verifying the authenticity of audio clips.
  • 🕵️‍♂️ Even experts struggle to rapidly and conclusively determine the authenticity of deepfakes, and the ease of falsely accusing real content as fake is increasing.
  • 🌟 The future presents profound challenges in protecting real content and detecting fakes, with deepfakes targeting politicians and influencing political ads and crisis reporting.
  • 🔑 The need for detection skills and tools to be accessible to those who need them most, such as journalists and human-rights defenders, is emphasized.
  • 💡 There is a call for better understanding of content provenance and disclosure through technologies like invisible watermarking and cryptographically signed metadata.
  • 🌳 A responsible pipeline from AI foundations to deployment in systems and platforms is necessary to ensure transparency, accountability, and liability in AI usage.

Q & A

  • What is the main challenge discussed in the script regarding the advancement of generative AI?

    -The main challenge is the increasing difficulty in distinguishing between real and fake content, as well as the potential for AI to both create convincing fakes and to be used as an excuse to dismiss genuine reality.

  • When did the speaker start working on deepfakes, and what was the initial concern?

    -The speaker started working on deepfakes in 2017, with the initial concern being the overhyped threat to trust in information and the harm caused by falsified sexual images.

  • What is the role of WITNESS as described in the script?

    -WITNESS is a human-rights group that helps people use video and technology to protect and defend their rights, and has coordinated a global effort called 'Prepare, Don't Panic' to address the manipulation and synthesis of reality.

  • What is the purpose of the deepfakes rapid-response task force mentioned in the script?

    -The task force, composed of media-forensics experts and companies, aims to debunk deepfakes and claims of deepfakes, providing a rapid response to such incidents.

  • How did the task force handle the audio clip from Sudan?

    -Experts used a machine-learning algorithm trained on over a million examples of synthetic speech to prove with high certainty that the Sudan audio clip was authentic.

  • What challenges did the task force face with the West Africa audio clip?

    -The task force couldn't reach a definitive conclusion due to challenges in analyzing audio from Twitter with background noise, which affected the analysis.

  • What was the outcome of the analysis of the Indian politician's leaked audio clip?

    -Despite the politician's claims that the audio was AI-falsified, experts concluded that it was at least partially real, not AI-generated.

  • What are the three steps proposed to address the challenges posed by deepfakes and AI in communication?

    -The three steps are: 1) Ensuring detection skills and tools are available to those who need them, 2) Understanding the content provenance and disclosure through metadata and watermarking, and 3) Establishing a pipeline of responsibility from AI models to the platforms where media is consumed.

  • Why is it important to have robust detection tools for deepfakes?

    -Robust detection tools are important to help journalists, community leaders, and human-rights defenders discern authenticity from simulation and to fortify the credibility of critical voices and images.

  • What is the significance of content provenance and disclosure in the context of AI-generated media?

    -Content provenance and disclosure, through metadata and watermarking, provide a 'recipe' of how AI and human input were mixed in the creation or editing of media, which is essential for building trust and literacy in AI-infused media.

  • How can we ensure that the infrastructure for authenticity does not compromise privacy or backfire globally?

    -By focusing on the 'how' of AI-human media making rather than the 'who', ensuring that the infrastructure respects rights and allows for anonymity where necessary, without obliging disclosure of personal information.

  • What is the role of governments in ensuring transparency, accountability, and liability in the pipeline of AI responsibility?

    -Governments need to ensure that there is a clear pipeline of responsibility for AI, including transparency in how AI is used, accountability for its effects, and liability for misuse, to prevent the repetition of social media failures in the next generation of technology.

Outlines

00:00

🤖 The Challenge of Deepfakes and Trust in Information

This paragraph discusses the increasing difficulty in distinguishing between real and AI-generated content, particularly with the advancement of deepfake technology. The speaker, from the human rights group WITNESS, outlines the evolution of deepfakes from a hyped threat in 2017 to a growing issue affecting women and girls globally. They highlight the broader implications of generative AI in creating and dismissing 'fake reality,' and the role of their organization in combating deepfakes through a rapid-response task force. The paragraph also details the task force's efforts in analyzing audio clips from Sudan, West Africa, and India, emphasizing the complexity and evolving nature of deepfake detection.

05:03

🔍 Addressing the Deepfake Phenomenon: Detection and Provenance

The speaker emphasizes the need for structural solutions to address the deepfake issue, focusing on three main steps. First, ensuring that detection tools and skills are accessible to those who need them most, such as journalists and human rights defenders. They discuss the limitations and challenges of current detection tools, which may not be reliable or universally effective. Second, the paragraph introduces the concept of content provenance and disclosure, including the use of invisible watermarking and cryptographically signed metadata to provide a 'recipe' of AI and human involvement in media creation. This approach is presented as a critical component of new media literacy in an AI-infused world. The speaker also touches on the importance of maintaining privacy and avoiding the disclosure of personal information in the process.

10:04

🌐 Creating a Pipeline of Responsibility for AI in Media

In the final paragraph, the speaker calls for a pipeline of responsibility that encompasses the entire lifecycle of AI in media, from foundational models to deployment in systems and platforms. They argue that governments must ensure transparency, accountability, and liability within this pipeline to prevent the misuse of AI technologies. The speaker warns of the dangers of a world where reality can be easily faked and genuine reality dismissed as potentially fake, referencing the political philosopher Hannah Arendt's views on the consequences of a society unable to discern truth from falsehood. The paragraph concludes with a call to action to prevent such a future by taking the necessary steps now.

Mindmap

Keywords

💡Generative AI

Generative AI refers to artificial intelligence systems that can create new content, such as text, images, or audio, that is similar to the data they were trained on. In the video's context, generative AI is central to the discussion of deepfakes and the challenges they pose to discerning real from fake content. The script mentions how advances in generative AI have made it easier to create convincingly fake realities.

💡Deepfakes

Deepfakes are a type of media forgery that use AI to manipulate or generate images or videos that appear real but are not. They are a significant concern in the video, as they illustrate the growing problem of misinformation and the potential for malicious use, such as falsified sexual images or impersonating public figures.

💡Media-forensics

Media-forensics is the field of analyzing digital media to determine its authenticity. In the video, a deepfakes rapid-response task force composed of media-forensics experts is mentioned, emphasizing the importance of this field in debunking deepfakes and verifying the authenticity of media content.

💡Synthetic speech

Synthetic speech is a technology that generates human-like speech using AI. The script refers to a machine-learning algorithm trained on over a million examples of synthetic speech, which was used to authenticate an audio clip, demonstrating the role of this technology in both creating and detecting deepfakes.

💡Authenticity

Authenticity in the context of the video refers to the genuineness and trustworthiness of information or media content. The speaker discusses the challenges of proving authenticity in the face of deepfakes and the importance of tools and techniques to verify and maintain the authenticity of media.

💡Human-rights defenders

Human-rights defenders are individuals or groups who promote and protect human rights. The video emphasizes the role of such defenders in using technology, like video and AI, to document and expose human rights abuses, and the challenges they face with the rise of deepfakes and misinformation.

💡AI-generated imagery

AI-generated imagery refers to visual content created by AI, which can be used to depict events or situations that did not actually occur. The script warns of the dangers of AI-generated imagery being shared as real, contributing to the spread of disinformation and confusion.

💡Content provenance

Content provenance is the origin and history of digital content, including how it was created and by whom. The video discusses the need for content provenance to establish trust in media, especially with the increasing use of AI in content creation and editing.

💡Cryptographically signed metadata

Cryptographically signed metadata refers to data about a digital file that is securely signed with a cryptographic key, ensuring its authenticity and integrity. In the video, this concept is presented as a solution for adding trust to AI-generated media by providing a record of how the content was created and distributed.

💡AI-infused media literacy

AI-infused media literacy is the ability to critically evaluate and understand media content that has been created or altered by AI. The video suggests that as AI becomes more prevalent in communication, media literacy must evolve to include understanding the role of AI in content creation and the implications for trust and authenticity.

💡Pipeline of responsibility

A pipeline of responsibility refers to a series of accountability measures that should be in place throughout the development, deployment, and use of AI technologies. The video argues for the necessity of such a pipeline to ensure transparency, accountability, and liability in AI applications, particularly in the context of media and communication.

Highlights

The increasing difficulty in distinguishing between real and fake content due to advances in generative AI and deepfakes.

The initial focus on deepfakes in 2017 was on falsified sexual images, which have since grown to affect women and girls worldwide.

The evolving threat of generative AI in creating fake realities and dismissing real ones as potentially faked.

The role of audiovisual AI in exacerbating societal problems, including electoral manipulation and human rights abuses.

Introduction of WITNESS, a human rights group that aids in using technology to defend rights against new forms of reality manipulation.

The 'Prepare, Don't Panic' initiative, a global effort to counteract the manipulation and synthesis of reality.

The deepfakes rapid-response task force, composed of media-forensics experts and companies, debunking deepfakes and claims of deepfakes.

Case studies from Sudan, West Africa, and India, illustrating the challenges and successes in deepfake detection.

The limitations of current deepfake detection methods, including the inability to conclusively separate true from false.

The future challenges in protecting real content and detecting fakes, with a focus on political leaders and crisis zones.

The importance of not relying solely on individuals to discern deepfakes, but rather implementing structural solutions.

The need for robust foundations to discern authenticity, fortify credibility, and develop powerful detection technology.

Three proposed steps to address the challenges of AI in communication: detection tools, content provenance, and a pipeline of responsibility.

Ensuring detection skills and tools are accessible to those who need them most, such as journalists and human rights defenders.

The concept of content provenance and disclosure, including invisible watermarking and cryptographically signed metadata.

The challenges in creating an infrastructure for authenticity that respects privacy and does not compromise anonymity.

The necessity for a pipeline of responsibility from AI foundations to platforms, ensuring transparency, accountability, and liability.

The risk of a world where reality is easily faked and genuine reality is dismissed, leading to a loss of capacity to act, think, and judge.

Transcripts

play00:03

It's getting harder, isn't it, to spot real from fake,

play00:07

AI-generated from human-generated.

play00:10

With generative AI,

play00:11

along with other advances in deep fakery,

play00:13

it doesn't take many seconds of your voice,

play00:16

many images of your face,

play00:17

to fake you,

play00:19

and the realism keeps increasing.

play00:21

I first started working on deepfakes in 2017,

play00:24

when the threat to our trust in information was overhyped,

play00:28

and the big harm, in reality, was falsified sexual images.

play00:32

Now that problem keeps growing, harming women and girls worldwide.

play00:38

But also, with advances in generative AI, we're now also approaching a world

play00:42

where it's broadly easier to make fake reality,

play00:46

but also to dismiss reality as possibly faked.

play00:50

Now, deceptive and malicious audiovisual AI

play00:54

is not the root of our societal problems,

play00:56

but it's likely to contribute to them.

play00:59

Audio clones are proliferating in a range of electoral contexts.

play01:03

"Is it, isn't it" claims cloud human-rights evidence from war zones,

play01:08

sexual deepfakes target women in public and in private,

play01:12

and synthetic avatars impersonate news anchors.

play01:16

I lead WITNESS.

play01:18

We're a human-rights group

play01:19

that helps people use video and technology to protect and defend their rights.

play01:23

And for the last five years, we've coordinated a global effort,

play01:26

"Prepare, Don't Panic,"

play01:27

around these new ways to manipulate and synthesize reality,

play01:30

and on how to fortify the truth

play01:32

of critical frontline journalists and human-rights defenders.

play01:37

Now, one element in that is a deepfakes rapid-response task force,

play01:42

made up of media-forensics experts

play01:44

and companies who donate their time and skills

play01:46

to debunk deepfakes and claims of deepfakes.

play01:50

The task force recently received three audio clips,

play01:54

from Sudan, West Africa and India.

play01:57

People were claiming that the clips were deepfaked, not real.

play02:01

In the Sudan case,

play02:02

experts used a machine-learning algorithm

play02:04

trained on over a million examples of synthetic speech

play02:07

to prove, almost without a shadow of a doubt,

play02:09

that it was authentic.

play02:11

In the West Africa case,

play02:13

they couldn't reach a definitive conclusion

play02:15

because of the challenges of analyzing audio from Twitter,

play02:18

and with background noise.

play02:20

The third clip was leaked audio of a politician from India.

play02:23

Nilesh Christopher of “Rest of World” brought the case to the task force.

play02:27

The experts used almost an hour of samples

play02:30

to develop a personalized model of the politician's authentic voice.

play02:35

Despite his loud and fast claims that it was all falsified with AI,

play02:39

experts concluded that it at least was partially real, not AI.

play02:44

As you can see,

play02:45

even experts cannot rapidly and conclusively separate true from false,

play02:50

and the ease of calling "that's deepfaked" on something real

play02:55

is increasing.

play02:57

The future is full of profound challenges,

play02:59

both in protecting the real and detecting the fake.

play03:03

We're already seeing the warning signs

play03:05

of this challenge of discerning fact from fiction.

play03:08

Audio and video deepfakes have targeted politicians,

play03:11

major political leaders in the EU, Turkey and Mexico,

play03:15

and US mayoral candidates.

play03:17

Political ads are incorporating footage of events that never happened,

play03:20

and people are sharing AI-generated imagery from crisis zones,

play03:25

claiming it to be real.

play03:27

Now, again, this problem is not entirely new.

play03:31

The human-rights defenders and journalists I work with

play03:33

are used to having their stories dismissed,

play03:36

and they're used to widespread, deceptive, shallow fakes,

play03:40

videos and images taken from one context or time or place

play03:43

and claimed as if they're in another,

play03:46

used to share confusion and spread disinformation.

play03:49

And of course, we live in a world that is full of partisanship

play03:53

and plentiful confirmation bias.

play03:57

Given all that,

play03:58

the last thing we need is a diminishing baseline

play04:01

of the shared, trustworthy information upon which democracies thrive,

play04:05

where the specter of AI

play04:07

is used to plausibly believe things you want to believe,

play04:10

and plausibly deny things you want to ignore.

play04:15

But I think there's a way we can prevent that future,

play04:17

if we act now;

play04:19

that if we "Prepare, Don't Panic,"

play04:21

we'll kind of make our way through this somehow.

play04:25

Panic won't serve us well.

play04:28

[It] plays into the hands of governments and corporations

play04:31

who will abuse our fears,

play04:33

and into the hands of people who want a fog of confusion

play04:36

and will use AI as an excuse.

play04:40

How many people were taken in, just for a minute,

play04:43

by the Pope in his dripped-out puffer jacket?

play04:45

You can admit it.

play04:46

(Laughter)

play04:47

More seriously,

play04:49

how many of you know someone who's been scammed

play04:51

by an audio that sounds like their kid?

play04:54

And for those of you who are thinking "I wasn't taken in,

play04:57

I know how to spot a deepfake,"

play04:59

any tip you know now is already outdated.

play05:02

Deepfakes didn't blink, they do now.

play05:06

Six-fingered hands were more common in deepfake land than real life --

play05:09

not so much.

play05:11

Technical advances erase those visible and audible clues

play05:15

that we so desperately want to hang on to

play05:17

as proof we can discern real from fake.

play05:20

But it also really shouldn’t be on us to make that guess without any help.

play05:24

Between real deepfakes and claimed deepfakes,

play05:27

we need big-picture, structural solutions.

play05:30

We need robust foundations

play05:31

that enable us to discern authentic from simulated,

play05:34

tools to fortify the credibility of critical voices and images,

play05:38

and powerful detection technology

play05:41

that doesn't raise more doubts than it fixes.

play05:45

There are three steps we need to take to get to that future.

play05:48

Step one is to ensure that the detection skills and tools

play05:52

are in the hands of the people who need them.

play05:54

I've talked to hundreds of journalists,

play05:57

community leaders and human-rights defenders,

play05:59

and they're in the same boat as you and me and us.

play06:02

They're listening to the audio, trying to think, "Can I spot a glitch?"

play06:05

Looking at the image, saying, "Oh, does that look right or not?"

play06:08

Or maybe they're going online to find a detector.

play06:12

And the detector they find,

play06:13

they don't know whether they're getting a false positive, a false negative,

play06:17

or a reliable result.

play06:18

Here's an example.

play06:19

I used a detector, which got the Pope in the puffer jacket right.

play06:23

But then, when I put in the Easter bunny image that I made for my kids,

play06:28

it said that it was human-generated.

play06:30

This is because of some big challenges in deepfake detection.

play06:34

Detection tools often only work on one single way to make a deepfake,

play06:37

so you need multiple tools,

play06:39

and they don't work well on low-quality social media content.

play06:43

Confidence score, 0.76-0.87,

play06:47

how do you know whether that's reliable,

play06:48

if you don't know if the underlying technology is reliable,

play06:51

or whether it works on the manipulation that is being used?

play06:54

And tools to spot an AI manipulation don't spot a manual edit.

play07:00

These tools also won't be available to everyone.

play07:04

There's a trade-off between security and access,

play07:07

which means if we make them available to anyone,

play07:09

they become useless to everybody,

play07:12

because the people designing the new deception techniques

play07:15

will test them on the publicly available detectors

play07:18

and evade them.

play07:20

But we do need to make sure these are available

play07:22

to the journalists, the community leaders,

play07:25

the election officials, globally, who are our first line of defense,

play07:28

thought through with attention to real-world accessibility and use.

play07:32

Though at the best circumstances,

play07:35

detection tools will be 85 to 95 percent effective,

play07:38

they have to be in the hands of that first line of defense,

play07:41

and they're not, right now.

play07:43

So for step one, I've been talking about detection after the fact.

play07:46

Step two -- AI is going to be everywhere in our communication,

play07:51

creating, changing, editing.

play07:53

It's not going to be a simple binary of "yes, it's AI" or "phew, it's not."

play07:58

AI is part of all of our communication,

play08:01

so we need to better understand the recipe of what we're consuming.

play08:06

Some people call this content provenance and disclosure.

play08:10

Technologists have been building ways to add invisible watermarking

play08:13

to AI-generated media.

play08:15

They've also been designing ways --

play08:17

and I've been part of these efforts --

play08:19

within a standard called the C2PA,

play08:20

to add cryptographically signed metadata to files.

play08:24

This means data that provides details about the content,

play08:28

cryptographically signed in a way that reinforces our trust

play08:32

in that information.

play08:33

It's an updating record of how AI was used to create or edit it,

play08:39

where humans and other technologies were involved,

play08:41

and how it was distributed.

play08:43

It's basically a recipe and serving instructions

play08:46

for the mix of AI and human

play08:48

that's in what you're seeing and hearing.

play08:51

And it's a critical part of a new AI-infused media literacy.

play08:57

And this actually shouldn't sound that crazy.

play08:59

Our communication is moving in this direction already.

play09:02

If you're like me -- you can admit it --

play09:04

you browse your TikTok “For You” page,

play09:07

and you're used to seeing videos that have an audio source,

play09:11

an AI filter, a green screen, a background,

play09:13

a stitch with another edit.

play09:15

This, in some sense, is the alpha version of this transparency

play09:19

in some of the major platforms we use today.

play09:21

It's just that it does not yet travel across the internet,

play09:24

it’s not reliable, updatable, and it’s not secure.

play09:27

Now, there are also big challenges

play09:30

in this type of infrastructure for authenticity.

play09:34

As we create these durable signs of how AI and human were mixed,

play09:38

that carry across the trajectory of how media is made,

play09:41

we need to ensure they don't compromise privacy or backfire globally.

play09:46

We have to get this right.

play09:48

We can't oblige a citizen journalist filming in a repressive context

play09:52

or a satirical maker using novel gen-AI tools

play09:56

to parody the powerful ...

play09:58

to have to disclose their identity or personally identifiable information

play10:03

in order to use their camera or ChatGPT.

play10:08

Because it's important they be able to retain their ability to have anonymity,

play10:12

at the same time as the tool to create is transparent.

play10:16

This needs to be about the how of AI-human media making,

play10:20

not the who.

play10:22

This brings me to the final step.

play10:25

None of this works without a pipeline of responsibility

play10:29

that runs from the foundation models and the open-source projects

play10:33

through to the way that is deployed into systems, APIs and apps,

play10:38

to the platforms where we consume media and communicate.

play10:43

I've spent much of the last 15 years fighting, essentially, a rearguard action,

play10:47

like so many of my colleagues in the human rights world,

play10:50

against the failures of social media.

play10:52

We can't make those mistakes again in this next generation of technology.

play10:59

What this means is that governments

play11:01

need to ensure that within this pipeline of responsibility for AI,

play11:05

there is transparency, accountability and liability.

play11:10

Without these three steps --

play11:12

detection for the people who need it most,

play11:15

provenance that is rights-respecting

play11:18

and that pipeline of responsibility,

play11:20

we're going to get stuck looking in vain for the six-fingered hand,

play11:24

or the eyes that don't blink.

play11:26

We need to take these steps.

play11:28

Otherwise, we risk a world where it gets easier and easier

play11:31

to both fake reality

play11:33

and dismiss reality as potentially faked.

play11:36

And that is a world that the political philosopher Hannah Arendt

play11:39

described in these terms:

play11:40

"A people that no longer can believe anything

play11:43

cannot make up its own mind.

play11:45

It is deprived not only of its capacity to act

play11:48

but also of its capacity to think and to judge.

play11:52

And with such a people you can then do what you please."

play11:56

That's a world I know none of us want,

play11:58

that I think we can prevent.

play12:00

Thanks.

play12:02

(Cheers and applause)

Rate This

5.0 / 5 (0 votes)

الوسوم ذات الصلة
DeepfakesReality CheckAI EthicsMedia LiteracyHuman RightsFake DetectionAuthenticitySocial MediaTech AdvancesInformation Trust
هل تحتاج إلى تلخيص باللغة الإنجليزية؟