Jeff Dean: AI isn't as smart as you think -- but it could be | TED

TED
12 Jan 202218:31

Summary

TLDR私はGoogleのAIリサーチとヘルスを担当しているジェフです。過去20年で、AIが視覚や言語理解などで大きな進歩を遂げましたが、まだ多くの課題があります。神経ネットワークと計算能力の向上が鍵でしたが、現在のモデルはタスクごとに独立しており、学習方法も非効率です。これを改善するために、複数タスク対応のモデルや異なるデータモダリティの統合、スパースなモデルが必要です。さらに、公平性やプライバシーを考慮した責任あるAI開発が重要です。これにより、医療や教育、環境問題の解決に貢献できると信じています。

Takeaways

  • 🌟 GoogleのAI研究と健康部門をリードしているジェフ・ディーンは、20年以上前に小さなオフィスで始まり、コンピューティングの変革を多くの目にしてきた。
  • 🤖 AIは過去10年間で、画像認識、言語理解、音声認識などの分野で飛躍的な進歩を遂げた。
  • 🔮 AIの進歩は、洪水予測、言語翻訳、病気の予測と診断など、人々の生活に変革をもたらしている。
  • 🧠 ニューラルネットワークは、過去15年で多くの問題解決のための突破的なアプローチとして現れ、実際には1960年代から考え出されているアイデアである。
  • 💪 ニューラルネットワークを効果的に動作させるには、膨大なコンピューティングパワーが必要で、過去15年間でそのパワーが実現された。
  • 🔧 AIの学習プロセスは、重み値の微調整を繰り返し行い、システム全体を望ましい行動に導くことで複雑なタスクをこなすことができる。
  • 🌐 Googleは、YouTubeのビデオからランダムに選択したフレームを使って非常に大きなニューラルネットワークをトレーニングし、物体認識の能力を開発させた。
  • 🛠️ Googleは、ニューラルネットワークの計算に特化したハードウェアであるTensor Processing Unit (TPU)を開発し、検索や翻訳などに使用している。
  • 🚧 AI分野では、現在のモデルが1つのタスクに特化し、多タスクモデルのトレーニングが不足していると指摘された。
  • 🌈 今後のAIモデルは、複数のモダリティ(画像、テキスト、音声など)を扱い、それらを融合してより包括的な世界理解を持つことを目指している。
  • 🏛️ 倫理的なAIの重要性が強調され、AIの開発と使用には、公平性、解釈可能性、プライバシー、セキュリティなどの原則が適用されるべきである。
  • 🌍 AIの進歩は、医療、教育、環境問題など、人類が直面する最大の課題に取り組む力となる可能性がある。

Q & A

  • ジェフはどの企業でどのような役職を務めていますか?

    -ジェフはGoogleでAI研究とヘルスのリーダーを務めています。

  • ジェフがGoogleに入社してからどのくらいの年月が経ちましたか?

    -ジェフは20年以上前にGoogleに入社しました。

  • AIが過去10年間でどのような進歩を遂げていますか?

    -過去10年間で、AIはコンピュータビジョン、言語理解、音声理解などの分野で大きな進歩を遂げています。

  • AIの進歩がもたらした具体的な応用例には何がありますか?

    -AIの進歩は洪水の予測、多言語翻訳、病気の予測と診断などに応用されています。

  • 最近のAIシステムの進歩の2つの重要な要素は何ですか?

    -最近のAIシステムの進歩の2つの重要な要素は神経ネットワークとコンピューテーションパワーです。

  • 神経ネットワークとはどのような概念ですか?

    -神経ネットワークは、実際の神経系のプロパティを緩やかに模倣する、相互接続された人工神経のシリーズです。

  • ジェフがなぜ神経ネットワークに興味を持ち始めましたか?

    -ジェフは1990年代に大学で神経ネットワークの授業を受け、その時点では小さな問題では印象的な結果を示していたものの、実際の大きな問題には拡張できませんでしたが、コンピュートパワーが増加すれば神経ネットワークの可能性が広がると感じ始めました。

  • GoogleのTPUとは何を意味していますか?

    -TPUはTensor Processing Unitの略で、神経ネットワークの計算に特化したハードウェアです。

  • ジェフがAIの現在の取り組み方で何が間違っていると思いますか?

    -ジェフは現在多くのAIモデルが1つのタスクに特化しており、多タスクに対応するモデルのトレーニングが不足していると指摘しています。また、データのモダリティやモデルの密度についても改善の余地があると言っています。

  • AIの将来のビジョンとしてジェフはどのようなシステムを提案していますか?

    -ジェフは多タスクに対応し、すべてのデータモダリティを扱い、効率的なスパースなモデルを用いるAIシステムを提案しています。

  • GoogleのAI原則とはどのようなものですか?

    -GoogleのAI原則は、AI技術を開発する際に公平性、解釈可能性、プライバシー、セキュリティを考慮する原則であり、これらの原則は常に更新されています。

  • ジェフがAIが直面している最大の課題としているものは何ですか?

    -ジェフはAIが既知のタスクから新しいタスクへの一般化をどのように効率的に行うかが最大の課題だとしています。

  • AIの進歩がもたらす可能性についてジェフは何を期待していますか?

    -ジェフはAIが新しいタスクに非常に少数の例で学習できるシステムを構築し、より多くの問題を解決できると期待しています。

  • AIの適用においてジェフはどのようなリスクや問題に注意を払うべきだと述べていますか?

    -ジェフはAIの適用において、その強力なシステムが善や悪のどちらに用いられるかに注意を払い、AIの原則に基づいた慎重で深思熟慮された適用が重要だと述べています。

  • GoogleのAI研究グループはどのようにして研究の独立性を保っていますか?

    -GoogleのAI研究グループは広告グループ、検索グループ、マップグループなどと協力しながらも、基本的な研究を行っており、その多くは公開されています。

Outlines

00:00

🧠 AIの進歩と課題

ジェフはGoogleのAI研究と健康部門をリードし、20年以上にわたってGoogleに所属しています。AIは過去10年間で画像認識、言語理解、音声認識などの分野で大きな進歩を遂げましたが、多くの点でまだ改善の余地があります。AIの進歩はニューラルネットワークとコンピューティングパワーの2つの要因に起因すると述べ、過去15年間でこれらの技術がどのように発展してきたかを説明しています。

05:02

🤖 AIの多面性と改善点

AIの進歩について語る前に、AIが洪水予測、言語翻訳、病気の予測と診断などの分野でどのように役立つかを紹介します。しかし、AIの現在の課題として、特定のタスクに特化したモデルが多く存在し、多タスクに対応するモデルの開発が遅れていると指摘します。また、AIは単一のデータモーダルに依存しており、人間の多感覚的な学習方式とは異なっています。最後に、現在のAIモデルは高密度で活性化されており、効率的なモデルではないと主張しています。

10:04

🛠️ AI改善の3つの方向性

ジェフはAIの改善に向けた3つの方向性について語ります。まず、多タスクに対応するモデルの開発が重要であり、人間の学習過程を模倣して新しいタスクに取り組む際に、既存の知識を活用することができるようにする必要があると述べています。次に、異なるデータモーダルを扱い、それらを融合できるモデルの開発が必要です。最後に、高密度なモデルの代わりに、特定のタスクに必要な部分のみを活性化するスパースなモデルを開発することが望ましいとしています。

15:05

🌐 責任あるAIの開発

AIの進歩に伴い、公平性、解釈可能性、プライバシー、セキュリティなどの観点からの責任あるAIの開発が求められます。Googleは2018年にAI原則を発表し、これらの原則が研究や製品開発に役立つかについて語ります。AIの強力なシステムが全世界の人々にとって益となるように、データの収集やモデルの開発において多様性と公平性に配慮する必要があると強調します。

🔮 AIの未来と社会的課題

ジェフはAIが人間の多様な問題に対処する力を持つと信じており、病気の診断や教育システムの向上、クリーンエネルギーの開発など、AIが社会に与える可能性について語ります。AIの研究は世界中の人々と多分野の専門家の協力なしには進展しないと指摘し、AIの将来について非常に前向きな姿勢を示しています。

🤝 AIの倫理と社会的責任

対話の最後に、AIの倫理的懸念と社会への影響について触れ、AIの適用方法によっては良い影響も悪い影響も生じうる可能性があると認めています。GoogleのAI原則が研究や製品開発の中心に位置しており、その原則に基づいてAIの適切な使用を検討することが重要だと述べています。

Mindmap

Keywords

💡人工知能(AI)

人工知能(AI)とは、人間のように思考し、判断し、学びを進化させるコンピュータシステムのことを指します。このビデオでは、AIがコンピュータビジョンや言語理解、音声認識など、これまで不可能だった多くの分野で進歩を遂げたと述べています。AIの進歩は、洪水予測、言語翻訳、病気の予測と診断など、人々の生活に変革をもたらす重要な役割を果たしています。

💡ニューラルネットワーク

ニューラルネットワークは、人間の神経系を緩やかにモデル化したアルゴリズムであり、AI分野における重要な突破的アプローチです。ビデオでは、過去15年でニューラルネットワークがどのようにさまざまな問題解決に輝いたかについて触れています。ニューラルネットワークは、個々のニューロンが重み付きの入力を受け取り、それに基づいて出力する仕組みを持っており、学習プロセスを通じて複雑なタスクをこなすことができます。

💡コンピューテーションパワー

コンピューテーションパワーとは、コンピュータの計算能力を指し、特にニューラルネットワークを効果的に動作させるために必要とされます。ビデオでは、過去15年間でコンピューテーションパワーが増加し、AIの進歩に寄与したと説明しています。例えば、YouTubeのビデオフレームからランダムに選んだ1000万フレームを使って訓練することで、ニューラルネットワークはさまざまな物体を認識する能力を獲得しています。

💡マルチタスクモデル

マルチタスクモデルとは、単一のタスクに特化したモデルの代わりに、何千もの異なるタスクをこなすことができるAIモデルのことを指します。ビデオでは、人間が新しいスキルを習得する際には、既存の知識を活用することができると比較して、コンピュータも同様にマルチタスクモデルを通じて新しいタスクに迅速に対応することができると述べています。

💡データモダリティ

データモダリティとは、画像、テキスト、音声など、異なるタイプのデータを指し、AIモデルがそれらを一度に処理し、統合することができる能力です。ビデオでは、AIモデルが単一のモダリティに依存するのではなく、複数の感覚を活用して学習し、判断を下す人間の様子を模倣することが重要であると強調しています。

💡スパースモデル

スパースモデルは、タスクに応じてモデルの一部のみを活性化するAIモデルのことを指します。ビデオでは、人間の脳がタスクに応じて異なる部位を活用するように、AIモデルも効率的に機能するためには、必要とされる部分のみを利用することができると述べています。これにより、モデルは高容量であるにもかかわらず、効率的で柔軟性が高くなると提案しています。

💡責任あるAI

責任あるAIとは、AI技術を開発・使用する際に、公正性、解釈可能性、プライバシー、セキュリティを考慮し、すべてのユーザーにとって有益であることを目指す取り組みを指します。ビデオでは、AIの強力なシステムがどのように社会全体に利益をもたらすかについて議論しており、特にデータの収集と代表的な性に関する重要な問題に焦点を当てています。

💡一般化

一般化とは、AIシステムが既知のタスクから新しいタスクへの適用を容易に行う能力を指します。ビデオでは、現在のAIモデルが特定の問題に特化しており、新しいタスクに取り組む際には大量のデータが必要であることが問題とされており、一方でマルチタスクモデルは新しいタスクへの適応が容易になる可能性があると述べています。

💡自己監督学習

自己監督学習は、AIシステムが最小限の例から新しいタスクを学習する能力を指し、ビデオではこれが望ましい目標であると示唆しています。この学習方式により、AIはより少ないデータで新しい問題を解決することができるため、データの収集と処理の効率が向上することが期待できます。

💡Pathways

Pathwaysは、ビデオで言及されているGoogleが開発しているAIシステムの名称であり、多様なタスクをこなし、新しいタスクを徐々に学習し、異なるデータモダリティを扱うことができることを目指しています。このシステムは、AIの進化において重要な一歩として位置づけられており、多様なタスクを一つのモデルで扱うことで、より効率的で柔軟なAIの実現を目指しています。

Highlights

Jeff Dean 是 Google AI 研究和健康部门的负责人,他分享了他在 Google 超过 20 年的经历。

过去十年,人工智能在计算机视觉、语言理解和语音识别方面取得了巨大进步。

机器学习可以更好地预测洪水,翻译超过 100 种语言,并预测和诊断疾病。

神经网络是解决复杂问题的关键方法,尽管它们并非新概念。

计算能力的提升是神经网络成功的关键,过去 15 年中计算能力显著增强。

Jeff Dean 讨论了当前 AI 系统存在的问题,包括训练单一任务模型的低效性。

神经网络的学习过程涉及不断调整权重值,以实现期望的行为。

Jeff Dean 在 1990 年对神经网络的兴趣源于他的本科课程。

Google 训练的大型神经网络能够识别 YouTube 视频中的各种对象,包括猫。

Google 为神经网络计算设计了专门的硬件,即张量处理单元(TPU)。

Jeff Dean 提出了三个 AI 系统需要改进的关键点:多任务模型、多模态数据处理和稀疏激活模型。

多任务模型可以更有效地利用已有的知识,快速学习新任务。

多模态模型可以处理文本、图像、语音等多种数据类型,并将其融合。

稀疏激活模型可以提高效率,仅在需要时激活模型的相关部分。

Google 正在开发一个名为“Pathways”的系统,以实现这些改进。

负责任的 AI 发展需要考虑公平性、可解释性、隐私和安全性。

Google 发布了 AI 原则,以指导技术的发展,并持续更新这些原则。

Jeff Dean 强调了 AI 在解决人类面临的重大问题中的潜力,如疾病诊断、教育系统改进和气候变化。

Jeff Dean 讨论了 AI 系统可能带来的意外后果,并强调了负责任使用 AI 的重要性。

Google 的 AI 研究团队与广告、搜索和地图等团队合作,同时进行开放性的基础研究。

Transcripts

play00:13

Hi, I'm Jeff.

play00:15

I lead AI Research and Health at Google.

play00:18

I joined Google more than 20 years ago,

play00:20

when we were all wedged into a tiny office space,

play00:23

above what's now a T-Mobile store in downtown Palo Alto.

play00:27

I've seen a lot of computing transformations in that time,

play00:30

and in the last decade, we've seen AI be able to do tremendous things.

play00:34

But we're still doing it all wrong in many ways.

play00:37

That's what I want to talk to you about today.

play00:39

But first, let's talk about what AI can do.

play00:41

So in the last decade, we've seen tremendous progress

play00:45

in how AI can help computers see, understand language,

play00:49

understand speech better than ever before.

play00:52

Things that we couldn't do before, now we can do.

play00:54

If you think about computer vision alone,

play00:57

just in the last 10 years,

play00:58

computers have effectively developed the ability to see;

play01:01

10 years ago, they couldn't see, now they can see.

play01:04

You can imagine this has had a transformative effect

play01:06

on what we can do with computers.

play01:08

So let's look at a couple of the great applications

play01:11

enabled by these capabilities.

play01:13

We can better predict flooding, keep everyone safe,

play01:15

using machine learning.

play01:17

We can translate over 100 languages so we all can communicate better,

play01:20

and better predict and diagnose disease,

play01:22

where everyone gets the treatment that they need.

play01:25

So let's look at two key components

play01:27

that underlie the progress in AI systems today.

play01:30

The first is neural networks,

play01:31

a breakthrough approach to solving some of these difficult problems

play01:35

that has really shone in the last 15 years.

play01:37

But they're not a new idea.

play01:39

And the second is computational power.

play01:41

It actually takes a lot of computational power

play01:43

to make neural networks able to really sing,

play01:45

and in the last 15 years, we’ve been able to have that,

play01:49

and that's partly what's enabled all this progress.

play01:51

But at the same time, I think we're doing several things wrong,

play01:55

and that's what I want to talk to you about

play01:57

at the end of the talk.

play01:58

First, a bit of a history lesson.

play02:00

So for decades,

play02:01

almost since the very beginning of computing,

play02:04

people have wanted to be able to build computers

play02:06

that could see, understand language, understand speech.

play02:10

The earliest approaches to this, generally,

play02:12

people were trying to hand-code all the algorithms

play02:14

that you need to accomplish those difficult tasks,

play02:17

and it just turned out to not work very well.

play02:19

But in the last 15 years, a single approach

play02:23

unexpectedly advanced all these different problem spaces all at once:

play02:27

neural networks.

play02:29

So neural networks are not a new idea.

play02:31

They're kind of loosely based

play02:32

on some of the properties that are in real neural systems.

play02:35

And many of the ideas behind neural networks

play02:37

have been around since the 1960s and 70s.

play02:40

A neural network is what it sounds like,

play02:42

a series of interconnected artificial neurons

play02:45

that loosely emulate the properties of your real neurons.

play02:48

An individual neuron in one of these systems

play02:50

has a set of inputs,

play02:51

each with an associated weight,

play02:53

and the output of a neuron

play02:55

is a function of those inputs multiplied by those weights.

play02:59

So pretty simple,

play03:00

and lots and lots of these work together to learn complicated things.

play03:04

So how do we actually learn in a neural network?

play03:07

It turns out the learning process

play03:09

consists of repeatedly making tiny little adjustments

play03:11

to the weight values,

play03:13

strengthening the influence of some things,

play03:15

weakening the influence of others.

play03:17

By driving the overall system towards desired behaviors,

play03:21

these systems can be trained to do really complicated things,

play03:24

like translate from one language to another,

play03:27

detect what kind of objects are in a photo,

play03:30

all kinds of complicated things.

play03:32

I first got interested in neural networks

play03:34

when I took a class on them as an undergraduate in 1990.

play03:37

At that time,

play03:38

neural networks showed impressive results on tiny problems,

play03:42

but they really couldn't scale to do real-world important tasks.

play03:46

But I was super excited.

play03:48

(Laughter)

play03:50

I felt maybe we just needed more compute power.

play03:52

And the University of Minnesota had a 32-processor machine.

play03:56

I thought, "With more compute power,

play03:58

boy, we could really make neural networks really sing."

play04:01

So I decided to do a senior thesis on parallel training of neural networks,

play04:05

the idea of using processors in a computer or in a computer system

play04:09

to all work toward the same task,

play04:11

that of training neural networks.

play04:12

32 processors, wow,

play04:14

we’ve got to be able to do great things with this.

play04:17

But I was wrong.

play04:20

Turns out we needed about a million times as much computational power

play04:23

as we had in 1990

play04:24

before we could actually get neural networks to do impressive things.

play04:28

But starting around 2005,

play04:30

thanks to the computing progress of Moore's law,

play04:33

we actually started to have that much computing power,

play04:35

and researchers in a few universities around the world started to see success

play04:40

in using neural networks for a wide variety of different kinds of tasks.

play04:44

I and a few others at Google heard about some of these successes,

play04:47

and we decided to start a project to train very large neural networks.

play04:51

One system that we trained,

play04:52

we trained with 10 million randomly selected frames

play04:56

from YouTube videos.

play04:57

The system developed the capability

play04:59

to recognize all kinds of different objects.

play05:01

And it being YouTube, of course,

play05:03

it developed the ability to recognize cats.

play05:05

YouTube is full of cats.

play05:07

(Laughter)

play05:08

But what made that so remarkable

play05:10

is that the system was never told what a cat was.

play05:13

So using just patterns in data,

play05:16

the system honed in on the concept of a cat all on its own.

play05:20

All of this occurred at the beginning of a decade-long string of successes,

play05:24

of using neural networks for a huge variety of tasks,

play05:26

at Google and elsewhere.

play05:27

Many of the things you use every day,

play05:30

things like better speech recognition for your phone,

play05:32

improved understanding of queries and documents

play05:34

for better search quality,

play05:36

better understanding of geographic information to improve maps,

play05:39

and so on.

play05:40

Around that time,

play05:41

we also got excited about how we could build hardware that was better tailored

play05:45

to the kinds of computations neural networks wanted to do.

play05:48

Neural network computations have two special properties.

play05:51

The first is they're very tolerant of reduced precision.

play05:53

Couple of significant digits, you don't need six or seven.

play05:56

And the second is that all the algorithms are generally composed

play05:59

of different sequences of matrix and vector operations.

play06:03

So if you can build a computer

play06:05

that is really good at low-precision matrix and vector operations

play06:09

but can't do much else,

play06:10

that's going to be great for neural-network computation,

play06:13

even though you can't use it for a lot of other things.

play06:16

And if you build such things, people will find amazing uses for them.

play06:19

This is the first one we built, TPU v1.

play06:21

"TPU" stands for Tensor Processing Unit.

play06:24

These have been used for many years behind every Google search,

play06:27

for translation,

play06:28

in the DeepMind AlphaGo matches,

play06:30

so Lee Sedol and Ke Jie maybe didn't realize,

play06:33

but they were competing against racks of TPU cards.

play06:35

And we've built a bunch of subsequent versions of TPUs

play06:38

that are even better and more exciting.

play06:40

But despite all these successes,

play06:42

I think we're still doing many things wrong,

play06:44

and I'll tell you about three key things we're doing wrong,

play06:47

and how we'll fix them.

play06:48

The first is that most neural networks today

play06:50

are trained to do one thing, and one thing only.

play06:53

You train it for a particular task that you might care deeply about,

play06:56

but it's a pretty heavyweight activity.

play06:58

You need to curate a data set,

play06:59

you need to decide what network architecture you'll use

play07:02

for this problem,

play07:04

you need to initialize the weights with random values,

play07:06

apply lots of computation to make adjustments to the weights.

play07:09

And at the end, if you’re lucky, you end up with a model

play07:12

that is really good at that task you care about.

play07:14

But if you do this over and over,

play07:16

you end up with thousands of separate models,

play07:19

each perhaps very capable,

play07:21

but separate for all the different tasks you care about.

play07:23

But think about how people learn.

play07:25

In the last year, many of us have picked up a bunch of new skills.

play07:28

I've been honing my gardening skills,

play07:30

experimenting with vertical hydroponic gardening.

play07:32

To do that, I didn't need to relearn everything I already knew about plants.

play07:36

I was able to know how to put a plant in a hole,

play07:40

how to pour water, that plants need sun,

play07:42

and leverage that in learning this new skill.

play07:46

Computers can work the same way, but they don’t today.

play07:49

If you train a neural network from scratch,

play07:51

it's effectively like forgetting your entire education

play07:55

every time you try to do something new.

play07:57

That’s crazy, right?

play07:58

So instead, I think we can and should be training

play08:02

multitask models that can do thousands or millions of different tasks.

play08:06

Each part of that model would specialize in different kinds of things.

play08:09

And then, if we have a model that can do a thousand things,

play08:12

and the thousand and first thing comes along,

play08:14

we can leverage the expertise we already have

play08:16

in the related kinds of things

play08:18

so that we can more quickly be able to do this new task,

play08:21

just like you, if you're confronted with some new problem,

play08:24

you quickly identify the 17 things you already know

play08:26

that are going to be helpful in solving that problem.

play08:29

Second problem is that most of our models today

play08:32

deal with only a single modality of data --

play08:34

with images, or text or speech,

play08:37

but not all of these all at once.

play08:39

But think about how you go about the world.

play08:41

You're continuously using all your senses

play08:43

to learn from, react to,

play08:46

figure out what actions you want to take in the world.

play08:49

Makes a lot more sense to do that,

play08:50

and we can build models in the same way.

play08:52

We can build models that take in these different modalities of input data,

play08:57

text, images, speech,

play08:58

but then fuse them together,

play09:00

so that regardless of whether the model sees the word "leopard,"

play09:04

sees a video of a leopard or hears someone say the word "leopard,"

play09:08

the same response is triggered inside the model:

play09:10

the concept of a leopard

play09:12

can deal with different kinds of input data,

play09:14

even nonhuman inputs, like genetic sequences,

play09:17

3D clouds of points, as well as images, text and video.

play09:20

The third problem is that today's models are dense.

play09:24

There's a single model,

play09:25

the model is fully activated for every task,

play09:28

for every example that we want to accomplish,

play09:30

whether that's a really simple or a really complicated thing.

play09:33

This, too, is unlike how our own brains work.

play09:36

Different parts of our brains are good at different things,

play09:39

and we're continuously calling upon the pieces of them

play09:42

that are relevant for the task at hand.

play09:44

For example, nervously watching a garbage truck

play09:47

back up towards your car,

play09:49

the part of your brain that thinks about Shakespearean sonnets

play09:51

is probably inactive.

play09:53

(Laughter)

play09:54

AI models can work the same way.

play09:56

Instead of a dense model,

play09:57

we can have one that is sparsely activated.

play10:00

So for particular different tasks, we call upon different parts of the model.

play10:04

During training, the model can also learn which parts are good at which things,

play10:08

to continuously identify what parts it wants to call upon

play10:12

in order to accomplish a new task.

play10:14

The advantage of this is we can have a very high-capacity model,

play10:18

but it's very efficient,

play10:19

because we're only calling upon the parts that we need

play10:21

for any given task.

play10:23

So fixing these three things, I think,

play10:25

will lead to a more powerful AI system:

play10:27

instead of thousands of separate models,

play10:29

train a handful of general-purpose models

play10:31

that can do thousands or millions of things.

play10:33

Instead of dealing with single modalities,

play10:35

deal with all modalities,

play10:36

and be able to fuse them together.

play10:38

And instead of dense models, use sparse, high-capacity models,

play10:42

where we call upon the relevant bits as we need them.

play10:45

We've been building a system that enables these kinds of approaches,

play10:48

and we’ve been calling the system “Pathways.”

play10:51

So the idea is this model will be able to do

play10:54

thousands or millions of different tasks,

play10:56

and then, we can incrementally add new tasks,

play10:58

and it can deal with all modalities at once,

play11:00

and then incrementally learn new tasks as needed

play11:03

and call upon the relevant bits of the model

play11:06

for different examples or tasks.

play11:07

And we're pretty excited about this,

play11:09

we think this is going to be a step forward

play11:11

in how we build AI systems.

play11:12

But I also wanted to touch on responsible AI.

play11:16

We clearly need to make sure that this vision of powerful AI systems

play11:21

benefits everyone.

play11:23

These kinds of models raise important new questions

play11:25

about how do we build them with fairness,

play11:28

interpretability, privacy and security,

play11:31

for all users in mind.

play11:33

For example, if we're going to train these models

play11:35

on thousands or millions of tasks,

play11:38

we'll need to be able to train them on large amounts of data.

play11:40

And we need to make sure that data is thoughtfully collected

play11:44

and is representative of different communities and situations

play11:47

all around the world.

play11:49

And data concerns are only one aspect of responsible AI.

play11:53

We have a lot of work to do here.

play11:55

So in 2018, Google published this set of AI principles

play11:57

by which we think about developing these kinds of technology.

play12:01

And these have helped guide us in how we do research in this space,

play12:05

how we use AI in our products.

play12:06

And I think it's a really helpful and important framing

play12:09

for how to think about these deep and complex questions

play12:12

about how we should be using AI in society.

play12:15

We continue to update these as we learn more.

play12:19

Many of these kinds of principles are active areas of research --

play12:22

super important area.

play12:24

Moving from single-purpose systems that kind of recognize patterns in data

play12:28

to these kinds of general-purpose intelligent systems

play12:31

that have a deeper understanding of the world

play12:33

will really enable us to tackle

play12:34

some of the greatest problems humanity faces.

play12:37

For example,

play12:38

we’ll be able to diagnose more disease;

play12:41

we'll be able to engineer better medicines

play12:43

by infusing these models with knowledge of chemistry and physics;

play12:46

we'll be able to advance educational systems

play12:48

by providing more individualized tutoring

play12:50

to help people learn in new and better ways;

play12:53

we’ll be able to tackle really complicated issues,

play12:55

like climate change,

play12:56

and perhaps engineering of clean energy solutions.

play12:59

So really, all of these kinds of systems

play13:02

are going to be requiring the multidisciplinary expertise

play13:05

of people all over the world.

play13:07

So connecting AI with whatever field you are in,

play13:10

in order to make progress.

play13:13

So I've seen a lot of advances in computing,

play13:15

and how computing, over the past decades,

play13:18

has really helped millions of people better understand the world around them.

play13:22

And AI today has the potential to help billions of people.

play13:26

We truly live in exciting times.

play13:28

Thank you.

play13:29

(Applause)

play13:39

Chris Anderson: Thank you so much.

play13:41

I want to follow up on a couple things.

play13:44

This is what I heard.

play13:47

Most people's traditional picture of AI

play13:51

is that computers recognize a pattern of information,

play13:54

and with a bit of machine learning,

play13:56

they can get really good at that, better than humans.

play13:59

What you're saying is those patterns

play14:01

are no longer the atoms that AI is working with,

play14:04

that it's much richer-layered concepts

play14:06

that can include all manners of types of things

play14:10

that go to make up a leopard, for example.

play14:13

So what could that lead to?

play14:16

Give me an example of when that AI is working,

play14:18

what do you picture happening in the world

play14:20

in the next five or 10 years that excites you?

play14:23

Jeff Dean: I think the grand challenge in AI

play14:26

is how do you generalize from a set of tasks

play14:28

you already know how to do

play14:29

to new tasks,

play14:31

as easily and effortlessly as possible.

play14:33

And the current approach of training separate models for everything

play14:36

means you need lots of data about that particular problem,

play14:40

because you're effectively trying to learn everything

play14:42

about the world and that problem, from nothing.

play14:45

But if you can build these systems

play14:46

that already are infused with how to do thousands and millions of tasks,

play14:51

then you can effectively teach them to do a new thing

play14:55

with relatively few examples.

play14:56

So I think that's the real hope,

play14:58

that you could then have a system where you just give it five examples

play15:03

of something you care about,

play15:04

and it learns to do that new task.

play15:06

CA: You can do a form of self-supervised learning

play15:09

that is based on remarkably little seeding.

play15:11

JD: Yeah, as opposed to needing 10,000 or 100,000 examples

play15:14

to figure everything in the world out.

play15:16

CA: Aren't there kind of terrifying unintended consequences

play15:19

possible, from that?

play15:20

JD: I think it depends on how you apply these systems.

play15:23

It's very clear that AI can be a powerful system for good,

play15:27

or if you apply it in ways that are not so great,

play15:29

it can be a negative consequence.

play15:33

So I think that's why it's important to have a set of principles

play15:36

by which you look at potential uses of AI

play15:38

and really are careful and thoughtful about how you consider applications.

play15:43

CA: One of the things people worry most about

play15:45

is that, if AI is so good at learning from the world as it is,

play15:49

it's going to carry forward into the future

play15:52

aspects of the world as it is that actually aren't right, right now.

play15:56

And there's obviously been a huge controversy about that

play15:59

recently at Google.

play16:02

Some of those principles of AI development,

play16:05

you've been challenged that you're not actually holding to them.

play16:10

Not really interested to hear about comments on a specific case,

play16:13

but ... are you really committed?

play16:16

How do we know that you are committed to these principles?

play16:19

Is that just PR, or is that real, at the heart of your day-to-day?

play16:23

JD: No, that is absolutely real.

play16:25

Like, we have literally hundreds of people

play16:27

working on many of these related research issues,

play16:29

because many of those things are research topics

play16:31

in their own right.

play16:33

How do you take data from the real world,

play16:35

that is the world as it is, not as we would like it to be,

play16:40

and how do you then use that to train a machine-learning model

play16:43

and adapt the data bit of the scene

play16:46

or augment the data with additional data

play16:48

so that it can better reflect the values we want the system to have,

play16:51

not the values that it sees in the world?

play16:54

CA: But you work for Google,

play16:56

Google is funding the research.

play16:59

How do we know that the main values that this AI will build

play17:03

are for the world,

play17:05

and not, for example, to maximize the profitability of an ad model?

play17:10

When you know everything there is to know about human attention,

play17:13

you're going to know so much

play17:14

about the little wriggly, weird, dark parts of us.

play17:17

In your group, are there rules about how you hold off,

play17:23

church-state wall between a sort of commercial push,

play17:27

"You must do it for this purpose,"

play17:29

so that you can inspire your engineers and so forth,

play17:31

to do this for the world, for all of us.

play17:33

JD: Yeah, our research group does collaborate

play17:36

with a number of groups across Google,

play17:37

including the Ads group, the Search group, the Maps group,

play17:40

so we do have some collaboration, but also a lot of basic research

play17:43

that we publish openly.

play17:45

We've published more than 1,000 papers last year

play17:48

in different topics, including the ones you discussed,

play17:51

about fairness, interpretability of the machine-learning models,

play17:54

things that are super important,

play17:56

and we need to advance the state of the art in this

play17:58

in order to continue to make progress

play18:01

to make sure these models are developed safely and responsibly.

play18:04

CA: It feels like we're at a time when people are concerned

play18:07

about the power of the big tech companies,

play18:09

and it's almost, if there was ever a moment to really show the world

play18:13

that this is being done to make a better future,

play18:17

that is actually key to Google's future,

play18:19

as well as all of ours.

play18:21

JD: Indeed.

play18:22

CA: It's very good to hear you come and say that, Jeff.

play18:25

Thank you so much for coming here to TED.

play18:27

JD: Thank you.

play18:28

(Applause)

Rate This

5.0 / 5 (0 votes)

Related Tags
人工知能Google研究マルチタスクニューラルネットワークデータ分析技術進歩責任AI未来予測教育
Do you need a summary in English?