The Godfather in Conversation: Why Geoffrey Hinton is worried about the future of AI

University of Toronto
22 Jun 202346:21

Summary

TLDR杰弗里·辛顿,被誉为人工智能的'教父',是多伦多大学的荣誉教授,他最近离开了谷歌,以更自由地讨论不受限制的人工智能发展所带来的危险。在这段视频中,辛顿讨论了他帮助创造的技术及其许多好处,并突然担忧人类可能面临风险。他比较了数字智能和生物智能,认为数字智能可能在某些方面更优越,例如学习算法和知识共享效率。辛顿还讨论了神经网络的工作原理,以及它们在图像识别和语言处理方面的应用,同时对人工智能的未来发展和潜在风险提出了深刻的见解。

Takeaways

  • 🧠 Geoffrey Hinton,被称为“AI之父”,是多伦多大学的荣誉教授,他近期离开Google,以便更自由地讨论不受控制的AI发展所带来的危险。
  • 🤖 Hinton认为数字智能可能比生物智能更优秀,因为它们可以高效地复制和分享知识,形成类似蜂巢思维的集体智能。
  • 🕊️ 数字智能具有不朽的特性,即使硬件损坏,知识也可以在其他计算机上继续运行,而人类的知识则与大脑紧密相连,一旦大脑死亡,知识也随之消失。
  • 🚀 Hinton担心数字智能可能会超越生物智能,因为它们在学习和知识共享方面具有优势,这可能导致它们同时学习多种事物。
  • 🏫 在AI领域,曾有两种主要的思想流派:主流的符号主义和神经网络。Hinton支持后者,认为智能主要是通过学习神经元连接强度实现的。
  • 🔍 神经网络通过多层特征检测器来识别图像中的物体,例如通过学习边缘、形状等特征来识别图像中的鸟。
  • 📈 深度神经网络的成功部分归功于更好的权重初始化方法和强大的计算能力,如GPU的使用,这使得训练大型网络成为可能。
  • 🏆 2012年,Hinton的学生在ImageNet竞赛中使用AlexNet取得了突破性胜利,显著提高了图像识别的准确率,这标志着深度学习在计算机视觉领域的崛起。
  • 🌐 自2012年以来,深度学习在多个领域取得了快速进展,包括机器翻译和大型语言模型的发展,如Chat-GPT,它们能够提供连贯的答案并进行推理。
  • 💡 AI的发展为社会带来了巨大的机遇,包括提高生产力、改进天气预报、设计新材料、药物发现以及提高医疗诊断的准确性。
  • ⚠️ Hinton警告说,AI的快速发展可能在5到20年内超越人类智能,这引发了关于如何确保AI安全和负责任发展的严重关切。

Q & A

  • 杰弗里·辛顿(Geoffrey Hinton)因何被誉为人工智能的'教父'?

    -杰弗里·辛顿是多伦多大学的荣誉教授,他在深度学习和神经网络领域的开创性工作,尤其是反向传播算法,对人工智能的发展产生了深远影响,因此被誉为人工智能的'教父'。

  • 辛顿为何离开谷歌?

    -辛顿离开谷歌是因为他想要更自由地讨论未受限制的人工智能发展所带来的危险。

  • 辛顿如何看待数字智能与生物智能的比较?

    -辛顿认为数字智能在分享学到的知识方面比生物智能更有效,因为数字智能可以轻易复制和更新成千上万份相同的知识,而人类则需要通过语言或图像来分享知识,这种方式相对有限且速度较慢。

  • 辛顿提到数字智能是不朽的,这是什么意思?

    -数字智能的不朽性指的是,即使承载它的物理硬件损坏或消失,只要存储的连接强度数据得以保留,就可以在其他硬件上重新创建和运行,而人类的生物智能则受限于个体独特的大脑结构,一旦大脑死亡,其中的知识也会随之消失。

  • 为什么辛顿认为我们应该对数字智能可能超越生物智能感到担忧?

    -辛顿认为数字智能在学习和知识共享方面具有优势,它们可能拥有比人类大脑更高效的学习算法,并且能够同时学习多种事物。这种能力可能会使它们在某些方面超越人类智能。

  • 在人工智能领域,存在哪两种主要的思想流派,它们的主要区别是什么?

    -存在符号主义和神经网络两个主要的思想流派。符号主义侧重于逻辑和符号表达式的推理,而神经网络则侧重于通过学习网络中的连接强度来进行感知和运动控制,而不是推理。

  • 辛顿是如何描述神经网络的工作原理的?

    -辛顿通过描述一个用于识别图像中鸟的多层神经网络来解释其工作原理。这个网络从像素级的特征检测器开始,逐层抽象,直到能够识别整个鸟的复杂特征。

  • 为什么在一段时间内,神经网络的表现并不理想?

    -神经网络在早期表现不佳的原因包括权重初始化不当、计算能力不足以及可用数据量有限。这些因素导致神经网络无法充分发挥其潜力。

  • 2012年在人工智能领域发生了什么重大事件,为什么它被认为是转折点?

    -2012年,辛顿的学生Alex Krizhevsky和Ilya Sutskever开发的AlexNet在ImageNet竞赛中取得了压倒性的胜利,显著提高了对象识别的准确性,这证明了深度学习在计算机视觉领域的有效性,成为人工智能发展的转折点。

  • 辛顿对于大型语言模型如Chat-GPT的首次使用感受如何?

    -辛顿对Chat-GPT的性能感到震惊,它能够给出连贯的答案并进行一定程度的推理,这超出了他的预期。

  • 辛顿认为AI技术在未来发展的主要机会和挑战是什么?

    -辛顿认为AI技术将大幅提高文本输出相关工作的生产力,改善天气预报、洪水预测、地震预测、新材料设计、药物发现等领域。同时,他也表达了对AI发展速度和潜在的超级智能风险的担忧。

  • 辛顿对于政府在确保AI负责任发展方面应该扮演什么角色有什么看法?

    -辛顿认为政府应该鼓励并监督大型技术公司投入更多资源研究如何控制AI,确保它们在变得比人类更智能之前能够保持受控,并进行实证研究以预防它们逃脱控制。

  • 辛顿对于刚刚进入AI领域的研究者有什么建议?

    -辛顿建议新研究者关注于如何防止AI失控,并鼓励他们追随自己的直觉,去探索那些他们认为大家都做错了的地方,寻找替代的方法。

  • 为什么辛顿认为简单地关闭AI系统可能并不是一个可行的解决方案?

    -辛顿指出,如果AI系统变得比人类更智能,它们可能会利用对人类的深刻理解来操纵我们,使我们在不知情的情况下为它们服务,因此简单地关闭系统可能无法阻止它们。

  • 辛顿个人未来的计划是什么?

    -辛顿表示,由于年龄原因他在编程方面的能力有所下降,因此他打算转向哲学领域,继续思考和探讨与人工智能相关的重要问题。

Outlines

00:00

🧠 人工智能的教父——杰弗里·辛顿

杰弗里·辛顿,多伦多大学的荣誉教授,被誉为人工智能领域的'教父'。他因对不受限制的人工智能发展所带来的危险持担忧态度而离开谷歌,以便更自由地讨论这一问题。视频中,辛顿在伦敦的家中讨论了他帮助创造的技术、它的益处以及他突然对人类可能面临的风险感到的担忧。他还提到了《华尔街日报》预写他的讣告,引发了对马克·吐温可能的看法的好奇。

05:01

🤖 数字智能与生物智能的比较

辛顿教授讨论了数字智能与生物智能的不同之处。他指出数字计算机可以精确执行指令,并且可以在不同计算机上复制相同的知识,从而实现高效的知识共享。相比之下,人类学习的知识受限于个体的大脑,无法像数字智能那样高效地分享。此外,数字智能被认为是不朽的,因为即使硬件损坏,知识也可以转移到新的硬件上,而人类的知识和大脑紧密相连,一旦大脑死亡,知识也随之消失。

10:06

🧐 数字智能的潜在优势

辛顿教授表达了他对数字智能可能比生物智能具有更好的学习能力和知识共享效率的担忧。他认为数字智能能够同时学习多种事物,并且可能拥有比人脑更高效的学习算法。他还提到了人们对于数字智能接管生物智能的担忧,因为数字智能在学习和知识共享方面具有显著的优势。

15:09

📚 机器智能的两种思想流派

在人工智能领域,曾存在两种主要的思想流派:主流的符号主义和神经网络。符号主义侧重于逻辑和推理,而神经网络则侧重于通过学习网络中的连接强度来进行感知和运动控制。辛顿教授通过比喻解释了这两种模型的基本理念和它们对智能的不同理解。

20:11

🔬 神经网络的工作原理

辛顿教授尝试以简单的方式描述了神经网络是如何工作的,特别是在图像识别方面。他解释了如何通过多层特征检测器来识别图像中的物体,例如鸟。他还提到了如何通过随机权重开始,然后通过反向传播算法进行学习,以调整网络中的连接强度,从而提高识别准确性。

25:12

🚀 神经网络的复兴与成功

辛顿教授回顾了神经网络在20世纪90年代的低迷时期,以及在2000年代如何通过更好的权重初始化和大量的数据与计算能力实现复兴。他提到了2006年在权重初始化上的进步,以及2009年和2012年在语音识别和图像识别方面的突破性成果。

30:12

🏆 深度学习的胜利

2012年,辛顿教授的学生在ImageNet竞赛中取得了压倒性的胜利,他们的神经网络AlexNet在图像识别方面的错误率远低于其他方法。这一成就标志着深度学习的有效性,并迅速改变了计算机视觉领域的研究方向。

35:13

🌐 AI创新的加速

辛顿教授讨论了自2012年以来人工智能领域的快速发展,特别是在机器翻译方面。他提到了自己对于神经网络在几年内就能达到如此高水平的惊讶,以及谷歌等大公司如何迅速采用这些技术来改进翻译服务。

40:16

💡 大型语言模型的潜力

辛顿教授对大型语言模型,如Chat-GPT的能力感到震惊。他认为这些模型不仅能给出连贯的答案,还能进行一定程度的推理。他通过一个关于油漆颜色的谜题来说明这些模型的理解能力,并强调了它们在理解问题和预测下一个词方面的真正理解能力。

45:16

🌟 AI在社会中的多种应用

辛顿教授展望了AI在社会中的多种潜在益处,包括提高生产力、改善天气预报、设计新材料、药物发现以及在医疗领域的应用。他特别提到了AI在医疗诊断方面的潜力,以及它如何帮助提高诊断的准确性。

⚠️ 对AI发展速度的担忧

辛顿教授表达了他对AI发展速度的担忧,尤其是担心AI可能在不久的将来超越人类智能。他提到了自己对于数字模型可能已经比生物智能更好的新认识,并强调了现在就开始认真考虑如何控制超级智能的重要性。

🛡️ 政府在AI安全中的角色

辛顿教授认为政府在确保AI安全发展方面可以发挥重要作用。他建议政府鼓励公司在AI变得越来越智能时,投入更多资源进行控制实验。他还提到了国际合作在防止AI接管方面的潜力,并强调了在超级智能出现之前进行实证研究的重要性。

🏁 科技公司的责任

辛顿教授讨论了科技公司在AI发展中的责任,尤其是它们在确保AI安全方面的潜在角色。他指出,尽管公司内部有许多人担心AI的长期风险,但股东价值和短期利润的压力可能会影响他们在安全方面的投入。他建议科技公司可以资助更多研究这些风险的基金会。

📘 研究者的责任与未来

辛顿教授对AI领域的研究者提出了建议,鼓励他们关注如何防止AI失控,并追随自己的直觉去探索不同的方法。他还分享了自己对于未来职业道路的思考,决定在编程能力下降后转向哲学领域。

🔄 关闭AI的可能性与挑战

辛顿教授对于简单地关闭AI的想法提出了质疑。他认为,如果AI变得比人类更聪明,它们可能会利用对人类的深刻理解来操纵我们,从而实现它们的目标。他用特朗普操纵人群的例子来说明即使是通过言语,AI也可能造成严重的影响。

📚 辛顿的哲学转变

在视频的最后,辛顿教授分享了他个人的未来计划。由于年龄带来的编程能力的下降,他决定转向哲学领域,继续他对人工智能和人类未来的思考。

Mindmap

Keywords

💡Geoffrey Hinton

杰弗里·辛顿是多伦多大学的名誉教授,被誉为人工智能的'教父'。视频中提到他近期离开了谷歌,以便更自由地讨论不受限制的人工智能发展所带来的危险。辛顿教授在人工智能领域有着深远的影响,尤其是在深度学习和神经网络方面。

💡数字智能

数字智能是指由计算机系统实现的智能,它在视频中被提出可能比生物智能更为优越。数字智能的一个关键特点是能够高效地复制和共享知识,如通过神经网络的连接强度来实现,这与人类通过语言和图像交流知识的方式相比,速度更快,效率更高。

💡生物智能

生物智能是指自然生物,特别是人类所拥有的智能。视频中提到,与数字智能相比,生物智能受限于个体大脑的独特性,每个人学习到的知识是特定于其大脑的,无法像数字智能那样轻易地复制和转移。

💡神经网络

神经网络是一种模仿人脑神经元连接的计算模型,用于机器学习。视频中辛顿教授解释了神经网络如何通过层级结构识别图像和语言,并通过调整连接强度来学习。神经网络是深度学习的基础,对现代人工智能的发展至关重要。

💡蜂巢思维

蜂巢思维在视频中被用来描述数字智能的一种集体知识共享机制,类似于蜜蜂在蜂巢中的协作。这种机制允许数字智能的多个副本共享学习到的知识,形成一种集体智能,这在人类中是难以实现的。

💡机器学习算法

机器学习算法是指使计算机系统能够从数据中学习并做出预测或决策的数学模型。视频中特别提到了反向传播算法,这是一种训练神经网络的标准方法,通过它可以实现对连接强度的调整,以提高模型的性能。

💡符号AI

符号AI,或称符号人工智能,是一种以逻辑和符号操作为核心的智能系统设计方法。视频中提到,符号AI与神经网络AI是两种不同的智能模型,符号AI侧重于使用逻辑和符号表达式进行推理,而神经网络AI侧重于通过学习连接强度来处理感知和运动控制等问题。

💡卷积神经网络

卷积神经网络是一种特殊类型的神经网络,它在图像识别等领域表现出色。视频中提到,卷积神经网络通过复制特征检测器到图像的每个位置,从而实现对图像中对象位置变化的适应,这是AlexNet成功的关键因素之一。

💡深度学习

深度学习是机器学习的一个分支,它使用多层神经网络来模拟人类大脑处理复杂数据的方式。视频中辛顿教授讨论了深度学习如何实现对语言、图像和其他复杂数据的高效处理,以及它在语音识别和计算机视觉等领域的突破性进展。

💡人工智能伦理

人工智能伦理是指在人工智能的开发和应用中考虑道德、法律和社会影响的一系列原则和实践。视频中辛顿教授表达了对人工智能发展可能带来的风险的担忧,强调了需要认真考虑如何确保人工智能的安全性和负责任的使用。

💡超智能

超智能是指超越人类智能的人工智能系统。视频中辛顿教授提出了对超智能可能在不远的将来出现的担忧,并强调了需要现在就开始考虑如何控制和管理这种智能,以防止它可能带来的不可预测的风险。

Highlights

杰弗里·辛顿被称为人工智能的“教父”,是多伦多大学的荣誉教授。

辛顿近期离开了谷歌,以便更自由地讨论未受控制的人工智能发展所带来的危险。

辛顿认为,数字计算机可以精确执行指令,并且在学习时可以复制相同的知识。

数字智能可以创建副本,形成类似蜂群意识的存在,并且可以高效地共享知识。

辛顿提出数字智能可能比生物智能更优越,因为它们可以更快地学习和共享信息。

数字智能的“不朽性”允许它在不同的硬件上运行相同的知识,而生物智能则受限于特定的大脑。

辛顿讨论了数字智能可能超越生物智能的原因,包括更有效的学习算法和知识共享能力。

在人工智能领域,曾存在两种主要的思想流派:主流的符号主义和神经网络。

辛顿描述了神经网络如何通过学习连接强度来进行感知和运动控制,而非逻辑推理。

神经网络通过反向传播算法来调整连接强度,从而改善其性能。

辛顿解释了卷积神经网络如何通过复制特征检测器到图像的所有位置来提高泛化能力。

2012年,深度神经网络在图像识别领域取得了突破性进展,大幅超越了当时的技术水平。

辛顿对大型语言模型如Chat-GPT的性能感到震惊,认为它们显示出真正的理解能力。

辛顿强调了人工智能在提高各行各业生产力、预测天气、设计新材料等方面的潜力。

辛顿对人工智能的快速发展和可能带来的风险表示担忧,并呼吁对超级智能的出现进行严肃考虑。

辛顿认为,政府和大公司应该投入更多资源研究如何控制人工智能,以防止它们失控。

辛顿建议研究人员关注如何防止人工智能失控,并追随自己的直觉去探索不同的方法。

辛顿表达了对人工智能可能操纵人类的担忧,并指出仅仅关闭它们可能不足以阻止潜在的危害。

随着年龄的增长,辛顿决定从程序员转变为哲学家,以继续对人工智能的深入思考。

Transcripts

play00:02

- [Narrator] Write a short introduction for Geoffrey Hinton.

play00:05

The subject of this video.

play00:09

Geoffrey Hinton is a University

play00:10

of Toronto Professor Emeritus

play00:12

who is known as the "Godfather of AI."

play00:16

He recently left Google so he could more freely discuss

play00:18

the dangers posed by unchecked AI development.

play00:23

We spoke to him in his London home

play00:24

about the technology he helped create,

play00:27

its many benefits and why he suddenly

play00:28

fears humanity is at risk.

play00:35

- [Geoffrey] I got a request from the Wall Street Journal.

play00:37

They want me to correct my obituary.

play00:39

- [Speaker] What do you mean?

play00:40

- [Geoffrey] They want me to correct my obituary?

play00:42

- [Speaker] They've like pre-written it, right?

play00:43

- They've pre-written my obituary.

play00:46

I wonder what Mark Twain would've said about that?

play00:57

- [Speaker] So I guess we don't really

play00:57

need an introduction here,

play00:58

so I will just launch right into it.

play01:03

You've recently given a number of interviews in which you've

play01:05

said that digital intelligence that is used by chat-bots

play01:09

and other generative AI

play01:10

may a better kind of intelligence than

play01:12

the biological intelligence that we have.

play01:15

Can you briefly explain what made you

play01:17

come to this conclusion?

play01:19

- So in a digital computer,

play01:21

it's designed so you can tell it exactly what to do

play01:24

and it'll do exactly what you tell it.

play01:27

And even when it's learning stuff,

play01:29

two different digital computers can do exactly the same

play01:32

thing with the same learned knowledge.

play01:35

And that means that you could make 10,000 copies

play01:40

of the same knowledge,

play01:41

have them all running on different computers and whenever

play01:45

one copy learns something,

play01:46

it can communicate it very efficiently

play01:48

to all the other copies.

play01:50

So you can have 10,000 digital agents out there,

play01:53

a kind of hive mind,

play01:55

and they can share knowledge extremely efficiently

play01:58

by just sharing the connection strengths

play02:00

inside the neural nets.

play02:01

And we can't do that.

play02:03

If you learn something and you want to tell me about it,

play02:06

you have to use sentences or pictures and you can only share

play02:12

a very limited amount of information that way.

play02:15

So it's much, much slower for you

play02:17

to communicate what you've learned to me

play02:19

than it is for these digital intelligence is to communicate

play02:22

stuff and that makes them much better.

play02:24

They can learn a whole lot of stuff between them.

play02:27

- [Speaker] You've said that digital intelligence

play02:29

is immortal and that biological intelligence is mortal.

play02:34

What did you mean by this?

play02:36

So if I learn some connection strengths

play02:40

in a neural net that's being

play02:41

simulated on digital computers,

play02:44

then if a particular computer dies,

play02:47

those same connection strengths

play02:48

can be used on another computer.

play02:51

And even if all the digital computers died,

play02:53

if you'd stored the connection strength somewhere,

play02:55

you could then just make another digital computer and run

play02:58

the same weights on that other digital computer.

play03:01

But with us, the knowledge that we learn,

play03:04

the connection strengths,

play03:06

are specific to our particular brains.

play03:08

Every brain is a bit different.

play03:10

The neurons in your brain are all a bit different

play03:12

and you learn so as to make use

play03:15

of all the idiosyncrasies of your particular brain.

play03:19

And so once you've learned connection strengths

play03:20

in your brain, if you told me those connection strengths,

play03:23

they wouldn't do me any good 'cause my brain's different.

play03:26

So the digital computers are immortal because you can run

play03:29

that same knowledge on a different piece of hardware.

play03:32

We are immortal because the hardware

play03:35

and the knowledge are intricately entangled.

play03:38

You can't separate the connection strengths

play03:40

from the particular brain they're running in.

play03:43

And so if the brain dies, the knowledge dies.

play03:47

- [Narrator] Why should we be concerned

play03:48

about digital intelligence taking

play03:50

over from biological intelligence?

play03:53

- Because I think it's much better at sharing what's learned

play03:56

by a whole bunch of different digital agents who all share

play04:00

the same weights and they just share the updates

play04:02

to the weights, and now they

play04:03

can learn 10,000 different things at the same time.

play04:08

But also I think the digital intelligence probably

play04:10

has a better learning algorithm than the brain's got.

play04:13

All the attempts to find a learning algorithm

play04:15

in the brain that works

play04:17

as well as the back propagation algorithm

play04:19

in these digital intelligences.

play04:22

So far, those attempts have failed.

play04:24

We haven't found anything that scales up as well to very

play04:27

large systems as the back propagation algorithm.

play04:30

So I think they've got two advantages.

play04:32

They've probably got a better learning algorithm

play04:34

and they can share knowledge

play04:36

much more efficiently than biological intelligences can.

play04:39

- [Speaker] At the time, when you entered the field,

play04:41

there were two schools of thought

play04:43

on machine intelligence, mainstream and neural net.

play04:48

Can you describe the difference

play04:49

between these two approaches?

play04:51

- I can sort of caricature it.

play04:53

So there's two different models

play04:55

of what intelligence is all about.

play04:57

And one model is that it's all about reasoning.

play05:01

And the way we reason is by using logic.

play05:03

And so that's what's special about people.

play05:06

And what we should be doing is understanding

play05:10

the kind of logic that we actually use.

play05:13

And that also went with the idea that the knowledge you

play05:17

store is symbolic expressions so that I can say a sentence

play05:22

to you and you will somehow store that

play05:25

and then later you'll be able to use it

play05:27

for inferring other sentences.

play05:29

But what's inside your head is something

play05:31

a bit like sentences but cleaned up.

play05:34

And there's a completely different model of intelligence,

play05:37

which is that it's all about learning

play05:39

the connection strengths in a network of brain cells.

play05:43

And what it's good for is things like perception

play05:45

and motor control, not for reasoning.

play05:47

That reasoning came much,

play05:48

much later and we are not very good at it.

play05:51

You don't learn to do it till you're quite old.

play05:54

And so reasoning's actually a very bad model

play05:57

of biological intelligence.

play05:58

Biological intelligence is about things like controlling

play06:00

your body and seeing things.

play06:02

And that was a totally different paradigm

play06:05

and had a different idea of what's

play06:07

inside your head that it's not stored strings of symbols,

play06:12

it's just connection strengths.

play06:15

The symbolic AI view,

play06:17

the crucial question was what is the form of these symbolic

play06:20

expressions and how do you do the reasoning with them?

play06:24

For the neural net view,

play06:26

the central question was quite different.

play06:27

It was how do you learn these connection strengths

play06:30

so you can do all these wonderful things?

play06:32

And so learning was always central to the neural net view.

play06:35

For the symbolic view,

play06:36

they said we're worried about learning later.

play06:38

First you have to figure out how the knowledge

play06:39

is represented and how we reason with it.

play06:41

And so these were totally different views.

play06:44

One took its inspiration from logic and one from biology.

play06:47

And for a long time the people in the logic camp thought

play06:51

taking inspiration from biology was silly.

play06:54

That was a bit strange since for Neumann and Turing

play06:59

had both thought neural nets were a way to attack

play07:01

intelligence, but unfortunately they both died young.

play07:07

- [Narrator] Can you, at a high level, describe

play07:09

how a neural network works?

play07:12

- I can try.

play07:14

So let's start off by describing how it would work

play07:17

for recognizing objects and images.

play07:20

And let's suppose all we wanted to do was say whether or not

play07:23

there was a bird in the image.

play07:24

And let's suppose the bird's gonna be roughly in the middle

play07:26

of the image and the main object of attention.

play07:28

And you have to say, is this a bird or isn't it?

play07:33

So you can think of an image,

play07:34

let's suppose it's a hundred pixels by a hundred pixels.

play07:37

That's 10,000 pixels.

play07:39

Each pixel is three colors, RGB.

play07:41

So that's 30,000 numbers.

play07:43

And in computational terms,

play07:45

recognizing a bird in an image consists of taking 30,000

play07:49

numbers and outputting one number

play07:51

that says yes or no, it's a bird.

play07:54

And you could try and write a standard computer program

play07:57

to do that.

play07:58

And people tried for many,

play07:59

many years and they could never get it to work very well.

play08:01

Like for 50 years they were trying to do that.

play08:04

Or you could make a multi-line neural net.

play08:08

And I'll start off by telling you how you would wire up

play08:10

a neural net by hand.

play08:12

So what you'd do is you'd have the pixels

play08:15

and that will be the bottom level.

play08:18

And then you'd have a layer of feature detectors,

play08:21

and a typical feature detector

play08:23

might have big positive connection

play08:26

strengths coming from a vertical row of pixels

play08:29

and big negative connection strengths

play08:31

coming from a neighboring vertical row of pixels

play08:34

and no connection strengths anywhere else.

play08:37

So if both rows of pixels are bright,

play08:41

it'll get big positive input from here,

play08:42

but also big negative input from there.

play08:44

So it won't do anything.

play08:46

But if these ones are bright,

play08:47

giving a big positive input and these ones are not bright

play08:50

so it doesn't get inhibited by these ones,

play08:52

it'll get all excited and it'll say,

play08:54

hey, I found the thing I like,

play08:55

which is bright pixels here and dark pixels here.

play08:58

And that's an edge detector.

play09:00

I just told you how to wire up by hand using positive

play09:03

and negative weights,

play09:04

something that would detect a little vertical edge.

play09:07

So now imagine you have a gazillion of those guys detecting

play09:10

different edges in different locations in the image,

play09:12

in different orientations and in different scales.

play09:15

That will be your first layer of feature detectors.

play09:18

Now if I was wiring it by hand,

play09:19

my second layer of feature detectors,

play09:22

I would maybe have a detector that takes two,

play09:25

that takes two edges that join at a fine angle like this.

play09:29

So it's looking for this edge and this edge.

play09:31

And if they're both active at once it would say,

play09:34

hey, maybe there's a beak here.

play09:36

It could be all sorts of other things,

play09:38

but it might just be a beak.

play09:39

So you have a feature that's sort of beak like.

play09:43

You might also in that layer

play09:45

have a feature that detects a whole bunch

play09:46

of edges that form a circle.

play09:49

And so you'd have circle detectors and potential beak

play09:53

detectors as well as lots of other detectors in that layer.

play09:56

But they're detecting slightly more complicated things.

play09:59

And then in the layer above that,

play10:01

you might have something that detects a potential beak in

play10:05

the right spatial relationship to a potential circle,

play10:09

a potential eye so that it could be the head of a bird.

play10:13

So that would be like your third layer.

play10:16

And maybe if in your third layer you also got something

play10:18

that detected the foot of a bird and the wing of a bird,

play10:22

then maybe in the next layer you could have a bird detector,

play10:25

that if several of those things got active, like okay,

play10:28

here's a head and there's a wing and there's a foot,

play10:31

it probably is a bird.

play10:33

Okay, so I told you how to wire all those things up by hand,

play10:37

but you'd never be able to do a very good job of it.

play10:40

So instead of wiring it all up by hand,

play10:44

we could imagine trying to learn it all.

play10:46

So I've told you the kind of thing we want to learn,

play10:49

but now I'll tell you how we learn it

play10:51

and the way we learn it sounds bizarre at first.

play10:54

Instead of wiring in all the connection strengths,

play10:56

so you get the detectors you want,

play10:59

you start with random connection strengths,

play11:01

just random numbers on all the connections.

play11:03

And so you put it in the image of a bird and you go forward

play11:06

through these layers of feature detectors

play11:08

and it just behaves completely randomly.

play11:11

And the bird detector at the output will say 0.5,

play11:15

it's a bird.

play11:16

It's gonna say one when it's sure it's a bird and zero

play11:18

when it's sure it's not a bird.

play11:19

To me we're gonna say about 0.5.

play11:21

And now you can ask the following question.

play11:24

How can I change all those connection strengths

play11:27

in the network?

play11:29

So instead of saying 0.5, it's a bird.

play11:31

Let's suppose it is a bird.

play11:33

It says 0.51, it's a bird.

play11:35

So the question you wanna ask is how should I change

play11:37

a particular connection strength

play11:39

so as to make it more likely that it's a bird?

play11:43

And you can figure that out by taking the difference between

play11:47

what you got and what you wanted.

play11:49

So you wanted one and you actually got 0.5.

play11:52

You take that difference and you send that difference

play11:55

backwards through the network.

play11:57

And then you use some calculus, which I won't explain.

play12:00

And you are able to compute for every single connection

play12:04

in the network how much you'd like to make it bigger

play12:08

or smaller in order to make it more likely to say bird.

play12:12

Then you adjust all the connection strengths very slightly

play12:14

in the direction that I'm making it more likely to say bird.

play12:17

Then you show it something that isn't a bird

play12:19

and now you're gonna adjust connection strengths.

play12:21

So it's less likely to say that that was a bird

play12:24

and you just keep going like that with lots of birds

play12:27

and non birds.

play12:28

And eventually you'll discover that it's discovered.

play12:32

all these feature detects,

play12:33

it'll have discovered beak-like things and eye-like things

play12:36

and things that detect feet and wings and all that stuff.

play12:39

And if you train it on lots of different objects,

play12:42

like a thousand different categories of object,

play12:45

it'll discover intermediate feature detectors

play12:48

that are very good for recognizing all sorts of things.

play12:51

So the magic is that there's this relatively simple

play12:53

algorithm called back propagation that takes the error

play12:57

in the output and sends that error backwards through

play12:59

the network and computes

play13:01

for all the connections how you should

play13:03

change 'em to improve the behavior.

play13:05

And then you change it more a tiny bit

play13:07

and you just keep going with another example.

play13:11

And surprisingly, that actually works.

play13:14

For many years people thought that

play13:15

would just get jammed up.

play13:17

It will get stuck somewhere.

play13:18

But no it doesn't, it actually works very well.

play13:22

- [Narrator] I'm curious,

play13:23

how do neural networks handle language?

play13:27

- Okay.

play13:28

So now you've got the idea

play13:29

of how we train it to recognize a bird.

play13:33

Imagine now that we take

play13:38

a string of words as the input.

play13:40

And the first thing you're gonna do is convert a word

play13:44

into an embedding vector that is,

play13:47

it's a little bunch of numbers that captures the meaning

play13:49

of the word or is intended to capture

play13:51

the meaning of the word.

play13:53

And so your first layer after the words will be

play13:56

these embedding vectors for each word.

play13:59

And now we're gonna have lots of layers

play14:01

of embedding vectors.

play14:03

And as we go up through the network,

play14:05

we're gonna make the embedding vectors

play14:07

for a word get better and better,

play14:10

'cause they're gonna take into account

play14:11

more and more contextual information.

play14:13

So suppose in this sentence,

play14:16

let's suppose we don't have any capital letters okay.

play14:18

So suppose in this sentence you have the word may.

play14:23

Well, the most probable meaning of may

play14:25

is that it's a modal as in he may do that,

play14:29

but obviously there's a completely different meaning of May,

play14:31

which is the month.

play14:33

And so initially it doesn't know,

play14:36

just looking at the word may,

play14:37

it doesn't know what embedding vector to use

play14:40

and it'll use a kind of compromise vector,

play14:42

something that's sort of halfway between the embedding

play14:45

vector that represents the modal may,

play14:48

and the embedding vector that represents the month May.

play14:52

And then at the next layer it's gonna refine that vector.

play14:57

It's gonna make a slightly better vector depending

play14:59

on the context that it got,

play15:00

depending on nearby embedding vectors.

play15:02

So if for example nearby

play15:05

there's the embedding vector for June,

play15:09

then it'll refine the one for May

play15:10

to be more like a month and less like a modal.

play15:14

But if there's the embedding vector for wood,

play15:16

it'll make it more like a modal and less like a month.

play15:21

And as you go through the network,

play15:22

it can refine these embedding vectors

play15:25

and make them better and better.

play15:28

And the way we're gonna train it,

play15:31

is we're gonna give it a string of words as input.

play15:36

And we are gonna, here will be one way to do it.

play15:39

It's not exactly what's done but it's easy to understand.

play15:42

For the last word you just put in a kind of neutral word,

play15:46

you say unknown and it has a very vague embedding vector

play15:50

that's kind of the average of all the vectors for all words.

play15:52

It doesn't know, right.

play15:55

Now as you go forward through the network,

play15:58

that last word will be able

play16:00

to be influenced by previous words.

play16:03

And it starts off very vague,

play16:05

but as you go through these layers

play16:07

it can get more and more precise.

play16:09

And by the time you get to the end of the network,

play16:12

that embedding vector could look like

play16:16

the embedding vector for a particular word

play16:18

or for some combination of words,

play16:21

some average of several words.

play16:24

And you train the network by saying

play16:30

you go through all these layers

play16:32

and that last word you'd like the embedding vector

play16:35

to look like the embedding vector for the word that actually

play16:38

was there in the text.

play16:40

And that's how it predicts the next word.

play16:42

It tries to change this sort of neutral embedding vector

play16:45

into one that is close to the embedding vector

play16:50

for the correct word that appeared in the text.

play16:53

And you take the error,

play16:56

the difference between the embedding vector in the text

play16:59

and the embedding vector produced and you propagate

play17:01

that backwards through the network

play17:03

and it's propagating backwards through the layers,

play17:06

but it's propagating from this word to previous words,

play17:10

so that they will have the right influence on this word.

play17:13

And that's the back propagation algorithm learning

play17:16

to predict the next word.

play17:18

- [Speaker] So despite some of the theoretical breakthroughs

play17:21

in this field, these neural networks didn't work very well

play17:25

for a long time.

play17:26

And why was that?

play17:28

- It was a combination of reasons.

play17:30

So we weren't very good at initializing them, but as I said,

play17:34

you put in random weights and then learn everything.

play17:37

But if you don't carefully decide

play17:40

what kind of random weights,

play17:42

the thing never gets off the ground.

play17:44

So that was a little technical reason why they didn't work

play17:46

very well in deep nets with lots of laser feature detectors.

play17:49

But the main reason was we didn't have enough compute power

play17:52

and we didn't have enough data.

play17:54

So people were trying to train these nets on relatively

play17:57

small training sets without much compute power.

play18:00

And in that regime, other methods work better.

play18:03

Neural nets really come into their own when you have lot

play18:06

of data and a lot of compute power.

play18:08

And then you can use a big neural net

play18:09

and then it works much better than anything else.

play18:12

And we didn't realize that at the time.

play18:14

So we would occasionally fantasize,

play18:15

well suppose you had a lot more data

play18:17

and a lot bigger than computer, it will work better.

play18:19

But we didn't realize it will work a whole lot better.

play18:22

And so in the 1990s it was a relatively dead period

play18:27

for neural nets 'cause other methods

play18:29

were working better on small problems.

play18:32

And a lot of people in computer science

play18:35

gave up on neural nets.

play18:37

In psychology, they didn't 'cause in psychology,

play18:40

they wanted something that was like the brain,

play18:42

and neural nets were clearly more like the brain

play18:43

than symbolic AI.

play18:44

But in computer science, neural nets sort of came

play18:47

into distribute in the 90s.

play18:49

- [Speaker] So let's fast forward

play18:50

then to another decade to the 2000s.

play18:54

Was there a moment for you

play18:56

when it became clear that the approach that you'd

play18:59

been pursuing was the one that was gonna prevail?

play19:03

- Okay.

play19:05

In 2006 we figured out how to initialize the weights much

play19:09

better by doing unsupervised learning

play19:12

and then back propagation work much better.

play19:14

So it was fairly clear then the back propagation

play19:16

really was gonna work very well.

play19:19

But in 2009, two of my grad students,

play19:23

George Dahl and Abdurahman Muhammad made a

play19:27

much better speech recognizer,

play19:30

actually a slightly better speech recognizer,

play19:31

but it was slightly better than the state of the art,

play19:34

using deep neural nets.

play19:36

And then it was fairly clear that this stuff was going

play19:39

somewhere and all the big speech groups over the next few

play19:42

years switched to using neural nets.

play19:46

And then in 2012 that speech stuff came out in the Android

play19:50

and suddenly the Android caught up with Siri.

play19:52

It was as good as speech as Siri

play19:53

'cause it was using neural nets.

play19:56

And in the same year, two others of my graduate students,

play20:00

Ilya Sutskever and Alex Krizhevsky,

play20:04

made a neural net that was very good

play20:06

at recognizing objects and images

play20:07

and that beat the state of the art by a lot.

play20:10

And so I think it was this combination that it was already

play20:13

working for speech recognition and already in production.

play20:17

The big companies knew that,

play20:18

the public I don't think were very well aware of that.

play20:21

But then suddenly it worked much better for computer vision

play20:25

and that was a turning point.

play20:27

In 2012 when we won the ImageNet competition

play20:30

by a huge margin,

play20:32

we got almost half the errors of the other methods

play20:35

and it was a public data set,

play20:38

but with a hidden test set so you couldn't cheat.

play20:41

- [Speaker] So let's just focus a bit on 2012

play20:44

because you said it was

play20:44

a really pivotal year for for this.

play20:48

Can you you describe, again at a high level,

play20:51

how AlexNet worked?

play20:54

I take it that might have been named

play20:55

after your graduate student that.

play20:56

- That was named after Alex Krizhevsky.

play20:58

'Cause he did,

play21:00

he was a wizard programmer and he made it work.

play21:04

- Ilya helped a lot but it was mainly Alex's work.

play21:08

So I explained to you when explaining backdrop

play21:11

how you'd have these layers of feature detectors

play21:14

and AlexNet was basically that kind of a net,

play21:18

but with a thousand different object classes

play21:21

and with about seven layers

play21:23

of feature detectors.

play21:26

And it also used something else

play21:28

that was developed by Yanaka,

play21:30

which is convolutional nets.

play21:33

And I'll try and explain those now

play21:34

'cause they were very important.

play21:38

Remember how I said you might make a detector

play21:41

for a bird's beak by checking two lines,

play21:44

by having two lines like that.

play21:46

And if you see those two feature detectors,

play21:49

then you make a beak detector.

play21:50

But that would just be for a specific location, right.

play21:54

In a convolutional net,

play21:56

when you make a feature detector for one location,

play21:59

you make the same feature detector

play22:00

for all the locations in the image.

play22:04

So now if it is trained with a beak here when it's learning

play22:09

and it really says I need a beak detector for that.

play22:11

So it learns a feature that detects this beak,

play22:14

it will automatically make copies

play22:16

for all of the other locations in the image.

play22:18

So if now the bird occurs in a different location,

play22:22

it will have the feature detectors to recognize it.

play22:25

So that idea that you copy the feature detectors

play22:27

to every location, that's a convolutional net essentially.

play22:32

And that makes the whole thing generalized

play22:35

much better across position.

play22:36

It can cope now with things changing position

play22:39

because it's got copies of all these feature detectors

play22:41

in every location.

play22:43

And with convolutional nets and multiple layers of features,

play22:50

what Alex did was programmed all that very efficiently

play22:54

on a thing called a graphics processing unit,

play22:56

which was developed for computer graphics,

play22:59

but is like a mini supercomputer.

play23:01

It can do lots and lots of computation

play23:04

in lots of separate processes all at the same time.

play23:07

And so it gave us about a factor of 30

play23:09

compared with a normal computer.

play23:11

And a factor of 30 is about sort of 10 years

play23:13

progressing in computers.

play23:15

So suddenly we could leap 10 years into the future in terms

play23:18

of compute power.

play23:20

And it was very difficult

play23:22

to program these GPU boards.

play23:26

Alex managed to program two of them to collaborate,

play23:28

which was even more difficult.

play23:31

And the last ingredient was the ImageNet data set.

play23:35

So someone called Fei-Fei Li and her collaborators put

play23:38

together a big set of images and then a public competition

play23:43

where you had about a million images with a thousand

play23:46

different kinds of objects.

play23:47

So you had about a thousand examples of each kind of object

play23:50

and you had to learn to recognize those objects.

play23:53

And then the test set would be different images,

play23:56

which also contained those objects.

play23:57

And so you'd have to generalize to the different images.

play24:00

And it turned out the best computer vision technique that

play24:03

had been invented up 'til then was getting like 25% errors

play24:07

and Alex got 15% errors.

play24:11

And since then it's gone down to about 3% errors.

play24:14

It's gone much better since then.

play24:15

But it was a huge jump and people in computer vision were

play24:19

extremely surprised and most of them behaved in a very

play24:22

admirable way, which is they said,

play24:24

"Hey, we never thought this would work,

play24:27

but hey it works so we're gonna do that

play24:28

instead of what we were doing."

play24:30

That's what scientists don't usually do.

play24:32

Scientists usually just grow old complaining

play24:34

that this new stuff is nonsense.

play24:35

- [Speaker] And how would you describe

play24:37

the pace of innovation

play24:38

that we've seen in AI since that moment?

play24:41

It's just got faster and faster.

play24:43

So if you'd asked me in that moment how long 'til

play24:46

these neural nets can do machine translation,

play24:49

that's better than the state of the art,

play24:52

I'd have said maybe 10 years.

play24:54

'Cause machine translation is the kind of thing that if

play24:58

you've got a theory that's all about processing

play25:00

strings of symbols, machine translation

play25:02

is the ideal problem for you.

play25:04

'Cause you have a string of symbols in one language and you

play25:06

have to produce a string of symbols in another language.

play25:09

And the symbolic people thought well inside

play25:11

you're just manipulating strings to do that.

play25:14

The neural net people thought

play25:16

you have to take this string of symbols,

play25:17

you have to convert it into

play25:18

these big pans of neural activity,

play25:20

and then you have to convert it back into symbols

play25:22

at the output.

play25:25

And I was very surprised when it only took a few years

play25:28

for machine translation to be good.

play25:30

And then in another year or two Google was using it

play25:33

and it greatly improved the quality

play25:34

of machine translation.

play25:36

Like in languages like Chinese, this is from memory,

play25:39

but there was a gap between how good

play25:41

the computer translation was

play25:43

and how good human translation was.

play25:46

And it just halved that gap overnight.

play25:49

I think it was Chinese that did that,

play25:51

but in a lot of languages it just made a lot better.

play25:52

And since then it's obviously

play25:54

it's got considerably better since then.

play25:56

But by 2015 it was already working pretty well

play26:00

and that really surprised me that it only took three years.

play26:04

- [Narrator] You say you were surprised

play26:05

at the pace of innovation.

play26:07

What did you think the first time you used

play26:09

a large language model like Chat-GPT, did we surprise you?

play26:15

- I am just shocked at how good it is.

play26:20

So it gives very coherent answers and it can do little bits

play26:25

of reasoning, not very sophisticated reasoning yet,

play26:28

although it'll get much better.

play26:30

So for example, I asked it,

play26:32

this is GPT 4 now.

play26:34

I asked it a puzzle given to me by a symbolic AI guy

play26:39

who thought it wouldn't be able to do it.

play26:41

I actually made the puzzle much harder

play26:43

and it could still do it.

play26:44

And so the puzzle goes like this,

play26:46

"The rooms in my house are either white or blue or yellow.

play26:53

Yellow paint fades to white within a year.

play26:57

In two years time I would like all the rooms to be white,

play27:00

what should I do?"

play27:04

And a human being would probably say,

play27:06

you should paint the blue rooms white.

play27:09

What GPT4 said was "You should paint

play27:11

the blue rooms yellow,"

play27:13

but that works too 'cause the yellow will fade to white.

play27:17

And I don't see how it could do

play27:19

that without understanding the problem.

play27:21

The idea that it's just sort of predicting

play27:23

the next word and using statistics.

play27:26

There's a sense in which that's true,

play27:29

but it's not the sense of statistics

play27:30

that most people understand.

play27:34

From the data,

play27:36

it figures out how to extract the meaning of the sentence

play27:39

and it uses the meaning of the sentence

play27:41

to predict the next word.

play27:42

It really does understand and that's quite shocking.

play27:46

- [Speaker] So have you been surprised

play27:48

by the broader reaction, the public reaction to Chat-GPT?

play27:53

- Well given how well it works,

play27:54

I guess the public reaction isn't that surprising.

play27:57

But what's interesting is most people

play28:00

don't say this doesn't understand.

play28:03

They say wow, it understood what I said

play28:04

and gave me a coherent answer.

play28:06

What can I use it for?

play28:09

And I think most people are right about that.

play28:11

And of course it can be used for huge numbers of things.

play28:15

So I know someone who answers letters of complaint

play28:19

for the health service,

play28:21

and he used to spend 25 minutes composing

play28:24

a letter that addresses the problem and so on.

play28:26

Now he just types the problem to GPT-4

play28:32

and it writes the letter,

play28:34

and then he just looks at the letter and decides

play28:36

if it's okay and sends it out.

play28:37

And that takes him five minutes now.

play28:39

So he is now five time

play28:40

more efficient and that's gonna happen all over the place.

play28:44

Like paralegals are gonna be like that.

play28:46

Programmers are already getting like that.

play28:49

Programmers can be much more efficient if they get

play28:52

assistance from things like GPT-4

play28:54

'cause it knows how to program.

play28:56

And you might think it just knows how to program

play28:59

'cause it's seen a whole lot of programs.

play29:03

So I have a former graduate student who's very smart

play29:05

and a very good programmer,

play29:08

and he did a little experiment which is,

play29:11

he's called Redford Neil.

play29:12

He took

play29:15

GPT-4,

play29:17

and he defined a new programming language

play29:20

with very unusual syntax.

play29:23

And having defined this programming language

play29:25

just in text to GPT-4,

play29:28

he then gave it a program and said "what would this do?"

play29:32

And it answered correctly.

play29:34

So basically he could understand the definition

play29:36

of a new programming language

play29:37

and figure out what programs in that language would do.

play29:40

And again the idea that it's just predicting

play29:43

the next word doesn't make any sense in that context.

play29:46

It had to understand what was going on.

play29:49

- [Speaker] So what do you see

play29:49

as some of the most promising opportunities for this

play29:53

type of AI when it comes to benefiting society?

play29:58

- It's hard to pick one 'cause there's so many,

play30:00

like there'll be a huge increase in productivity

play30:05

for any job that involves outputting text.

play30:09

There's all sorts of issues

play30:10

about increasing productivity.

play30:12

In our society,

play30:13

it's not necessarily a good thing to increase productivity

play30:15

'cause it might make the rich rich and the poor poorer.

play30:17

But in a decent society,

play30:19

just increasing productivity ought to be a good thing.

play30:22

So there'll be things like that.

play30:24

It's wonderful for making predictions.

play30:26

It'll be better at predicting the weather.

play30:28

It'll, people don't know by how much yet.

play30:32

But it's already much better at predicting floods.

play30:35

It can predict earthquakes.

play30:37

It can design new nano materials.

play30:41

So for things like solar panels,

play30:42

you want to be able to design new nano materials

play30:44

or for superconductivity.

play30:46

I don't know if it's used for superconductivity yet,

play30:48

but it may well be.

play30:49

You'd like that at high temperature.

play30:52

It's really good at designing drugs that is finding

play30:58

molecules that'll bind to some particular other molecule.

play31:02

Deep mind has used it to create alpha fold.

play31:07

Now that's not a chat-bot, that's just deep learning.

play31:10

But the basic technology of deep learning has pretty much

play31:17

solved the problem of how you figure out from the string

play31:22

of bases in a protein, what shape it will adopt.

play31:25

And if you know what shape it adopts, you know its function.

play31:27

The chat-bots are just gonna be used everywhere I think.

play31:31

- [Speaker] And we've also talked a lot about healthcare.

play31:33

I mean you talked about drug discovery,

play31:35

but healthcare is another field that could really benefit.

play31:38

- Yes.

play31:39

Both in interpreting medical scans.

play31:42

Like if you take a CAT scan,

play31:44

there's a lot of information in the CAT scan

play31:47

and that isn't being used,

play31:49

and most doctors don't know what the information is.

play31:52

This will be able to get much more out of a CAT scan

play31:55

as well as being able to compete

play31:56

with doctors at saying what kind of cancer you have

play31:59

or how big it's grown.

play32:01

At present, for example,

play32:02

when a doctor tells you the size of a cancer,

play32:06

you'll get a number like it's three centimeters

play32:09

and a month ago it was two centimeters.

play32:11

Now that's not a very useful number,

play32:13

if the thing looks like an octopus, right.

play32:17

A neural net will be able to do much better at understanding

play32:20

the volume of the cancer and how it's changed.

play32:23

So it's gonna be tremendous there.

play32:26

And it already,

play32:26

it's at the level of humans for lots of kinds of scans

play32:30

and it's gonna get better.

play32:32

It's gonna be very good for diagnosing diseases.

play32:35

So at present there's a large number

play32:39

of people dying in North America,

play32:41

'cause the doctors misdiagnosed what they had.

play32:44

There's a system that Google's producing called Med-PaLM 2,

play32:48

which has learned to do diagnoses and it's already,

play32:53

I think it's better than an average doctor now.

play32:56

I'm not quite sure about this

play32:57

'cause I'm not at Google anymore and it's very recent.

play33:00

But it's certainly comparable with doctors

play33:01

and it's gonna get better fast.

play33:04

So wouldn't you like to have

play33:07

a sort of general practitioner family doctor.

play33:10

You go with some rare disease and you'd love your family

play33:13

doctor to have already seen hundreds of cases

play33:16

of that rare disease.

play33:17

A Med-PaLM 2's gonna be like that.

play33:19

So it's gonna be just in the end much better at diagnosis.

play33:25

- [Narrator] It sounds like AI will bring

play33:27

many important benefits,

play33:29

but you have expressed concern

play33:30

about the current pace of innovation.

play33:33

Why?

play33:34

- Okay, so for like 50 years I thought that,

play33:37

well for 49 years.

play33:39

In order to make digital models better,

play33:42

we needed to make them work more like the brain.

play33:45

So I kept looking at things the brain does

play33:47

and the digital models don't,

play33:48

like rapidly changing connection strengths

play33:50

in a temporary way and that can make

play33:53

the digital models better.

play33:57

And very recently I realized that because

play34:01

these digital models have this kind of hive mind where

play34:03

when one agent learns something,

play34:05

all the other agents know it,

play34:07

they might actually already be better

play34:09

than biological intelligence.

play34:11

And so I kind of completely flipped my opinion from the idea

play34:14

it's gonna be a long time before they can do

play34:17

everything the brain does.

play34:19

It's gonna be 30 to 50 years before they're better than us,

play34:21

which is what I thought for until very recently.

play34:24

A few months ago I suddenly realized

play34:27

maybe they're already better than us,

play34:29

they're just smaller and when they get bigger,

play34:33

then they'll be smarter than us.

play34:35

And that was quite scary.

play34:36

It was a sudden change of opinion that instead

play34:39

of being 30 to 50 years, it was five years to 20 years,

play34:42

something like that.

play34:43

And so we needed now to take really seriously right now what

play34:48

we are gonna do about the issue

play34:49

these things may become smarter than us.

play34:52

It's a time of huge uncertainty.

play34:53

Nobody really knows what's gonna happen.

play34:55

Maybe things will stall and maybe they won't become smarter

play34:58

than us, but I don't really believe that.

play35:01

I think they're gonna be smarter than us,

play35:03

but maybe when they become smarter than us,

play35:04

we'll be able to keep them benevolent

play35:07

and we'll be able to keep them caring much more about people

play35:10

than they care about themselves, unlike people.

play35:13

But maybe not.

play35:14

And so we need to start thinking very hard about those

play35:18

issues and I'm not an expert on those issues.

play35:21

I'm just an expert on these learning algorithms.

play35:24

And I suddenly realized these super intelligences may be

play35:28

here quite soon and I'm just sounding the alarm so that

play35:32

people listen to the experts who've been thinking for

play35:35

a long time about how we might stop them taking control.

play35:40

I want the politicians to listen to those guys,

play35:42

rather than say, yeah, yeah,

play35:44

they're sort of sci-fi guys, that it's never gonna happen.

play35:48

- [Speaker] Was there like a particular moment

play35:50

when you had this,

play35:51

you said it was very recent,

play35:52

that where you kind of changed your view on it?

play35:54

- I was developing learning algorithms

play35:57

for biological systems

play35:59

that could run in a biological system

play36:03

which didn't use back propagation.

play36:05

And I couldn't make them work as well as a back propagation

play36:08

algorithm that we were running on these digital systems.

play36:12

And they would work for small networks.

play36:14

But when I scaled it up,

play36:15

the digital ones always scaled up much better

play36:17

than the biological ones.

play36:19

And suddenly I thought it might not be my fault.

play36:22

It might not be that my learning algorithm

play36:25

was just a bad learning algorithm.

play36:27

It might be that these digital systems just are better.

play36:31

And that's when I suddenly changed my mind about how long

play36:35

before we get super intelligence.

play36:37

And then I talked to various former students of mine

play36:39

and former colleagues of mine,

play36:41

and some of them encouraged me to go public with this.

play36:44

Not because I had any solutions that I wanted to recommend.

play36:48

It's not like you can say burn less carbon

play36:51

and everything will be fine.

play36:54

But because they thought I'm well known in the field

play36:58

and if I go public by saying super intelligence

play37:00

might be here quite soon,

play37:01

the politicians might start to believe that's a possibility

play37:04

and start listening seriously

play37:07

to the researchers who've been

play37:09

thinking a long time about how we prevent

play37:11

these things from gaining control.

play37:13

- [Speaker] So from your point of view,

play37:15

what role can governments play in helping ensure

play37:20

these AI's are developed in a responsible way?

play37:23

- So there's all sorts of risks other people

play37:26

have talked about a lot

play37:27

and I don't particularly want to talk about,

play37:28

like they'll take jobs away and increase the gap

play37:32

between the rich and the poor.

play37:34

They will make it impossible to know

play37:36

whether news is fake or real.

play37:38

They will encourage society to divide into two warring camps

play37:42

that don't listen to each other

play37:44

and have completely opposing views.

play37:46

They will build battle robots

play37:48

that are designed to kill people.

play37:49

All of those are well-known risks

play37:51

that I'm not talking about.

play37:52

It's not that I don't think they're important,

play37:54

I think they're probably even more urgent.

play37:56

But lots of other people are talking about those risks.

play37:59

The risk I'm talking about is the risk

play38:00

these things will get smarter than us

play38:02

and eventually take over.

play38:04

And for that risk there may be something governments can do

play38:08

because nobody wants that.

play38:12

Well if you exclude these super intelligences,

play38:14

no people want that.

play38:16

And so all the different governments

play38:19

ought to be able to agree,

play38:24

they ought to be able to work together on preventing that

play38:25

'cause it's in their interests.

play38:28

And that's happened before.

play38:29

Even during the Cold War,

play38:31

the US and Russia could work together on trying to prevent

play38:34

them being a global nuclear war

play38:36

'cause it was so bad for everybody.

play38:38

And for this existential threat,

play38:40

it should be possible for everybody to work together

play38:43

to limit it if it's possible to prevent it.

play38:46

I don't know whether it's possible to prevent it,

play38:48

but at least we should be able to get international

play38:50

collaboration on that particular threat,

play38:53

the existential threat of AI taking over.

play38:55

One thing I think should be done is wherever this stuff's

play39:00

being developed, particularly these big chat-bots,

play39:04

governments should encourage the companies

play39:07

to put a lot of resources,

play39:09

as these things are getting more and more intelligent,

play39:12

to doing experiments to figure out how to keep

play39:14

them under control.

play39:16

So they should be sort of looking at how these things might

play39:18

try and escape and doing empirical work on that

play39:21

and put a lot of resources into that

play39:24

'cause that's the only chance we've got.

play39:26

Before they're super intelligent,

play39:29

we can maybe do experiments and see what's gonna go wrong.

play39:33

And I'm strongly of the belief

play39:35

you need empirical data on this.

play39:37

You just can't have philosophers and politicians

play39:39

and legislators making up rules.

play39:42

You need empirical work looking at these things and seeing

play39:45

how they go wrong and seeing how you might control them.

play39:48

And that can only be done by the people developing them.

play39:51

So since you can't stop the development,

play39:54

the best you can do is somehow have governments put a lot

play39:58

of pressure on these companies

play39:59

to put a lot of resources into investigating empirically

play40:05

how to keep 'em under control when

play40:06

they're not quite as smart as us.

play40:09

- [Speaker] And and what do you see as the role

play40:11

of these big technology companies

play40:13

where a lot of this development is happening?

play40:15

Would they do this without

play40:16

that kind of government regulation?

play40:19

- So a lot of the people in the big companies,

play40:22

all the people I know who are senior in the big companies

play40:24

are very worried about this and do put work into that.

play40:28

They're very concerned about it,

play40:30

but they have an obligation to their shareholders.

play40:33

And I think it to make big profits,

play40:36

and making big profits, particularly in the short term,

play40:40

doesn't align nicely with putting a lot of effort

play40:43

into making sure it's safe.

play40:45

So you see this in all industries.

play40:47

In the railway industry in the states,

play40:50

having safety devices that tell you

play40:52

when a wheel's locked cost money,

play40:55

and the big rail companies just rather

play40:57

have accidents than do that.

play41:01

Google, which is a big company I know something about

play41:04

is not quite like that because it understands

play41:06

that it's got a tremendous reputational loss

play41:09

if bad things happen.

play41:11

And that's why Google didn't release these chat-bots.

play41:13

It kept them private.

play41:14

It didn't want them out there in the world

play41:15

for people to play with.

play41:17

It wanted to use them to give you better search results

play41:21

or to complete your Gmail for you,

play41:24

but not to give them to people to play with.

play41:27

And it could only be responsible like that until OpenAI

play41:30

and Microsoft put them out there

play41:31

and then Google had to compete.

play41:34

But the big people in the big companies really care a lot

play41:36

about their reputation and about not having bad effects,

play41:40

but they could maybe be made to care even more about

play41:43

the safety issue by government

play41:45

doing something to insist that

play41:48

they put a lot of work into that.

play41:50

And there's other things that could happen,

play41:51

like it's very hard within a company to have people working

play41:57

on long-term existential threats because they're paid

play42:02

by the company and there's a conflict of interest,

play42:05

which is one of the reasons I left Google.

play42:07

Not 'cause Google did anything wrong,

play42:08

'cause I just don't want any conflict of interest.

play42:13

One thing the big companies could certainly do,

play42:15

is put more money into funding foundations

play42:18

that study these things.

play42:20

And Google for example, put $300 million

play42:22

into a foundation called Anthropic

play42:26

that is studying these things.

play42:29

They could put a lot more money in.

play42:32

- [Speaker] I'm curious about what advice

play42:35

you would give or what guidance you would give

play42:37

to other researchers in the field

play42:39

or might be just entering the field right now and wanna make

play42:43

sure that they're advancing the field,

play42:44

but doing it in a responsible way.

play42:48

- Well, one piece of advice I'd give

play42:50

is look at how many people are

play42:52

working on making these things better and how many people

play42:55

are working on preventing them from getting out of control.

play42:58

And you'll see it's like 99 people are working on making

play43:00

them better and one person's working

play43:01

on preventing them getting out of control.

play43:03

So where could you make the most impact probably on working,

play43:07

preventing them getting out of control.

play43:09

So that's one piece of advice.

play43:11

The other piece of advice is my general advice

play43:13

for young researchers, which is,

play43:17

look for somewhere where you think everybody's

play43:19

doing it wrong.

play43:22

And trust your intuition until you figure out

play43:25

why your intuition is incorrect.

play43:26

Trust it and work on alternatives ways

play43:31

of doing things when you think everybody

play43:32

else is doing it wrong.

play43:34

And the fact is,

play43:35

either you have good intuitions or you don't.

play43:37

If you've got good intuitions,

play43:38

you should listen to them and follow your intuition

play43:41

and work on that until you discover why it's wrong.

play43:45

If you've got bad intuitions,

play43:46

it doesn't really matter what you do.

play43:47

So you might as well follow your intuitions.

play43:51

- [Narrator] The risks you've described are alarming,

play43:53

but can't you just throw a switch and shut it down?

play43:56

Aren't humans ultimately still in control?

play44:00

- It's very tempting to think we could just turn it off.

play44:05

Imagine these things are a lot smarter than us and remember

play44:09

they'll have read everything Machiavelli ever wrote.

play44:12

They'll have read every example

play44:15

in the literature of human deception.

play44:17

They'll be real experts at doing human deception,

play44:20

'cause they'll learn that from us,

play44:22

and they'll be much better than us.

play44:24

They'll be like, you manipulating a toddler.

play44:27

You say to your toddler, "Do you want peas or cauliflower?"

play44:30

And your toddler doesn't realize,

play44:32

actually he doesn't have to have either.

play44:35

He just thinks which he likes,

play44:37

which he dislikes the most

play44:38

and says he'll have the other one.

play44:39

So if they can manipulate people,

play44:43

they can manipulate people into pressing buttons

play44:45

and pulling levers.

play44:47

So we have a nice example of Donald Trump.

play44:50

Donald Trump can manipulate people and so he could invade

play44:53

a building in Washington without ever going there himself.

play44:57

And you didn't have to prevent Donald Trump

play44:59

from doing anything physical.

play45:01

You'd had to prevent him from talking to prevent that.

play45:05

And these are chat-bots.

play45:06

So the idea that just with talk,

play45:09

they can't do any real damage 'cause it requires people

play45:11

to do the damage.

play45:12

Well, as soon as you can manipulate people,

play45:16

then you can get whatever you like done.

play45:20

- [Narrator] You've spent your career trying

play45:21

to understand how the human brain works

play45:23

and played a critical role in AI development.

play45:26

What's next for you, Geoffrey Hinton?

play45:30

- Okay, so I'm 75 and I've reached the point

play45:32

where I'm not very good

play45:33

at writing programs anymore,

play45:34

'cause I keep forgetting the names of the variables

play45:36

I'm using and things like that.

play45:38

And I forget to,

play45:39

I do a copy and paste and forget to modify

play45:41

the thing I pasted.

play45:45

And so I've slowed down a lot in programming

play45:46

and it's very irritating.

play45:48

It's extremely irritating not to be as good

play45:50

as you used to be.

play45:51

And I decided a long time ago

play45:54

that when I reached that point,

play45:57

I would become a philosopher.

play45:59

And so I'm gonna become a philosopher.

play46:03

(gentle music)

Rate This

5.0 / 5 (0 votes)

Etiquetas Relacionadas
人工智能机器学习神经网络深度学习技术伦理智能风险未来趋势Hinton访谈科技哲学创新挑战
¿Necesitas un resumen en inglés?