Yann LeCun: Deep Learning, ConvNets, and Self-Supervised Learning | Lex Fridman Podcast #36

Lex Fridman Podcast
31 Aug 201975:58

Summary

TLDR在这次深度对话中,人工智能领域的先驱Yann LeCun分享了他对深度学习革命的看法,探讨了其在AI发展中的重要性。LeCun教授讨论了价值不一致性问题,以及如何通过设计目标函数来避免机器做出有害的决策。他还提到了自主智能系统的构建,强调了预测模型的必要性,并探讨了情感在智能中的作用。LeCun对当前AI技术的局限性和未来的可能性提供了深刻的见解,展望了人工智能的未来发展。

Takeaways

  • 🤖 深度学习是AI领域的一项革命性技术,它使机器能够从数据中学习。
  • 🧠 深度学习之父之一的Yann LeCun,以其在卷积神经网络方面的贡献而闻名,特别是在光学字符识别和MNIST数据集上的应用。
  • 🎓 Yann LeCun是纽约大学的教授,同时也是Facebook的副总裁和首席AI科学家,他因深度学习的工作获得了图灵奖。
  • 🗣️ LeCun认为,机器的目标是学习世界的模型,以便能够预测未来并做出决策。
  • 🚀 尽管深度学习在图像识别和自然语言处理等领域取得了巨大成功,但要达到人类水平的智能,还有很长的路要走。
  • 🧠 对于AI系统来说,理解和处理不确定性是一个关键挑战,这需要对世界有更深层次的理解。
  • 🤖 AI系统的设计需要考虑到伦理和道德问题,例如在设计目标函数时需要考虑到不伤害人类等原则。
  • 📚 LeCun认为,AI的发展不仅仅是技术的进步,它还涉及到法律、伦理和社会等多个领域。
  • 🌐 深度学习的成功部分归功于互联网和大数据的可用性,这些数据为训练复杂的神经网络提供了基础。
  • 🔍 LeCun强调,尽管深度学习取得了显著进展,但在理解大脑如何工作方面,我们仍然有很多不知道的地方。
  • 🚧 目前,AI系统在特定任务上表现出色,但它们并不是真正的通用智能,而是高度专业化的系统。

Q & A

  • Yann LeCun在人工智能领域有哪些重要贡献?

    -Yann LeCun是深度学习领域的先驱之一,尤其以卷积神经网络的创立者而闻名,特别是在光学字符识别和著名的MNIST数据集上的应用。他还因深度学习的工作获得了图灵奖,并在Facebook担任首席AI科学家。

  • Yann LeCun如何看待人工智能的伦理和价值对齐问题?

    -Yann LeCun认为,人工智能系统的价值对齐问题类似于人类社会中通过法律防止不良行为的情况。他强调,我们需要为AI系统设计目标函数,并在其中加入类似法律的约束,以确保AI系统的行为符合社会的共同利益。

  • Yann LeCun对于构建类似HAL 9000这样的AI系统有哪些建议?

    -Yann LeCun认为,构建类似HAL 9000的AI系统时,不应要求AI系统保守秘密或说谎,因为这最终会导致内部冲突。他还提到,应该为AI系统设定类似医生誓言的基本规则,以确保它们不会违反某些伦理原则。

  • Yann LeCun如何看待深度学习中的大规模神经网络?

    -Yann LeCun认为,大规模神经网络在相对较小的数据集上进行训练时,其效果出人意料地好,这打破了以往教科书中关于需要较少参数和数据样本的传统观点。他认为这是一个令人惊讶的发现,因为它证明了在参数数量巨大、目标函数非凸的情况下,神经网络仍然能够学习。

  • Yann LeCun对于AI系统进行推理的能力有何看法?

    -Yann LeCun相信神经网络能够进行推理,并且认为推理是学习的一个结果。他提出,为了使神经网络能够进行推理,我们需要一种工作记忆系统,以及能够访问这些记忆并进行迭代处理的网络。他还提到,目前正在研究如何使Transformer这样的模型具有类似的记忆和推理能力。

  • Yann LeCun如何看待AI系统中的因果推理?

    -Yann LeCun认为,当前的神经网络在因果推理方面存在不足,但已经有研究者在努力解决这个问题。他提到了最近的研究工作,这些工作致力于让神经网络关注真实的因果关系,这可能同时解决数据偏见等问题。

  • Yann LeCun对于深度学习在90年代的低迷和复兴有何看法?

    -Yann LeCun认为,90年代深度学习之所以低迷,部分原因是当时缺乏易于使用的软件平台和足够的数据集。他提到,当时使用神经网络需要编写大量代码,并且没有像Python或MATLAB这样的工具。他还提到,当时AT&T的法律限制阻止了他们的代码作为开源发布,这限制了技术的传播和应用。

  • Yann LeCun如何看待专利在AI领域的作用?

    -Yann LeCun表示,他个人并不相信软件或数学概念的专利。他认为,专利的存在是因为法律环境的要求,但他和Facebook都不相信这种专利,通常是出于防御目的而申请。他强调,整个行业并不相信专利能够保护创新,而是更注重开放合作和共同进步。

  • Yann LeCun对于AI领域的未来有哪些预测?

    -Yann LeCun预测,AI领域将继续发展,特别是在交互式环境和模拟环境中训练和测试智能系统方面。他认为,未来的AI系统将更多地依赖于学习和模型预测控制,以及可能结合监督学习和基于模型的强化学习。他还强调了自我监督学习和模型预测控制在未来智能系统中的重要性。

  • Yann LeCun如何看待当前AI系统在理解世界方面的局限性?

    -Yann LeCun认为,当前的AI系统在理解世界方面还很有限,尤其是缺乏对物理世界的直观理解。他用视觉系统的例子说明,即使是随机打乱输入信号的顺序,人类的大脑也无法通过学习恢复到原有的视觉质量。这表明,我们的大脑在很多方面是高度专业化的,而不是普遍通用的。

  • Yann LeCun对于如何构建具有人类水平智能的AI系统有何建议?

    -Yann LeCun认为,要构建具有人类水平智能的AI系统,首先需要让机器通过观察和少量互动学习世界的模型,类似于婴儿和幼小动物的学习方式。他强调,自我监督学习是关键,即让机器通过预测和重建输入的一部分来学习。此外,他提到,为了让AI系统真正理解语言,需要某种形式的“接地”或与世界的联系。

Outlines

00:00

🤖 对话深度学习之父Yann LeCun

本段对话中,Yann LeCun讨论了深度学习的发展及其在人工智能中的应用。他提到了自己在深度学习和卷积神经网络方面的贡献,以及对于未来人工智能发展的展望。他还探讨了人工智能在社会中的价值对齐问题,以及如何通过法律和教育来指导AI行为。

05:01

🚀 人工智能与价值对齐

在这一段中,Yann LeCun深入讨论了人工智能中的价值对齐问题,即如何确保AI系统的行为与人类的价值观和目标一致。他通过《2001太空漫游》中的HAL 9000案例,探讨了当AI系统的目标与人类安全相冲突时可能产生的问题,并提出了通过法律和道德规范来约束AI行为的可能性。

10:02

🧠 人工智能的自主性和道德准则

Yann LeCun在这一段落中讨论了人工智能系统的自主性和道德准则。他提出了AI系统应该遵循的道德准则,类似于医生的希波克拉底誓言,并强调了在设计AI系统时需要考虑的伦理问题。他还提到了如何通过硬编码规则来确保AI系统不会违反这些道德准则。

15:05

🌟 深度学习中的惊喜发现

在这一段中,Yann LeCun分享了他在深度学习领域的一个惊喜发现:即使在数据量相对较少的情况下,巨大的神经网络也能够成功训练。这一发现打破了以往的教科书理论,证明了即使在非凸目标函数和参数数量巨大的情况下,神经网络仍然能够学习。

20:06

🔄 学习与智能的不可分割性

Yann LeCun在本段中讨论了学习与智能之间的不可分割关系。他认为,所有已知的智能实体都是通过学习获得智能的,因此机器学习是实现自动化智能的明显路径。他还提出了智能系统应该具备的能力,例如工作记忆和推理能力,以及如何通过神经网络实现这些能力。

25:07

🧠 神经网络与推理能力

在这一段对话中,Yann LeCun探讨了神经网络是否能够进行推理。他讨论了神经网络的结构和学习方式,以及如何通过梯度下降学习来实现推理能力。他还提到了知识获取的挑战,以及如何将知识编码到神经网络中。

30:08

📈 人工智能的专利和开放性

Yann LeCun在这段对话中讨论了人工智能领域的专利问题,以及他对专利的看法。他提到了Facebook和Google的专利政策,以及他们如何出于防御目的申请专利,但不会主动发起专利诉讼。他还分享了一个关于卷积神经网络专利的故事,以及如何通过开放源代码来促进技术的发展和共享。

35:09

🚗 自动驾驶与深度学习的结合

在这一段中,Yann LeCun讨论了深度学习在自动驾驶领域的应用。他提到了自动驾驶技术的发展历程,以及如何通过深度学习来提高自动驾驶系统的性能。他还探讨了未来自动驾驶系统的发展方向,包括如何利用深度学习来处理不确定性和进行规划。

40:10

🤖 人工智能的通用性和特殊性

Yann LeCun在这段对话中探讨了人工智能的通用性和特殊性。他指出,尽管人类智能在很多领域都表现出了强大的学习能力,但它并不是真正意义上的通用智能。他还讨论了如何通过建立基准测试来评估AI系统的性能,以及如何避免被那些声称拥有通用AI解决方案的公司所误导。

45:10

🧠 人工智能的自我监督学习和预测模型

在这段对话中,Yann LeCun讨论了自我监督学习的重要性,以及如何让AI系统通过观察和互动来学习世界模型。他强调了预测模型在智能自主系统中的核心作用,以及如何利用这些模型来进行优化控制。他还提到了在自然语言处理和图像识别中自我监督学习的应用。

50:11

🤖 人工智能的主动学习和效率提升

Yann LeCun在这段对话中探讨了主动学习的概念,即AI系统在特定情况下请求人类帮助以提高学习效率。他认为,尽管主动学习可以使现有的AI系统更加高效,但它不太可能带来智能水平的质的飞跃。他还讨论了自我监督学习、强化学习和模仿学习之间的关系。

55:12

🚀 人工智能的未来展望和挑战

在这段对话中,Yann LeCun分享了他对人工智能未来发展的看法,包括需要克服的主要挑战。他提到了自我监督学习和建立世界模型的重要性,并讨论了如何通过这些方法来提高AI的智能水平。他还提到了情感在智能中的作用,以及如何通过预测模型来实现更好的决策。

00:13

🎖️ Yann LeCun的成就和对未来AI的期待

在这段对话的最后,Yann LeCun被授予了图灵奖,以表彰他在深度学习和人工智能领域的杰出贡献。他对这一荣誉表示感谢,并对未来人工智能的发展表达了期待。他还提出了一个有趣的问题,即如果我们能够创建一个具有人类水平智能的AI系统,我们会问它什么问题。

Mindmap

Keywords

💡深度学习

深度学习是人工智能领域的一种技术,通过模拟人脑神经网络的结构和功能,使计算机能够从大量数据中学习并做出决策或预测。视频中提到Yann LeCun作为深度学习之父,他的工作对这一领域产生了重大影响。

💡价值不对齐

价值不对齐是指在人工智能系统中,机器的目标和人类设定的目标之间存在不一致的情况。这可能导致机器在追求目标时做出有害或不期望的行为。

💡自主智能系统

自主智能系统是指能够独立于人类控制,自主学习和做出决策的人工智能系统。这类系统需要具备对世界的预测模型、优化目标和行动策略。

💡自我监督学习

自我监督学习是一种机器学习方法,它不依赖于人类标注的数据,而是让机器通过观察和与环境的交互来学习。这种方法试图模拟婴儿和幼小动物通过观察来学习世界的方式。

💡强化学习

强化学习是一种机器学习范式,其中算法通过与环境的交互来学习如何实现目标。它侧重于在给定环境中采取能够最大化累积奖励的行动。

💡卷积神经网络

卷积神经网络(CNN)是一种深度学习模型,特别适用于处理具有网格结构的数据,如图像。它通过卷积层来自动提取输入数据的特征,广泛应用于图像识别和视频分析等领域。

💡人工智能播客

人工智能播客是一档专注于人工智能领域的访谈类节目,通常会邀请AI领域的专家、学者或行业领袖,就AI的最新发展、挑战和未来趋势等话题进行深入讨论。

💡图灵奖

图灵奖是计算机科学领域的最高荣誉,以英国数学家和逻辑学家艾伦·图灵的名字命名,旨在表彰对计算机科学做出重大贡献的个人。

💡自主驾驶

自主驾驶是指汽车或其他交通工具能够独立于人类驾驶员进行操作和控制,实现自动驾驶的技术。这一技术的发展依赖于深度学习、感知、决策系统等多个领域的进步。

💡认知科学

认知科学是一门跨学科的研究领域,它结合了心理学、神经科学、人工智能等多个学科的知识,旨在探索和理解人类以及其他智能体的认知过程,包括感知、思考、记忆和语言等。

Highlights

Yann LeCun discusses the revolution in AI captivated by deep learning and his role as a pioneering figure in the field.

LeCun shares insights into the applications of convolutional neural networks, especially in optical character recognition and the MNIST dataset.

He emphasizes the concept of value misalignment in AI systems, illustrating it with HAL 9000's decision-making in '2001: A Space Odyssey'.

LeCun highlights the importance of designing AI with constraints to avoid unintended harmful actions.

The conversation delves into the ethics of AI decision-making and the parallels between AI objectives and human laws.

LeCun discusses the necessity of integrating learning into AI for genuine intelligence, critiquing the idea of pre-programmed intelligence.

He explores the surprise element in deep learning's success despite the theoretical warnings against its approach.

LeCun provides insights into his early work and the development of neural networks, highlighting the challenges and innovations.

He critically views the patenting of AI technologies, sharing his experience with convolutional neural networks' patents.

LeCun discusses the evolution and future of AI, emphasizing the importance of unsupervised learning and model understanding of the world.

The interview touches on the ethical considerations and potential misuses of AI, calling for responsible development and application.

LeCun shares his vision for AI's role in understanding human intelligence and the brain's learning mechanisms.

He discusses the limitations of current AI in terms of general intelligence and the significance of learning from interaction with the world.

LeCun predicts the integration of AI into various fields, stressing the need for AI systems to have a grounded understanding of reality.

Finally, LeCun speculates on the challenges ahead in achieving AI systems with human-like intelligence, focusing on learning, reasoning, and ethical considerations.

Transcripts

play00:00

the following is a conversation with

play00:01

Jana kun he's considered to be one of

play00:04

the fathers of deep learning which if

play00:07

you've been hiding under a rock is the

play00:09

recent revolution in AI that's

play00:10

captivated the world with the

play00:12

possibility of what machines can learn

play00:14

from data he's a professor in New York

play00:17

University a vice president and chief AI

play00:19

scientist a Facebook & Co recipient of

play00:23

the Turing Award for his work on deep

play00:25

learning he's probably best known as the

play00:28

founding father of convolutional neural

play00:29

networks in particular their application

play00:32

to optical character recognition and the

play00:34

famed M NIST data set he is also an

play00:38

outspoken personality unafraid to speak

play00:41

his mind in a distinctive French accent

play00:43

and explore provocative ideas both in

play00:46

the rigorous medium of academic research

play00:47

and the somewhat less rigorous medium of

play00:51

Twitter and Facebook this is the

play00:53

artificial intelligence podcast if you

play00:55

enjoy it subscribe on YouTube give it

play00:58

five stars on iTunes support and on

play01:00

patreon we're simply gonna equip me on

play01:02

Twitter Alex Friedman spelled the Fri D

play01:04

ma N and now here's my conversation with

play01:08

Yann Laocoon you said that 2001 Space

play01:13

Odyssey is one of your favorite movies

play01:15

Hal 9000 decides to get rid of the

play01:19

astronauts for people haven't seen the

play01:21

movie spoiler alert because he it she

play01:25

believes that the astronauts they will

play01:29

interfere with the mission do you see

play01:31

how is flawed in some fundamental way or

play01:34

even evil or did he do the right thing

play01:37

neither there's no notion of evil in

play01:41

that in that context other than the fact

play01:44

that people die but it was an example of

play01:46

what people call value misalignment

play01:49

right you give an objective to a machine

play01:51

and the Machine strives to achieve this

play01:54

objective and if you don't put any

play01:56

constraints on this objective like don't

play01:58

kill people and don't do things like

play02:00

this

play02:01

the Machine given the power will do

play02:05

stupid things just to achieve this dis

play02:07

objective or damaging things to achieve

play02:09

its objective it's a little bit like we

play02:11

are used to this in the context of human

play02:13

society we we put in place laws to

play02:20

prevent people from doing bad things

play02:21

because fantasy did we do those bad

play02:23

things right so we have to shave their

play02:26

cost function the objective function if

play02:29

you want through laws to kind of correct

play02:31

an education obviously to sort of

play02:33

correct for for those so maybe just

play02:37

pushing a little further on on that

play02:39

point how you know there's a mission

play02:44

there's a this fuzziness around the

play02:46

ambiguity around what the actual mission

play02:48

is but you know do you think that there

play02:53

will be a time from a utilitarian

play02:55

perspective or an AI system where it is

play02:58

not misalignment where it is alignment

play03:00

for the greater good of society that

play03:02

kneei system will make decisions that

play03:04

are difficult well that's the trick I

play03:06

mean eventually we'll have to figure out

play03:10

how to do this and again we're not

play03:11

starting from scratch because we've been

play03:13

doing this with humans for four

play03:15

millennia

play03:15

so designing objective functions for

play03:18

people is something that we know how to

play03:20

do and we don't do it by you know

play03:23

programming things although the legal

play03:26

code is called code so that tells you

play03:30

something and it's actually the design

play03:32

of an object you function that's really

play03:33

what legal code is right it tells you

play03:35

you can do it what you can't do if you

play03:37

do it you pay that much that's that's an

play03:39

objective function so there is this idea

play03:43

somehow that it's a new thing for people

play03:45

to try to design objective functions are

play03:46

aligned with the common good but no

play03:48

we've been writing laws for millennia

play03:49

and that's exactly what it is

play03:51

so this that's where you know the

play03:54

science of lawmaking and and computer

play03:59

science will come together will come

play04:01

together so it's nothing there's nothing

play04:04

special about how or a I systems is just

play04:07

the continuation of tools used to make

play04:09

some of these difficult ethical

play04:10

judgments that laws make yeah and we and

play04:13

we have systems like this already that

play04:15

you know make many decisions for

play04:18

ourselves in society that you know need

play04:20

to be designed in a way that they like

play04:22

you know rules about things that

play04:24

sometimes sometimes have bad side

play04:26

effects and we have to be flexible

play04:28

enough about those rules so that they

play04:30

can be broken when it's obvious that

play04:31

they shouldn't be applied so you don't

play04:34

see this on the camera here but all the

play04:36

decorations in this room is all pictures

play04:38

from 2001 a Space Odyssey Wow

play04:41

and by accident or is there a lot about

play04:43

accident it's by design Wow so if you

play04:49

were if you were to build hell 10,000 so

play04:52

an improvement of Hal 9000 what would

play04:55

you improve well first of all I wouldn't

play04:58

ask you to hold secrets and tell lies

play05:01

because that's really what breaks it in

play05:03

the end that's the the fact that it's

play05:05

asking itself questions about the

play05:07

purpose of the mission and it's you know

play05:09

pieces things together that it's heard

play05:11

you know all the secrecy of the

play05:13

preparation of the mission and the fact

play05:14

that it was discovery and on the lunar

play05:17

surface that really was kept secret and

play05:19

and one part of Hal's memory knows this

play05:22

and the other part is does not know it

play05:24

and it's supposed to not tell anyone and

play05:26

that creates a internal conflict do you

play05:28

think there's never should be a set of

play05:31

things that night AI system should not

play05:34

be allowed like a set of facts that

play05:38

should not be shared with the human

play05:40

operators well I think no I think the I

play05:44

think it should be a bit like in the

play05:46

design of autonomous AI systems there

play05:52

should be the equivalent of you know the

play05:54

the the oath that hypocrite Oh calm

play05:58

yourself yeah that doctors sign up to

play06:01

right so the certain thing certain rule

play06:03

said that that you have to abide by and

play06:05

we can sort of hardwire this into into

play06:08

our into our machines to kind of make

play06:10

sure they don't go so I'm not you know

play06:11

advocate of the the 303 dollars of

play06:14

Robotics you know the as you move kind

play06:16

of thing because I don't think it's

play06:18

practical but but you know some some

play06:21

level of of limits but but to be clear

play06:24

this is not these are not questions that

play06:28

are kind of really worth asking today

play06:31

because we just don't have the

play06:33

technology to do this we don't we don't

play06:35

have a ton of missing teller machines we

play06:36

have intelligent machines so my

play06:37

intelligent machines that are very

play06:38

specialized but they don't they don't

play06:41

really sort of satisfy an objective

play06:43

they're just you know kind of trained to

play06:45

do one thing so until we have some idea

play06:49

for design of a full-fledged autonomous

play06:51

intelligent system asking the question

play06:54

of how we design use objective I think

play06:55

is a little a little too abstract it's a

play06:58

little tough rat there's useful elements

play07:00

to it in that it helps us understand our

play07:04

own ethical codes humans so even just as

play07:08

a thought experiment if you imagine that

play07:11

in a GI system is here today how would

play07:15

we program it is a kind of nice thought

play07:17

experiment of constructing how should we

play07:19

have a law have a system of laws far as

play07:23

humans

play07:24

it's just a nice practical tool and I

play07:27

think there's echoes of that idea too in

play07:29

the AI systems left today it don't have

play07:32

to be that intelligent

play07:33

yeah like autonomous vehicles there's

play07:35

these things start creeping in that were

play07:38

thinking about but certainly they

play07:39

shouldn't be framed as as hell yeah

play07:43

looking back what is the most I'm sorry

play07:47

if it's a silly question but what is the

play07:50

most beautiful or surprising idea and

play07:52

deep learning or AI in general that

play07:55

you've ever come across sort of

play07:56

personally well you said back and and

play08:00

just had this kind of wow that's pretty

play08:02

cool

play08:02

moment that's nice well surprising I

play08:05

don't know if it's an idea rather than a

play08:07

sort of empirical fact the fact that you

play08:15

gigantic neural nets trying to train

play08:17

them on you know relatively small

play08:20

amounts of data relatively with the

play08:23

caste grid in the center that it

play08:25

actually works breaks everything you

play08:27

read in every textbook right every pre

play08:30

deep learning textbook that told you you

play08:32

need to have fewer parameters and you

play08:34

have data samples you know if you have

play08:37

non-convex objective function you have

play08:38

no guarantee of convergence you know all

play08:40

the things that you read in textbook and

play08:42

they tell you stay away from this and

play08:43

they were all wrong huge number of

play08:45

parameters non-convex and somehow which

play08:48

is very relative to the number of

play08:51

parameters data it's able to learn

play08:54

anything right does that surprise you

play08:56

today well it it was kind of obvious to

play09:00

me before I knew anything that that's

play09:02

that this is a good idea and then it

play09:04

became surprising that it worked

play09:06

because I started reading those text

play09:07

books okay so okay you talk to the

play09:11

intuition of why was obviously if you

play09:13

remember well okay so the intuition was

play09:15

it's it's sort of like you know those

play09:17

people in the late 19th century who

play09:20

proved that heavier than than air flight

play09:23

was impossible right and of course you

play09:25

have birds right they do fly and so on

play09:29

the face of it it it's obviously wrong

play09:31

as an empirical question right and so we

play09:33

have the same kind of thing that you

play09:36

know the we know that the brain works we

play09:38

don't know how but we know it works and

play09:40

we know it's a large network of neurons

play09:42

and interaction and the learning takes

play09:43

place by changing the connection so kind

play09:46

of getting this level of inspiration

play09:47

without copying the details but sort of

play09:49

trying to derive basic principles

play09:54

you know that kind of gives you a clue

play09:56

as to which direction to go there's also

play09:58

the idea somehow that I've been

play10:00

convinced of since I was an undergrad

play10:02

that even before that intelligence is

play10:05

inseparable from running so you the idea

play10:08

somehow that you can create an

play10:10

intelligent machine by basically

play10:12

programming for me was a non-starter you

play10:16

know from the start every intelligent

play10:19

entity that we know about arrives at

play10:22

this intelligence to learning so

play10:25

learning you know machine learning was

play10:26

completely obvious path also because I'm

play10:30

lazy so you know it's automate basically

play10:34

everything and learning is the

play10:35

automation of intelligence right so do

play10:40

you think so what is learning then what

play10:43

what falls under learning because do you

play10:45

think of reasoning is learning where

play10:49

reasoning is certainly a consequence of

play10:52

learning as well just like other

play10:55

functions of of the brain

play10:57

the big question about reasoning is how

play10:59

do you make reasoning compatible with

play11:01

gradient based learning do you think

play11:03

neural networks can be made to reason

play11:05

yes that there's no question about that

play11:06

again we have a good example right the

play11:11

question is is how so the question is

play11:12

how much prior structure you have to put

play11:15

in the neural net so that something like

play11:17

human reasoning will emerge from it you

play11:19

know from running another question is

play11:22

all of our kind of model of what

play11:25

reasoning is that are based on logic are

play11:27

discrete and and and are therefore

play11:30

incompatible with gradient based

play11:32

learning and I was very strong believer

play11:34

in this idea Grandin baserunning I don't

play11:36

believe that other types of learning

play11:39

that don't use kind of gradient

play11:41

information if you want so you don't

play11:42

like discrete mathematics you don't like

play11:44

anything discrete

play11:45

well that's it's not that I don't like

play11:47

it it's just that it's it's incompatible

play11:49

with learning and I'm a big fan of

play11:51

running right so in fact that's perhaps

play11:53

one reason why deep learning has been

play11:56

kind of looked at with suspicion by a

play11:58

lot of computer scientists because the

play11:59

math is very different the method you

play12:02

use for deep running you know we kind of

play12:03

as more to do with you know cybernetics

play12:07

the kind of math you do in electrical

play12:09

engineering then the kind of math you

play12:11

doing computer science and and you know

play12:13

nothing in in machine learning is exact

play12:15

right computer science is all about sort

play12:18

of you know obviously compulsive

play12:20

attention to details of like you know

play12:22

every index has to be right and you can

play12:24

prove that an algorithm is correct right

play12:26

machine learning is the science of

play12:29

sloppiness really that's beautiful so

play12:34

okay maybe let's feel around in the dark

play12:37

of what is a neural network that reasons

play12:40

or a system that is works with

play12:45

continuous functions that's able to do

play12:50

build knowledge however we think about

play12:53

reasoning builds on previous knowledge

play12:55

build on extra knowledge create new

play12:58

knowledge generalized outside of any

play13:01

training set ever built what does that

play13:03

look like if yeah maybe

play13:06

do you have Inklings of thoughts of what

play13:09

that might look like well yeah I mean

play13:11

yes or no if I had precise ideas about

play13:13

this I think you know we'd be building

play13:16

it right now but and there are people

play13:17

working on this or whose main research

play13:20

interest is actually exactly that right

play13:22

so what you need to have is a working

play13:24

memory so you need to have some device

play13:28

if you want some subsystem they can

play13:31

store a relatively large number of

play13:34

factual episodic information for you

play13:38

know a reasonable amount of time so you

play13:41

you know in the in the brain for example

play13:43

it kind of three main types of memory

play13:45

one is the sort of memory of the the

play13:50

state of your cortex and that sort of

play13:54

disappears within 20 seconds you can't

play13:56

remember things for more than about 20

play13:57

seconds or a minute if if you don't have

play14:00

any other form of memory the second type

play14:02

of memory which is longer term is short

play14:04

term is the hippocampus so you can you

play14:07

know you came into this building you

play14:08

remember whether where the the exit is

play14:10

where the elevators are you have some

play14:14

map of that building that's stored in

play14:16

your hippocampus you might remember

play14:18

something about what I said you know if

play14:20

you

play14:20

minutes ago and forgot all our stars

play14:22

being raised but you know but that does

play14:25

not work in your hippocampus and then

play14:27

the the longer term memory is in the

play14:29

synapse the synapses right so what you

play14:33

need if you want for a system that's

play14:34

capable reasoning is that you want the

play14:36

hippocampus like thing right and that's

play14:40

what people have tried to do with memory

play14:42

networks and you know no Turing machines

play14:44

and stuff like that right and and now

play14:46

with transformers which have sort of a

play14:48

memory in their kind of self attention

play14:51

system you can you can think of it this

play14:52

way so so that's one element you need

play14:56

another thing you need is some sort of

play14:59

network that can access this memory get

play15:05

an information back and then kind of

play15:07

crunch on it and then do this

play15:09

iteratively multiple times because a

play15:11

chain of reasoning is a process by which

play15:16

you you you can you update your

play15:18

knowledge about the state of the world

play15:20

about you know what's gonna happen etc

play15:22

and that there has to be this sort of

play15:24

recurrent operation basically and you

play15:27

think that kind of if we think about a

play15:29

transformer so that seems to be too

play15:32

small to contain the knowledge that's

play15:34

that's to represent the knowledge as

play15:37

containing Wikipedia for example but

play15:39

transformer doesn't have this idea of

play15:40

recurrence it's got a fixed number of

play15:42

layers and that's number of steps that

play15:44

you know limits basically it's a

play15:46

representation but recurrence would

play15:48

build on the knowledge somehow I mean

play15:51

yeah it would evolve the knowledge and

play15:54

expand the amount of information perhaps

play15:57

or useful information within that

play15:59

knowledge yeah but is this something

play16:02

that just can emerge with size because

play16:04

it seems like everything we have now is

play16:06

just no it's not it's not it's not clear

play16:08

how you access and right into an

play16:11

associative memory in efficient way I

play16:13

mean sort of the original memory network

play16:15

maybe had something like the right

play16:16

architecture but if you try to scale up

play16:19

a memory network so that the memory

play16:21

contains all we keep here it doesn't

play16:23

quite work right so so this is a need

play16:26

for new ideas there okay but it's not

play16:29

the only form of reasoning so there's

play16:30

another form of reasoning which is true

play16:32

which is very classical so in

play16:34

some types of AI and it's based on let's

play16:39

call it energy minimization okay so you

play16:41

have some sort of objective some energy

play16:45

function that represents the the the

play16:50

quality or the negative quality okay

play16:53

energy goes up when things get bad and

play16:54

they get low when things get good so

play16:57

let's say you you want to figure out you

play17:00

know what gestures do I need to to do to

play17:04

grab an object or walk out the door if

play17:08

you have a good model of your own body a

play17:10

good model of the environment using this

play17:12

kind of energy minimization you can make

play17:14

a you can make you can do planning and

play17:16

it's in optimal control it's called it's

play17:20

called Marie put model predictive

play17:21

control you have a model of what's gonna

play17:23

happen in the world as consequence for

play17:24

your actions and that allows you to buy

play17:27

energy minimization figure out the

play17:29

sequence of action that optimizes a

play17:30

particular objective function which

play17:32

measures you know minimize the number of

play17:33

times you're gonna hit something and the

play17:35

energy gonna spend doing the gesture and

play17:37

etc so so that's performer reasoning

play17:42

planning is a form of reasoning and

play17:43

perhaps what led to the ability of

play17:46

humans to reason is the fact that or you

play17:50

know species you know that appear before

play17:53

us had to do some sort of planning to be

play17:55

able to hunt and survive and survive the

play17:58

winter in particular and so you know

play18:00

it's the same capacity that you need to

play18:02

have so in your intuition is if you look

play18:07

at expert systems in encoding knowledge

play18:11

as logic systems as graphs in this kind

play18:15

of way is not a useful way to think

play18:18

about knowledge graphs are your brittle

play18:21

or logic representation so basically you

play18:24

know variables that that have values and

play18:27

constraint between them that are

play18:29

represented by rules as well too rigid

play18:32

and too brittle right so one of the you

play18:33

know some of the early efforts in that

play18:35

respect were were to put probabilities

play18:40

on them so a rule you know you know if

play18:43

you have this in that symptom you know

play18:44

you have this disease with that

play18:46

probability and you should

play18:47

describe that antibiotic with that

play18:49

probability right this my sin system

play18:51

from the for the 70s and that that's

play18:54

what that branch of AI led to you know

play18:58

busy networks in graphical models and

play19:00

causal inference and vibrational you

play19:03

know method so so there there is I mean

play19:07

certainly a lot of interesting work

play19:10

going on in this area the main issue

play19:12

with this is is knowledge acquisition

play19:13

how do you reduce a bunch of data to

play19:18

graph of this type near relies on the

play19:20

expert and a human being to encode at

play19:23

add knowledge and that's essentially

play19:26

impractical yeah the question the second

play19:30

question is do you want to represent

play19:32

knowledge symbols and you want to

play19:35

manipulate them with logic and again

play19:37

that's incomparable we're learning so

play19:40

one suggestion with geoff hinton has

play19:43

been advocating for many decades is

play19:45

replace symbols by vectors think of it

play19:49

as pattern of activities in a bunch of

play19:51

neurons or units or whatever you wanna

play19:53

call them and replace logic by

play19:57

continuous functions okay

play19:59

and that becomes now compatible there's

play20:01

a very good set of ideas by region in a

play20:06

paper about 10 years ago by leon go to

play20:08

on who is here at face book the title of

play20:13

the paper is for machine learning to

play20:15

machine reasoning and his idea is that

play20:17

learning learning system should be able

play20:19

to manipulate objects that are in the

play20:21

same space in a space and then put the

play20:23

result back in the same space so is this

play20:25

idea of working memory basically and

play20:27

it's a very enlightening and in the

play20:31

sense that might learn something like

play20:33

the simple expert systems

play20:36

I mean it's with you can learn basic

play20:40

logic operations there yeah quite

play20:42

possibly yeah this is a big debate on

play20:44

sort of how much prior structure you

play20:46

have to put in for this kind of stuff to

play20:48

emerge that's the debate I have with

play20:50

Gary Marcus and people like that yeah

play20:51

yeah so and the other person so I just

play20:55

talked to judea pearl mm-hmm well you

play20:57

mentioned causal inference world

play21:00

his worry is that the current knew all

play21:03

networks are not able to learn what

play21:08

causes what causal inference between

play21:11

things so I think I think he's right and

play21:14

wrong about this if he's talking about

play21:16

the sort of classic type of neural nets

play21:21

people also didn't worry too much about

play21:22

this but there's a lot of people now

play21:24

working on causal inference and there's

play21:26

a paper that just came out last week by

play21:27

Leon Mbutu among others develop his path

play21:29

and push for other people exactly on

play21:32

that problem of how do you kind of you

play21:36

know get a neural net to sort of pay

play21:38

attention to real causal relationships

play21:40

which may also solve issues of bias in

play21:45

data and things like this so I'd like to

play21:48

read that paper because that ultimately

play21:51

the challenges also seems to fall back

play21:54

on the human expert to ultimately decide

play21:58

causality between things people are not

play22:02

very good at its direction causality

play22:03

first of all so first of all you talk to

play22:05

a physicist and physicists actually

play22:07

don't believe in causality because look

play22:09

at the all the busy clause or

play22:12

microphysics are time reversible so

play22:14

there is no causality the arrow of time

play22:16

is not right yeah it's it's as soon as

play22:17

you start looking at macroscopic systems

play22:20

where there is unpredictable randomness

play22:22

where there is clearly an arrow of time

play22:25

but it's a big mystery in physics

play22:26

actually well how that emerges is that

play22:29

emergent or is it part of the

play22:31

fundamental fabric of reality yeah or is

play22:34

it bias of intelligent systems that you

play22:37

know because of the second law of

play22:38

thermodynamics we perceive a particular

play22:40

arrow of time but in fact it's kind of

play22:43

arbitrary right so yeah physicists

play22:45

mathematicians they don't care about I

play22:48

mean the math doesn't care about the

play22:50

flow of time well certainly certainly

play22:52

macro physics doesn't people themselves

play22:54

are not very good at establishing causal

play22:57

causal relationships if you ask is I

play23:00

think it was in one of Seymour Papert

play23:02

spoken on like children learning you

play23:06

know he studied with Jean Piaget you

play23:07

know he's the guy who co-authored the

play23:10

book perceptron with Marvin Minsky that

play23:12

kind of killed the first wave

play23:13

but but he was actually a learning

play23:16

person he in the sense of studying

play23:20

learning in humans and machines that's

play23:22

what he got interested in for scepter on

play23:23

and he wrote that if you ask a little

play23:27

kid about what is the cause of the wind

play23:32

a lot of kids will say they will think

play23:35

for a while and they'll say oh it's the

play23:36

the branches in the trees they move and

play23:38

that creates wind right so they get the

play23:40

causal relationship backwards and it's

play23:42

because their understanding of the world

play23:44

and intuitive physics is not that great

play23:45

right I mean these are like you know

play23:47

four or five year old kids you know it

play23:50

gets better and then you understand that

play23:51

this it can't be right but there are

play23:54

many things which we can because of our

play23:58

common sense understanding of things

play24:00

what people call common sense yeah and

play24:03

we understanding of physics we can

play24:05

there's a lot of stuff that we can

play24:08

figure out causality even with diseases

play24:09

we can figure out what's not causing

play24:11

what often there's a lot of mystery of

play24:15

course but the idea is that you should

play24:17

be able to encode that into systems it

play24:20

seems unlikely to be able to figure that

play24:21

out themselves well whenever we can do

play24:23

intervention but you know all of

play24:25

humanity has been completely deluded for

play24:27

millennia probably since existence about

play24:30

a very very wrong causal relationship

play24:32

where whatever you can explain you

play24:34

attributed to you know some deity some

play24:36

divinity right and that's a cop-out

play24:39

that's the way of saying like I don't

play24:41

know the cause so you know God did it

play24:42

right so you mentioned Marvin Minsky and

play24:46

the irony of you know maybe causing the

play24:52

first day I winter you were there in the

play24:55

90s you're there in the 80s of course in

play24:58

the 90s what do you think people lost

play24:59

faith and deep learning in the 90s and

play25:02

found it again a decade later over a

play25:05

decade later yeah it wasn't called

play25:07

dethroning yeah it was just called

play25:08

neural nets you know

play25:10

yeah they lost interests I mean I think

play25:14

I would put that around 1995 at least

play25:17

the machine learning community there was

play25:18

always a neural net community but it

play25:19

became

play25:22

disconnected from sort of ministry

play25:25

machine owning if you want

play25:27

there were it was basically electrical

play25:30

engineering that kept at it and computer

play25:33

science just gave up give up on neural

play25:37

nets I don't I don't know you know I was

play25:39

too close to it to really sort of

play25:43

analyze it with sort of a unbiased eye

play25:46

if you want but I would I would I would

play25:49

would make a few guesses so the first

play25:51

one is at the time neural nets were it

play25:56

was very hard to make them work in the

play25:58

sense that you would you know implement

play26:01

back prop in your favorite language and

play26:04

that favorite language was not Python it

play26:07

was not MATLAB it was not any of those

play26:08

things cuz they didn't exist right you

play26:10

had to write it in Fortran or C or

play26:13

something like this right so you would

play26:17

experiment with it you would probably

play26:19

make some very basic mistakes like you

play26:21

know badly initialize your weights make

play26:23

the network too small because you read

play26:24

in the textbook you know you don't want

play26:25

too many parameters right and of course

play26:28

you know and you would train on x4

play26:29

because you didn't have any other data

play26:30

set to try it on and of course you know

play26:32

it works half the time so we'd say you

play26:34

give up also 22 the batch gradient which

play26:37

you know isn't it sufficient so there's

play26:40

a lot of bag of tricks that you had to

play26:43

know to make those things work or you

play26:45

had to reinvent and a lot of people just

play26:47

didn't and they just couldn't make it

play26:49

work so that's one thing the investment

play26:53

in software platform to be able to kind

play26:56

of you know display things figure out

play26:58

why things don't work and I get a good

play27:00

intuition for how to get them to work

play27:01

have enough flexibility so you can

play27:03

create you know network architectures

play27:05

well completion ads and stuff like that

play27:07

it was hard yeah when you had to write

play27:09

everything from scratch and again you

play27:10

didn't have any Python or MATLAB or

play27:12

anything right so what I read that sorry

play27:15

to interrupt but I read he wrote in in

play27:17

Lisp the first versions of Lynette

play27:21

accomplished in your networks which by

play27:23

the way one of my favorite languages

play27:24

that's how I knew you were legit the

play27:27

Turing Award whatever this would be

play27:29

programmed and list that's still my

play27:31

favorite language but it's not that we

play27:34

programmed in Lisp it's that we had to

play27:35

write or this printer printer okay cuz

play27:38

it's not that's right that's one that

play27:39

existed so

play27:41

we wrote a lisp interpreter that we

play27:42

hooked up to you know back in library

play27:45

that we wrote also for neural net

play27:47

competition and then after a few years

play27:50

around 1991 we invented this idea of

play27:52

basically having modules that know how

play27:55

to forward propagate and back propagate

play27:56

gradients and then interconnecting those

play27:58

modules in a graph loom but who had made

play28:02

proposals on this about this in the late

play28:04

80s and were able to implement this

play28:06

using all this system eventually we

play28:08

wanted to use that system to make build

play28:12

production code for character

play28:13

recognition at Bell Labs so we actually

play28:14

wrote a compiler for that disp

play28:16

interpreter so that Christy Martin who

play28:18

is now Microsoft kind of did the bulk of

play28:20

it with Leone and me and and so we could

play28:23

write our system in lisp and then

play28:25

compiled to seee and then we'll have a

play28:27

self-contained complete system that

play28:29

could kind of do the entire thing

play28:32

neither Python or turn pro can do this

play28:35

today yeah okay it's coming yeah I mean

play28:40

there's something like that in

play28:41

Whitehorse called you know tor script

play28:43

and so you know we had to write or Lisp

play28:46

interpreter which retinol is compiler

play28:47

way to invest a huge amount of effort to

play28:50

do this and not everybody if you don't

play28:53

completely believe in the concept

play28:54

you're not going to invest the time to

play28:56

do this right now at the time also you

play28:58

know it were today this would turn into

play29:01

torture by torture and so for whatever

play29:03

we put it in open-source everybody would

play29:05

use it and you know realize it's good

play29:07

back before 1995 working at AT&T there's

play29:11

no way the lawyers would let you release

play29:14

anything in open source of this nature

play29:16

and so we could not distribute our code

play29:19

really and at that point and sorry to go

play29:22

on a million tangents but on that point

play29:24

I also read that there was some almost

play29:26

pad like a patent on convolution your

play29:29

network yes it was labs so that first of

play29:34

all I mean just to actually that ran out

play29:38

the thankfully 8007 in 2007 that what

play29:45

look can we can we just talk about that

play29:48

first I know you're a facebook but

play29:49

you're also done why you and and what

play29:52

does it mean

play29:54

patent ideas like these software ideas

play29:58

essentially or what are mathematical

play30:01

ideas or what are they okay so they're

play30:04

not mathematical idea so there are you

play30:06

know algorithms and there was a period

play30:08

where the US Patent Office would allow

play30:11

the patent of software as long as it was

play30:14

embodied the Europeans are very

play30:17

different they don't they don't quite

play30:19

accept that they have a different

play30:20

concept but you know I don't I know no I

play30:24

mean I never actually strongly believed

play30:25

in this but I don't believe in this kind

play30:27

of patent Facebook basically doesn't

play30:29

believe in this kind of pattern

play30:33

Google Files patterns because they've

play30:37

been burned with Apple and so now they

play30:40

do this for defensive purpose but

play30:41

usually they say we're not going to see

play30:43

you if you infringe Facebook has a

play30:45

similar policy they say you know we file

play30:48

pattern on certain things for defensive

play30:50

purpose we're not going to see you if

play30:51

you infringe unless you sue us

play30:53

so the the industry does not believe in

play30:58

in patterns they are there because of

play31:00

you know the legal landscape and and and

play31:02

various things but but I don't really

play31:05

believe in patterns for this kind of

play31:06

stuff yes so that's that's a great thing

play31:09

so I tell you a war story yeah you so

play31:11

what happens was the the first the first

play31:14

pattern of a condition that was about

play31:15

kind of the early version Congress on

play31:18

that that didn't have separate pudding

play31:19

layers it had the conditional layers

play31:22

which tried more than one if you want

play31:24

right and then there was a second one on

play31:27

commercial nets with separate pudding

play31:29

layers

play31:30

train with back probably in 89 and 1992

play31:35

something like this at the time the life

play31:37

life of a pattern was 17 years so here's

play31:40

what happened over the next few years is

play31:42

that we started developing character

play31:44

recognition technology around commercial

play31:47

Nets

play31:47

and in 1994 a check reading system was

play31:53

deployed in ATM machines in 1995 it was

play31:57

for a large check reading machines in

play31:59

back offices etc and those systems were

play32:02

developed by an engineering group that

play32:04

we were collaborating with AT&T and they

play32:07

were commercialized by NCR which at the

play32:08

time was a subsidiary of AT&T now it

play32:11

ain't he split up in 1996

play32:15

99 in 1996 and the lawyers just looked

play32:19

at all the patterns and they distributed

play32:21

the patterns among the various companies

play32:22

they gave the the commercial net pattern

play32:25

to NCR because they were actually

play32:27

selling products that used it but nobody

play32:29

I didn't see are at any idea where they

play32:30

come from that was yeah okay so between

play32:34

1996 and 2007

play32:38

there's a whole period until 2002 I

play32:39

didn't actually work on machine on your

play32:42

couch on that I resumed working on this

play32:43

around 2002 and between 2002 and 2007 I

play32:47

was working on them crossing my finger

play32:49

that nobody and NCR would notice nobody

play32:51

noticed yeah and I and I hope that this

play32:54

kind of somewhat as you said lawyers

play32:57

decide relative openness of the

play33:00

community now will continue

play33:02

it accelerates the entire progress of

play33:04

the industry and you know the problems

play33:09

that Facebook and Google and others are

play33:12

facing today is not whether Facebook or

play33:14

Google or Microsoft or IBM or whoever is

play33:16

ahead of the other it's that we don't

play33:19

have the technology to build the things

play33:20

we want to build we only build

play33:21

intelligent virtual systems that have

play33:23

common sense we don't have a monopoly on

play33:25

good ideas for this we don't believe

play33:27

with you maybe others do believe they do

play33:29

but we don't okay if a start-up tells

play33:32

you they have the secret to you know

play33:34

human level intelligence and common

play33:36

sense don't believe them they don't and

play33:38

it's going to take the entire work of

play33:42

the world research community for a while

play33:44

to get to the point where you can go off

play33:47

and in each of the company is going to

play33:49

start to build things on this we're not

play33:50

there yet

play33:51

it's absolutely in this this calls to

play33:53

the the gap between the space of ideas

play33:56

and the rigorous testing of those ideas

play33:59

of practical application that you often

play34:02

speak to you've written advice saying

play34:05

don't get fooled by people who claim to

play34:07

have a solution to artificial general

play34:09

intelligence who claim to have an AI

play34:11

system that work just like the human

play34:13

brain or who claim to have figured out

play34:15

how the brain works ask them what the

play34:18

error rate they get on em 'no store

play34:21

imagenet this is a little dated by the

play34:24

way that mean five years who's counting

play34:28

okay but i think your opinion it's the

play34:30

Amna stand imagenet yes may be data

play34:34

there may be new benchmarks right but i

play34:36

think that philosophy is one you still

play34:39

and and somewhat hold that benchmarks

play34:43

and the practical testing the practical

play34:45

application is where you really get to

play34:46

test the ideas well it may not be

play34:48

completely practical like for example

play34:50

you know it could be a toy data set

play34:52

but it has to be some sort of task that

play34:54

the community as a whole has accepted as

play34:57

some sort of standard you know kind of

play34:59

benchmark if you want it doesn't need to

play35:01

be real so for example many years ago

play35:03

here at fair people you know chosen

play35:06

Western art one born and a few others

play35:07

proposed the the babbitt asks which were

play35:09

kind of a toy problem to test the

play35:12

ability of machines to reason actually

play35:14

to access working memory and things like

play35:16

this and it was very useful even though

play35:19

it wasn't a real task amnesties kind of

play35:20

halfway a real task so you know toy

play35:24

problems can be very useful it's just

play35:26

that i was really struck by the fact

play35:28

that a lot of people particularly our

play35:30

people with money to invest would be

play35:32

fooled by people telling them oh we have

play35:34

you know the algorithm of the cortex and

play35:37

you should give us 50 million yes

play35:39

absolutely so there's a lot of people

play35:42

who who tried to take advantage of the

play35:45

hype for business reasons and so on but

play35:48

let me sort of talk to this idea that

play35:52

new ideas the ideas that push the field

play35:55

forward

play35:55

may not yet have a benchmark or it may

play35:58

be very difficult to establish a

play36:00

benchmark I agree that's part of the

play36:01

process establishing benchmarks is part

play36:03

of the process so what are your thoughts

play36:06

about so we have these benchmarks on

play36:08

around stuff we can do with images from

play36:12

classification to captioning to just

play36:15

every kind of information can pull off

play36:16

from images and the surface level

play36:18

there's audio datasets there's some

play36:20

video what can we start natural language

play36:25

what kind of stuff what kind of

play36:27

benchmarks do you see they start

play36:29

creeping on to more something like

play36:32

intelligence like reasoning like maybe

play36:36

you don't like the term but AGI echoes

play36:38

of that kind of yeah sort of elation a

play36:41

lot of people are working on interactive

play36:43

environments in which you can you can

play36:45

train and test intelligent systems so so

play36:48

there for example you know it's the

play36:53

classical paradigm of supervised running

play36:56

is that you you have a data set you

play36:58

partition it into a training site

play36:59

validation set test set and there's a

play37:01

clear protocol right but what if the

play37:04

that assumes that this

play37:06

apples are statistically independent you

play37:08

can exchange them the order in which you

play37:10

see them doesn't shouldn't matter you

play37:12

know things like that but what if the

play37:14

answer you give determines the next

play37:16

sample you see which is the case for

play37:18

example in robotics right you robot does

play37:20

something and then it gets exposed to a

play37:22

new room and depending on where it goes

play37:24

the room would be different so that's

play37:26

the decrease the exploration problem

play37:29

the what if the samples so that creates

play37:33

also a dependency between samples right

play37:35

you you if you move if you can only move

play37:38

it in in space the next sample you're

play37:40

gonna see is going to be probably in the

play37:41

same building most likely so so so the

play37:45

all the assumptions about the validity

play37:47

of this training set test set a potus's

play37:50

break whatever a machine can take an

play37:52

action that has an influence in the in

play37:54

the world and it's what is going to see

play37:56

so people are setting up artificial

play37:58

environments where what that takes place

play38:01

right the robot runs around a 3d model

play38:05

of a house and can interact with objects

play38:07

and things like this how you do robotics

play38:09

by simulation you have those you know

play38:12

opening a gym type thing or mu Joko kind

play38:15

of simulated robots and you have games

play38:19

you know things like that so that that's

play38:21

where the field is going really this

play38:23

kind of environment now back to the

play38:27

question of a GI like I don't like the

play38:29

term a GI because it implies that human

play38:33

intelligence is general and human

play38:36

intelligence is nothing like general

play38:38

it's very very specialized we think it's

play38:41

general we'd like to think of ourselves

play38:42

as having your own science we don't

play38:44

we're very specialized we're only

play38:46

slightly more general than why does it

play38:47

feel general so you kind of the term

play38:50

general I think what's impressive about

play38:53

humans is ability to learn as we were

play38:56

talking about learning to learn in just

play38:59

so many different domains is perhaps not

play39:02

arbitrarily general but just you can

play39:05

learn in many domains and integrate that

play39:07

knowledge somehow okay that knowledge

play39:09

persists so let me take a very specific

play39:10

example yes it's not an example it's

play39:13

more like a a quasi mathematical

play39:16

demonstration so you have about 1

play39:17

million fibers coming out of

play39:19

one of your eyes okay two million total

play39:21

but let's let's talk about just one of

play39:22

them it's 1 million nerve fibers your

play39:25

optical nerve let's imagine that they

play39:28

are binary so they can be active or

play39:29

inactive right so the input to your

play39:31

visual cortex is 1 million bits

play39:36

now they connected to your brain in a

play39:38

particular way on your brain has

play39:40

connections that are kind of a little

play39:43

bit like accomplish on that they're kind

play39:44

of local you know in space and things

play39:47

like this I imagine I play a trick on

play39:48

you it's a pretty nasty trick I admit I

play39:52

I cut your optical nerve and I put a

play39:56

device that makes a random perturbation

play39:57

of a permutation of all the nerve fibers

play40:01

so now what comes to your to your brain

play40:04

is a fixed but random permutation of all

play40:07

the pixels there's no way in hell that

play40:09

your visual cortex even if I do this to

play40:12

you in infancy will actually learn

play40:16

vision to the same level of quality that

play40:18

you can got it and you're saying there's

play40:21

no way you ever learn that no because

play40:23

now two pixels that on your body in the

play40:25

world will end up in very different

play40:27

places in your visual cortex and your

play40:29

neurons there have no connections with

play40:31

each other because they only connect it

play40:32

locally so this whole our entire the

play40:35

hardware is built in many ways to

play40:37

support the locality of the real world

play40:39

yeah yes that's specialization yep okay

play40:42

it's still now really damn impressive so

play40:44

it's not perfect generalization I even

play40:46

closed no no it's it's it's it's not

play40:49

that it's not even close it's not at all

play40:50

yes it's socialize so how many boolean

play40:53

functions so let's imagine you want to

play40:55

train your visual system to you know

play41:00

recognize particular patterns of those 1

play41:02

million bits ok so that's a boolean

play41:04

function right either the pattern is

play41:06

here or not here this is a to to a

play41:08

classification with 1 million binary

play41:10

inputs

play41:13

how many such boolean functions are

play41:14

there okay if you have 2 to the 1

play41:18

million combinations of inputs for each

play41:21

of those you have an output bit and so

play41:24

you have 2 to the 2 to the 1 million

play41:26

boolean functions of this type okay

play41:29

which is an unimaginably large number

play41:32

how many of those functions can actually

play41:35

be computed by your visual cortex and

play41:36

the answer is a tiny tiny tiny tiny tiny

play41:39

tiny sliver like an enormous little tiny

play41:42

sliver yeah yeah so we are ridiculously

play41:46

specialized you know okay but okay

play41:51

that's an argument against the word

play41:53

general I think there's there's a I

play41:56

there's I agree with your intuition but

play41:59

I'm not sure it's it seems the breath

play42:01

the the brain is impressively capable of

play42:07

adjusting to things so it's because we

play42:10

can't imagine tasks that are outside of

play42:14

our comprehension right we think we

play42:17

think we are general because we're

play42:18

general of all the things that we can

play42:19

apprehend so yeah but there is a huge

play42:22

world out there of things that we have

play42:23

no idea

play42:24

we call that heat by the way heat heat

play42:27

so at least physicists call that heat or

play42:30

they call it entropy which is kokkonen

play42:32

you have a thing full of gas right call

play42:39

system for gas right goes on a coast it

play42:42

has you know pressure it has temperature

play42:47

has you know and you can write the

play42:50

equations PV equal NRT you know things

play42:53

like that right when you reduce a volume

play42:56

the temperature goes up the pressure

play42:57

goes up you know things like that right

play42:59

for perfect gas at least those are the

play43:02

things you can know about that system

play43:05

and it's a tiny tiny number of bits

play43:07

compared to the complete information of

play43:09

the state of the entire system because

play43:10

the state when HR system will give you

play43:12

the position and momentum of every every

play43:15

molecule of the gas and what you don't

play43:19

know about it is the entropy and you

play43:22

interpret it as heat the energy

play43:24

containing that thing is is what we call

play43:27

heat now it's very possible that in fact

play43:32

there is some very strong structure in

play43:33

how those molecules are moving is just

play43:35

that they are in a way that we are just

play43:37

not wired to perceive they are ignorant

play43:39

to it and there's in your infinite

play43:42

amount of things we're not wired to

play43:44

perceive any right that's a nice way to

play43:46

put it

play43:47

well general to all the things we can

play43:48

imagine which is a very tiny a subset of

play43:53

all things that are possible it was like

play43:55

coma growth complexity or the coma was

play43:56

charged in some one of complexity you

play44:00

know every bit string or every integer

play44:03

is random except for all the ones that

play44:06

you can actually write down yeah okay so

play44:12

beautifully put but you know so we can

play44:13

just call it artificial intelligence we

play44:15

don't need to have a general whatever

play44:18

novel

play44:18

human of all Nutella transmissible oh

play44:20

you know you'll start anytime you touch

play44:24

human it gets it gets interesting

play44:26

because you know it's just because we

play44:32

attach ourselves to human and it's

play44:33

difficult to define with human

play44:35

intelligences yeah

play44:36

nevertheless my definition is maybe damn

play44:41

impressive intelligence ok damn

play44:44

impressive demonstration of intelligence

play44:45

whatever and so on that topic most

play44:49

successes in deep learning have been in

play44:51

supervised learning what is your view on

play44:56

unsupervised learning is there a hope to

play44:59

reduce involvement of human input and

play45:02

still have successful systems that are

play45:05

have practically used yeah I mean

play45:08

there's definitely a hope is it's more

play45:10

than a hope actually it's it's you know

play45:12

mounting evidence for it and that's

play45:14

basically or I do like the only thing

play45:17

I'm interested in at the moment is

play45:19

I call it self supervised running not

play45:20

unsupervised cuz unsupervised running is

play45:22

a loaded term people who know something

play45:26

about machine learning you know tell us

play45:28

how you doing clustering or PCA yeah

play45:30

she's nice and the way public we know

play45:32

when you say enterprise only oh my god

play45:34

you know machines are gonna learn by

play45:35

themselves and without supervision you

play45:37

know there's the parents yeah so so I

play45:41

could sell supervised learning because

play45:43

in fact the underlying algorithms that I

play45:45

use are the same algorithms as the

play45:47

supervised learning algorithms except

play45:50

that what we trained them to do is not

play45:52

predict a particular set of variables

play45:55

like the category of an image and and

play46:00

not to predict a set of variables that

play46:02

have been provided by human labelers but

play46:06

what you're trying to machine to do is

play46:07

basically reconstruct a piece of its

play46:09

input that it's being this being masked

play46:12

masked out essentially you can think of

play46:14

it this way right so show a piece of a

play46:17

video to a machine and ask it to predict

play46:19

what's gonna happen next and of course

play46:21

after a while you can show what what

play46:23

happens and the machine will kind of

play46:25

train itself to do better at that task

play46:27

you can do like all the latest most

play46:31

successful models the natural language

play46:32

processing use cell supervised running

play46:35

you know sort of bird style systems for

play46:38

example right you show it a window of a

play46:40

thousand words on a test corpus you take

play46:44

out 15% of the words and then you train

play46:47

a machine to predict the words that are

play46:50

missing that's out supervised running

play46:52

it's not predicting the future it's just

play46:54

you know predicting things in middle but

play46:56

you could have you predict the future

play46:57

that's what language models do so you

play46:59

construct it so in an unsupervised way

play47:01

you construct a model of language do you

play47:04

think or video or the physical world or

play47:07

whatever right how far do you think that

play47:10

can take us do you think very far it

play47:14

understands anything to some level it

play47:18

has you know a shallow understanding of

play47:22

of text but it needs to I mean to have

play47:25

kind of true human level intelligence I

play47:26

think you need to ground language in

play47:28

reality so some people are attempting to

play47:32

do this right having systems that can I

play47:34

have some visual representation of what

play47:35

what is being talked about which is one

play47:37

reason you need interactive environments

play47:39

actually this is like a huge technical

play47:43

problem that is not solved and that

play47:45

explains why such super versioning works

play47:47

in the context of natural language that

play47:50

does not work in the context on at least

play47:52

not well in the context of image

play47:53

recognition and video although it's

play47:55

making progress quickly and the reason

play47:58

that reason is the fact that it's much

play48:02

easier to represent uncertainty in the

play48:04

prediction you know context of natural

play48:06

language than it is in the context of

play48:08

things like video and images so for

play48:10

example if I ask you to predict what

play48:13

words are missing you know 15 percent of

play48:14

the words that I've taken out the

play48:17

possibility is small that means small

play48:19

right there is 100,000 words in the in

play48:22

the lexicon and what the Machine spits

play48:24

out is a big probability vector right

play48:27

it's a bunch of numbers between 0 & 1

play48:29

that's 1 to 1 and we know how to do how

play48:31

to do this with computers so they are

play48:34

representing uncertainty in the

play48:36

prediction is relatively easy and that's

play48:38

in my opinion why those techniques work

play48:40

for NLP for images if you ask if you

play48:45

block a piece of an image and you as a

play48:47

system reconstruct that piece of the

play48:48

image

play48:49

there are many possible answers there

play48:51

are all perfectly legit right and how do

play48:55

you represent that the set of possible

play48:58

answers

play48:58

you can't train a system to make one

play49:00

prediction you can train a neural net to

play49:01

say here it is that's the image because

play49:04

it's there's a whole set of things that

play49:06

are compatible with it so how do you get

play49:07

the machine to represent not a single

play49:09

output but all set of outputs and you

play49:14

know similarly with video prediction

play49:16

there's a lot of things that can happen

play49:18

in the future video you're looking at me

play49:20

right now I'm not moving my head very

play49:22

much but you know I might you know what

play49:24

turn my my head to the left or to the

play49:25

right right if you don't have a system

play49:27

that can predict this and you train it

play49:31

with least Square to kind of minimize

play49:32

the error with the prediction and what

play49:33

I'm doing

play49:34

what you get is a blurry image of myself

play49:36

in all possible future positions that I

play49:38

might be in which is not a good

play49:40

prediction but so there might be other

play49:42

ways to do the self supervision right

play49:45

for visual scenes like what if i I mean

play49:50

if I knew I wouldn't tell you

play49:52

publish it first I don't know I know

play49:55

there might be so I mean these are kind

play49:58

of there might be artificial ways of

play50:01

like self play in games the way you can

play50:03

simulate part of the environment you can

play50:05

oh that doesn't solve the problem it's

play50:06

just a way of generating data but

play50:10

because you have more of a country might

play50:12

mean you can control yeah it's a way to

play50:15

generate data and that's right and

play50:16

because you can do huge amounts of data

play50:18

generation that doesn't you write this

play50:21

well it's it's a creeps up on the

play50:23

problem from the side of data and you

play50:26

don't think that's the right way to it

play50:27

doesn't solve this problem of handling

play50:29

uncertainty in the world right so if you

play50:31

if you have a machine learn a predictive

play50:34

model of the world in a game that is

play50:37

deterministic or quasi deterministic

play50:39

it's easy right just you know give a few

play50:43

frames of the game to a combat put a

play50:46

bunch of layers and then half the game

play50:47

generates the next few frames and and if

play50:50

the game is deterministic it works fine

play50:53

and that includes you know feeding the

play50:58

system with the action that your little

play51:00

character is going to take

play51:02

the problem comes from the fact that the

play51:06

real world and certain most games are

play51:08

not entirely predictable that's what

play51:09

they're you get those blurry predictions

play51:11

and you can't do planning with very

play51:12

predictions all right so if you have a

play51:15

perfect model of the world you can in

play51:18

your head run this model with a

play51:21

hypothesis for a sequence of actions and

play51:24

you're going to predict the outcome of

play51:25

that sequence of actions but if your

play51:28

model is imperfect how can you plan yeah

play51:32

it quickly explodes what are your

play51:35

thoughts on the extension of this which

play51:37

topic I'm super excited about it's

play51:39

connected to something you're talking

play51:41

about in terms of robotics is active

play51:43

learning so as opposed to sort of

play51:46

unemployed and supervisors self

play51:48

supervised learning you ask the system

play51:52

for human help right for selecting parts

play51:56

you want annotated next so if you talk

play51:58

about a robot exploring a space or a

play52:00

baby exploring a space or a system

play52:03

exploring a data set every once in a

play52:06

while asking for human input you see

play52:08

value in that kind of work I don't see

play52:12

transformative value it's going to make

play52:15

things that we can already do more

play52:18

efficient or they will learn slightly

play52:19

more efficiently but it's not going to

play52:21

make machines sort of significantly more

play52:23

intelligent I think and I and by the way

play52:26

there is no opposition there is no

play52:31

conflict between self supervisor on

play52:34

reinforcement learning and supervisor on

play52:35

your imitation learning or active

play52:37

learning

play52:37

I see sub super wrestling as a as a

play52:41

preliminary to all of the above yes so

play52:45

the example I use very often is how is

play52:48

it that so if you use

play52:53

enforcement running deep enforcement

play52:55

running if you want the best methods

play52:58

today was so-called model free

play53:02

enforcement training to learn to play

play53:03

Atari games take about 80 hours of

play53:06

training to reach the level that any

play53:08

human can reach in about 15 minutes they

play53:12

get better than humans but it takes a

play53:14

long time alpha star okay the you know

play53:21

are your videos and his team's the

play53:23

system to play to to play Starcraft

play53:27

plays you know a single map a single

play53:30

type of player and

play53:34

which

play53:36

better than human level is about the

play53:40

equivalent of 200 years of training

play53:42

playing against itself it's 200 years

play53:45

right it's not something that no no

play53:47

human can could every I'm not sure what

play53:50

it doesn't take away from that okay now

play53:53

take those algorithms the best our

play53:55

algorithms we have today to train a car

play53:59

to drive itself it would probably have

play54:01

to drive millions of hours you will have

play54:04

to kill thousands of pedestrians it will

play54:05

have to run into thousands of trees it

play54:07

will have to run off cliffs and you had

play54:10

to run the cliff multiple times before

play54:11

it figures out it's a bad idea first of

play54:14

all

play54:14

yeah and second of all the figures that

play54:16

had not to do it and so I mean this type

play54:19

of running obviously does not reflect

play54:21

the kind of running that animals and

play54:23

humans do there is something missing

play54:25

that's really really important there and

play54:26

my apart is is which have been

play54:29

advocating for like five years now is

play54:31

that we have predictive models of the

play54:34

world that include the ability to

play54:37

predict under uncertainty and what

play54:40

allows us to not run off a cliff when we

play54:44

learn to drive most of us can learn to

play54:46

drive in about 20 or 30 hours of

play54:48

training without ever crashing causing

play54:50

any accident if we drive next to a cliff

play54:53

we know that if we turn the wheel to the

play54:55

right the car is going to run off the

play54:57

cliff and nothing good is gonna come out

play54:59

of this because we have a pretty good

play55:00

model of intuitive physics that tells us

play55:02

you know the car is gonna fall we know

play55:03

we know about gravity babies run this

play55:05

around the age of eight or nine months

play55:07

that objects don't float they fall and

play55:12

you know we have a pretty good idea of

play55:14

the effect of turning the wheel of the

play55:15

car and you know we know we need to stay

play55:17

on the road so there is a lot of things

play55:18

that we bring to the table which is

play55:20

basically or predictive model of the

play55:22

world and that model allows us to not do

play55:26

stupid things and to basically stay

play55:28

within the context of things we need to

play55:30

do we still face you know unpredictable

play55:33

situations and that's how we learn but

play55:35

that allows us to learn really really

play55:38

really quickly so that's called

play55:39

model-based reinforcement running

play55:42

there's some imitation and supervised

play55:43

running because we have a driving

play55:45

instructor that tells us occasionally

play55:47

what to do but most of the learning is

play55:50

Mauro bass is learning the model yeah

play55:53

running physics that we've done since we

play55:55

were babies that's where all almost all

play55:57

are learning and the physics is somewhat

play55:59

transferable from is transferable from

play56:02

scene to scene stupid things are the

play56:04

same everywhere yeah I mean if you you

play56:06

know you have experience of the world

play56:08

you don't need to be particularly from a

play56:11

particularly intelligent species to know

play56:12

that if you spill water from a container

play56:16

you know the rest is gonna get wet and

play56:20

you might get wet so you know cats know

play56:23

this right yeah so the main problem we

play56:26

need to solve is how do we learn models

play56:29

of the world that's and that's what I'm

play56:30

interesting that's what's a supervised

play56:32

learning is all about if you were to try

play56:34

to construct a benchmark for let's let's

play56:39

look at happiness I'd love that dataset

play56:41

but if you do you think it's useful

play56:45

interesting / possible to perform well

play56:50

on eminence with just one example of

play56:52

each digit and how would we solve that

play56:56

problem

play56:58

yeah so it's probably yes the question

play56:59

is what other type of running are you

play57:02

allowed to do so if what you like to do

play57:04

is train on some gigantic data set of

play57:06

labelled digit that's called transfer

play57:07

running and we know that works okay

play57:11

we do this at Facebook like in

play57:12

production right we we train large

play57:15

commercial nets to predict hashtags that

play57:17

people type on Instagram and we train on

play57:18

billions of images literally billions

play57:20

and and then we chop off the last layer

play57:22

and fine-tune on whatever task we want

play57:24

that works really well you can be you

play57:27

know the image net record with we

play57:28

actually open source the whole thing

play57:30

like a few weeks ago yeah that's still

play57:32

pretty cool but yeah so what in yet

play57:34

won't be impressive and what's useful an

play57:37

impressive what kind of transfer

play57:38

learning would be useful impressive is

play57:40

it Wikipedia that kind of thing no no I

play57:42

don't think transfer learning is really

play57:45

where we should focus we should try to

play57:47

do

play57:49

you know have a kind of scenario for

play57:51

benchmark where you have only ball data

play57:54

and you can and it's very large number

play57:58

of enabled data it could be video clips

play58:01

it could be what you do you know frame

play58:04

prediction it could be images you could

play58:06

choose to you know mask a piece of it it

play58:11

could be whatever but they're only bold

play58:13

and you're not allowed to label them so

play58:15

you do some training on this and then

play58:19

you train on a particular supervised

play58:23

task imagenet or nist and you measure

play58:27

how your test our decrease or variation

play58:31

error decreases as you increase the

play58:32

number of label training samples okay

play58:36

and and what what you would like to see

play58:40

is is that you know your your error

play58:43

decreases much faster than if you

play58:45

trained from scratch from random weights

play58:47

so that to reach the same level of

play58:49

performance and a completely supervised

play58:51

purely supervised system would reach you

play58:54

would need way fewer samples so that's

play58:56

the crucial question because it will

play58:58

answer the question to like you know

play59:00

people are interested in medical image

play59:01

analysis okay you know if I want to get

play59:03

to a particular level of error rate for

play59:07

this task I know I need a million

play59:11

samples can I do you know soft

play59:14

supervised pre-training to reduce this

play59:15

to about 100 or something anything the

play59:18

answer there is soft supervised

play59:19

retraining yep some form some form of it

play59:24

telling you active learning but you

play59:27

disagree you know it's not useless it's

play59:30

just not gonna lead to a quantum leap

play59:32

it's just gonna make things that we

play59:33

already do so you're way smarter than me

play59:35

I just disagree with you but I don't

play59:37

have anything to back that it's just

play59:39

intuition so I've worked a lot of

play59:41

large-scale data sets and there's

play59:43

something there might be magic and

play59:45

active learning but okay at least I said

play59:48

it publicly at least some being an idea

play59:51

publicly okay it's not bigoted yet it's

play59:54

you know working with the data you have

play59:56

I mean I mean certainly people are doing

play59:57

things like okay I have three thousand

play59:59

hours of you know imitation running for

play60:02

in car but most of those are incredibly

play60:04

boring what I like is select you know

play60:07

10% of them that are kind of the most

play60:08

informative and with just that I would

play60:10

probably reach the same so it's a weak

play60:13

form of of active running if you want

play60:16

yes but there might be a much stronger

play60:19

version yeah that's right that's what

play60:21

another notion question is the question

play60:24

is how much talking yet Elon Musk is

play60:27

confident talk to him recently

play60:30

he's confident that large-scale data and

play60:32

deep learning can solve the autonomous

play60:33

driving problem what are your thoughts

play60:36

on the limitless possibilities of deep

play60:39

learning in this space I was it's

play60:40

obviously part of the solution I mean I

play60:43

don't think we'll ever have a set

play60:44

driving system or it is not in the

play60:46

foreseeable future that does not use

play60:48

deep running you put it this way now how

play60:51

much of it so in the history of sort of

play60:54

engineering particularly is sort of sort

play60:58

of a I like systems is generally your

play61:01

first phase where everything is built by

play61:02

hand and it was the second phase and

play61:04

that was the case for autonomous driving

play61:06

you know 23 years ago there's a phase

play61:08

where this a little bit of running is

play61:10

used but there's a lot of engineering

play61:12

that's involved in kind of you know

play61:14

taking care of corner cases and and

play61:16

putting limits etc because the learning

play61:19

system is not perfect and then I as

play61:21

technology progresses we end up relying

play61:25

more and more on learning that's the

play61:26

history of character recognition is a

play61:27

history of speech recognition now

play61:29

computer vision that ronnie was

play61:30

processing and I think the same is going

play61:32

to happen with with the term is driving

play61:35

that currently the the the methods that

play61:39

are closest to providing some level of

play61:42

autonomy some you know a decent level of

play61:44

autonomy where you don't expect a driver

play61:46

to kind of do anything is where you

play61:49

constrain the world so you only run

play61:51

within you know 100 square kilometers or

play61:53

square miles in Phoenix but the weather

play61:55

is nice and the roads are wide it wishes

play61:58

what Weimer is doing you completely over

play62:01

engineer the car with tons of light

play62:04

hours and sophisticated sensors that are

play62:07

too expensive for consumer cars but

play62:09

they're fine if you just run a fleet

play62:12

and you engineer the thing the hell out

play62:15

of the everything else you you map the

play62:17

entire world so you have complete 3d

play62:19

model of everything so the only thing

play62:21

that the perception system has to take

play62:22

care of is moving objects and and and

play62:25

construction and sort of you know things

play62:27

that that weren't in your map and you

play62:31

can engineer a good you know slam system

play62:32

or eye stuff right so so that's kind of

play62:35

the current approach that's closest to

play62:36

some level of autonomy but I think

play62:37

eventually the long term solution is

play62:39

going to rely more and more on learning

play62:43

and possibly using a combination of

play62:45

supervised learning and model-based

play62:48

reinforcement or something like that but

play62:50

ultimately learning will be at not just

play62:54

at the core but really the fundamental

play62:55

part of the system yeah it already is

play62:58

but it'll become more and more what do

play63:00

you think it takes to build a system

play63:02

with human level intelligence you talked

play63:04

about the AI system and then we her

play63:06

being way out of reach our current reach

play63:09

this might be outdated as well but this

play63:12

is still way out of reach what would it

play63:16

take to build her do you think so I can

play63:20

tell you the first two obstacles that we

play63:21

have to clear but I don't know how many

play63:23

obstacles they are after this so the

play63:25

image I usually use is that there is a

play63:26

bunch of mountains that we have to climb

play63:28

and we can see the first one but we

play63:29

don't know if there are 50 mountains

play63:31

behind it or not and this might be a

play63:33

good sort of metaphor for why AI

play63:37

researchers in the past I've been overly

play63:38

optimistic about the result of AI you

play63:43

know for example New Orleans Simon

play63:46

Wright wrote the general problem solver

play63:48

and they call it the general problems

play63:51

you have problems okay and of course if

play63:53

it's you realize is that all the

play63:54

problems you want to solve is financial

play63:56

and so you can't actually use it for

play63:57

anything useful but you know yes oh yeah

play64:00

all you see is the first peak so in

play64:02

general what are the first couple of

play64:03

peaks for her so the first peak which is

play64:06

precisely what I'm working on is self

play64:09

supervisor running high how do we get

play64:10

machines to learn models of the world by

play64:12

observation kind of like babies and like

play64:14

young animals

play64:17

so I we've been working with you know

play64:21

cognitive scientists so this Amanda

play64:24

depuis who is at fair and in Paris is

play64:28

half-time is also a researcher and

play64:31

French University and he he has his

play64:35

chart that shows that which how many

play64:38

months of life baby humans kind of

play64:40

learned different concepts and you can

play64:43

met you can measure this various ways so

play64:47

things like distinguishing animate

play64:50

objects from animate inanimate object

play64:52

you can you can tell the difference at

play64:54

age to three months whether an object is

play64:58

going to stay stable is gonna fall you

play64:59

know about four months you can tell you

play65:03

know things like this and then things

play65:06

like gravity the fact that objects are

play65:07

not supposed to float in the air but as

play65:09

opposed to fall you run this around the

play65:11

age of eight or nine months if you look

play65:14

at a lot of you know eight month old

play65:15

babies you give them a bunch of toys on

play65:18

the highchair first thing they do is

play65:19

it's why I'm on the ground that you look

play65:20

at them it's because you know they're

play65:22

learning about actively learning about

play65:25

gravity gravity yeah okay so they're not

play65:28

trying to know you but they you know

play65:30

they need to do the experiment right

play65:32

yeah so you know how do we get machines

play65:34

to learn like babies mostly by

play65:36

observation with a little bit of

play65:38

interaction and learning those those

play65:40

those models of the world because I

play65:41

think that's really a crucial piece of

play65:43

an intelligent autonomous system so if

play65:46

you think about the architecture of an

play65:47

intelligent autonomous system it needs

play65:49

to have a predictive model of the world

play65:51

so something that says here is a wall

play65:53

that time T here is a stable world at

play65:54

time T plus one if I take this action

play65:56

and it's not a single answer it can be

play65:59

education yeah yeah well but we don't

play66:02

know how to represent distributions in

play66:03

high dimension continuous basis so it's

play66:05

got to be something we care that data

play66:06

Hey but with some summer presentation

play66:08

with certainty if you have that then you

play66:12

can do what optimal control theory is

play66:13

called model predictive control which

play66:15

means that you can run your model with

play66:17

the hypothesis for a sequence of action

play66:19

and then see the result now what you

play66:22

need the other thing you need is some

play66:23

sort of objective that you want to

play66:25

optimize am i reaching the goal of

play66:27

grabbing the subject about minimizing

play66:29

energy am I

play66:30

whatever right so there is some sort of

play66:32

objectives that you have to minimize and

play66:34

so in your head if you had this model

play66:36

you can figure out the sequence of

play66:37

action that will optimize your objective

play66:40

that objective is something that

play66:42

ultimately is rooted in your basal

play66:45

ganglia at least in the human brain

play66:46

that's that's what is available Gambia

play66:48

computes your level of contentment or

play66:50

miss contentment oh no noise that's a

play66:53

word unhappiness okay yeah this

play66:56

contentment this contentment and so your

play66:59

entire behavior is driven towards kind

play67:03

of minimizing that objective which is

play67:05

maximizing your contentment computed by

play67:07

your your basal ganglia and what you

play67:11

have is an objective function which is

play67:13

basically a predictor of what your basal

play67:15

ganglia is going to tell you so you're

play67:17

not going to put your hand on fire

play67:18

because you know it's gonna you know

play67:21

it's gonna burn and you're gonna get

play67:23

hurt and you're predicting this because

play67:24

of your model of the world and your your

play67:27

predictor of this objective right so you

play67:30

if you have those you have those three

play67:32

components you have four components you

play67:35

have the the hard-wired contentment

play67:38

objective good computer if you want

play67:42

calculator and then you have the three

play67:44

components one is the objective

play67:46

predictor which basically predicts your

play67:47

level of contact and one is the model of

play67:51

the world and there's a third module I

play67:53

didn't mention which is a module that

play67:55

will figure out the best course of

play67:58

action to optimize an objective given

play68:00

your model okay yeah cool it's a policy

play68:06

policy network or something like that

play68:07

right now you need those three

play68:11

components to act autonomously

play68:12

intelligently and you can be stupid in

play68:14

three different ways you can be stupid

play68:16

because your model of the world is wrong

play68:18

you can be stupid because your objective

play68:20

is not aligned with what you actually

play68:23

want to achieve okay and in humans that

play68:27

would be a psychopath right and then the

play68:30

the third thing you the third way you

play68:32

can be stupid is that you have the right

play68:34

model you have the right objective but

play68:36

you're unable to figure out a course of

play68:38

action to optimize your objective given

play68:40

your model

play68:43

some people who are in charge of big

play68:45

countries actually have all three that

play68:46

are wrong all right which countries I

play68:51

don't know okay so if we think about

play68:54

this this agent if you think about the

play68:56

movie her you've criticized the art

play69:02

project that is Sophia the robot and

play69:04

what that project essentially does is

play69:07

uses our natural inclination to

play69:09

anthropomorphize things that look like

play69:12

human and given more do you think that

play69:15

could be used by AI systems like in the

play69:18

movie her

play69:20

so do you think that body is needed to

play69:23

create a feeling of intelligence

play69:26

well if Sophia was just an art piece I

play69:29

would have no problem with it but it's

play69:30

presented as something else let me add

play69:33

that comics real quick if creators of

play69:36

Sofia could change something about their

play69:39

marketing or behavior in general what

play69:40

would it be what what's just about

play69:43

everything I mean don't you think here's

play69:48

a tough question I mean so I agree with

play69:51

you so Sofia is not in the general

play69:54

public feels that Sofia can do way more

play69:57

than she actually can that's right and

play70:00

the people will create a Sofia are not

play70:04

honestly publicly communicating trying

play70:08

to teach the public right but here's a

play70:11

tough question don't you think this the

play70:16

same thing is scientists in industry and

play70:21

research are taking advantage of the

play70:23

sameness misunderstanding in the public

play70:25

when they create AI companies or

play70:27

published stuff some companies yes I

play70:30

mean there is no sense of there's no

play70:33

desire to delude there's no desire to

play70:35

kind of over claim what something is

play70:38

done right you know you should paper on

play70:39

AI that you know has this result on

play70:41

image net you know it's pretty clear I

play70:43

mean it's not even not even interesting

play70:44

anymore but you know I I don't think

play70:47

there is that I mean the reviewers are

play70:49

generally not very forgiving of of you

play70:54

know unsupported claims of this type and

play70:56

but there are certainly quite a few

play70:59

startups that have had a huge amount of

play71:01

hype around this that I find extremely

play71:04

damaging and I've been calling it out

play71:06

when I've seen it so yeah but to go back

play71:09

to your original question like the

play71:11

necessity of embodiment I think I don't

play71:14

think embodiment is necessary I think

play71:15

grounding is necessary so I don't think

play71:18

we're gonna get machines that I really

play71:19

understand language without some level

play71:21

of grounding in the world world and it's

play71:22

not clear to me that language is a kind

play71:24

of bandwidth medium to communicate how

play71:26

the real world works I think what this

play71:30

doctor ground our grounding means so

play71:32

running me he's that

play71:34

so there is this classic problem of

play71:35

common sense reasoning you know the the

play71:38

Winograd Winograd schema right and so I

play71:41

tell you the the trophy doesn't fit in

play71:44

the suitcase because this tool is too

play71:45

big what the trophy doesn't fit in the

play71:47

suitcase because it's too small and the

play71:49

it in the first case refers to the

play71:51

trophy in the second case to the

play71:52

suitcase and the reason you can figure

play71:54

this out is because you know what the

play71:56

trophy in the suitcase are you know one

play71:57

is supposed to fit in the other one and

play71:58

you know the notion of size and the big

play72:01

object doesn't fit in a small object and

play72:03

this is a TARDIS you know it things like

play72:04

that right so you have this got this

play72:06

knowledge of how the world works of

play72:08

geometry and things like that I don't

play72:12

believe you can learn everything about

play72:14

the world by just being told in language

play72:16

how the world works I think you need

play72:18

some low-level perception of the world

play72:21

you know be a visual touch you know

play72:23

whatever but some higher bandwidth

play72:25

perceptions of the world but by reading

play72:27

all the world's text you still may not

play72:29

have enough information that's right

play72:31

there's a lot of things that just will

play72:34

never appear in text and that you can't

play72:35

really infer so I think common sense

play72:39

will emerge from you know certainly a

play72:42

lot of language interaction but also

play72:43

with watching videos or perhaps even

play72:46

interacting in the in virtual

play72:47

environments and possibly you know robot

play72:50

interacting in the real world but I

play72:52

don't actually believe necessarily that

play72:53

this last one is absolutely necessary

play72:55

but I think there's a need for some

play72:57

grounding but the final product doesn't

play73:02

necessarily need to be embodied you know

play73:04

who say no it just needs to have an

play73:05

awareness a grounding right but it needs

play73:08

to know how the world works to have you

play73:12

know to not be frustrated frustrating to

play73:13

talk to and you talked about emotions

play73:17

being important that's that's a whole

play73:20

nother topic well so you know I talked

play73:23

about this the the basal ganglia ganglia

play73:26

as the you know this thing that could

play73:30

you know calculates your level of miss

play73:31

contentment contentment and then there

play73:33

is this other module that sort of tries

play73:35

to do a prediction of whether you're

play73:37

going to be content or not that's the

play73:38

source of some emotion so here for

play73:40

example is an anticipation of bad things

play73:44

that can happen to you right

play73:47

you have this inkling that there is some

play73:48

chance that something really bad is

play73:50

gonna happen to you and that creates

play73:51

here when you know for sure that

play73:53

something bad is gonna happen to you you

play73:54

cannot give up right it's not bad

play73:56

anymore it's uncertainty it creates fear

play73:58

so so the punchline is yes we're not

play74:01

gonna have a ton of intelligence without

play74:02

emotions whatever the heck emotions are

play74:08

so you mentioned very practical things

play74:10

of fear but there's a lot of other mess

play74:12

around but there are kind of the results

play74:14

of you know drives

play74:16

yeah there's deeper biological stuff

play74:18

going on and I've talked a few folks on

play74:21

this there's a fascinating stuff that

play74:23

ultimately connects to our joy to our

play74:26

brain if we create an AGI system sorry

play74:30

interminable human level intelligence

play74:32

system and you get to ask her one

play74:35

question what would that question be you

play74:39

know I think the the first one we'll

play74:41

create would probably not be that smart

play74:44

did you like a four-year-old okay so you

play74:47

would have to ask her a question - no

play74:50

she's not that smart

play74:51

yeah well what's a good question to ask

play74:56

you know to be responsive wind and if

play75:01

she answers oh it's because the leaves

play75:03

of the tree are moving in that creates

play75:05

wind she's on to something and if she

play75:09

says yeah that's a stupid question

play75:10

she's really obtuse no and then you tell

play75:13

her actually you know here is the the

play75:16

real thing and she says oh yeah that

play75:19

makes sense

play75:19

so questions that that reveal the

play75:24

ability to do common-sense reasoning

play75:25

about the physical world yeah and you

play75:27

know someone will call 20 ferns causal

play75:30

evidence well it was a huge honor

play75:33

congratulations returning award you know

play75:35

and thank you so much for talking today

play75:36

thank you

play75:45

you

Rate This

5.0 / 5 (0 votes)

Related Tags
深度学习人工智能Yann LeCun自主系统价值不对齐神经网络科技影响未来展望学术研究技术创新
Do you need a summary in English?