Ilya sutskever | Humanity will eventually move towards AGI | The intelligent body will soon appear

Me&ChatGPT
22 Jul 202475:20

Summary

TLDR在这段访谈中,Ilya Sutskever回顾了深度学习和神经网络在计算机视觉领域的重大突破。他分享了自己对于如何训练深度网络的见解,以及这些网络如何彻底改变了机器学习领域。Sutskever讨论了他在OpenAI的工作,包括开发GPT系列模型和CLIP模型,以及这些技术如何推动了AI的边界。他还探讨了AI的未来,包括对提高模型可靠性和效率的期望,以及AI如何可能改变我们的工作和生活方式。

Takeaways

  • 📈 神经网络在计算机视觉领域的突破性进展,大幅超越了以往的方法。
  • 🤖 深度学习的成功部分归功于James Martin等人的研究,他们证明了可以从监督数据中端到端训练深度网络。
  • 🧠 人类视觉的快速反应意味着不需要太多层次的神经网络就能实现可观的视觉效果。
  • 🚀 随着大型数据集和GPU的兴起,为训练大型神经网络提供了可能,从而推动了深度学习的发展。
  • 🌐 深度学习的成功也改变了人们对机器学习模型的看法,从追求简单优雅的数学证明转向接受更强大的模型。
  • 🔑 神经网络被视为一种编程工具,通过反向传播算法对这些“小计算机”进行编程。
  • 🔑 神经网络的并行计算特性使其能够处理复杂的模式识别任务,如围棋和机器翻译。
  • 🌟 深度学习在图像识别和语言处理等领域的成功,展示了其在解决人类能够快速解决的问题上的巨大潜力。
  • 💡 深度学习的发展推动了对神经网络更深层次理解的探索,包括其在强化学习、自然语言处理和计算机视觉等领域的应用。
  • 🔮 未来的AI研究可能会继续探索如何使神经网络更加可靠、高效,并在更广泛的任务中发挥作用。

Q & A

  • 神经网络是如何在计算机视觉领域取得突破的?

    -神经网络通过端到端的训练方法,利用大量标记数据进行学习,从而在计算机视觉领域取得了突破。这种训练方式最初由Hinton等人提出,并在ImageNet竞赛中得到验证,神经网络的表现大幅超越了以往的方法。

  • 为什么深度学习在早期没有被广泛接受?

    -早期深度学习没有被广泛接受的原因是存在一些普遍的误解,比如认为深度网络无法训练,因为梯度消失或梯度爆炸问题。此外,当时普遍认为需要复杂的数学证明来保证机器学习模型的有效性,而神经网络则没有这样的证明。

  • 人类视觉的快速识别能力对神经网络设计有何启示?

    -人类视觉系统能够在几百毫秒内识别物体,而我们的神经元反应速度相对较慢。这说明不需要太多的层次就能实现有效的视觉识别,这对神经网络设计提供了启示,即可以通过构建较大但不需要太多层的网络来实现高效的计算机视觉。

  • 为什么选择使用GPU来训练神经网络?

    -GPU提供了并行处理大量数据的能力,这对于训练大型神经网络至关重要。在Alex Krizhevsky开发出能够在GPU上高效训练的卷积神经网络(CNN)之后,这种方法开始变得可行,因为它大大缩短了训练时间,使得在ImageNet等大规模数据集上训练成为可能。

  • 神经网络在语言处理方面的应用是如何实现的?

    -神经网络在语言处理方面的应用是通过将语言视为一种可以通过模式识别来处理的连续信号来实现的。通过训练神经网络来预测文本序列中的下一个词,网络能够学习语言的结构和语义,从而在机器翻译等任务中取得显著效果。

  • 为什么选择DOTA作为强化学习的研究项目?

    -DOTA是一个复杂的实时策略游戏,具有高度的挑战性,它要求玩家具有快速反应、战略思维和直觉。选择DOTA作为研究项目是为了测试和推动强化学习技术的极限,特别是在处理多变量、多步骤决策和长期规划方面。

  • GPT模型是如何通过预测来实现语言理解的?

    -GPT模型通过预测文本序列中的下一个词来实现语言理解。如果模型能够准确预测接下来的单词,这意味着它已经理解了前面的文本内容。随着预测精度的提高,模型能够捕捉到更复杂的语言特征,如词汇、语法和语义。

  • 什么是CLIP模型,它如何将语言和视觉结合起来?

    -CLIP模型是一个多模态神经网络,它通过大规模的自然语言监督学习视觉概念。CLIP能够将文本描述与图像内容关联起来,从而实现对图像的理解。这种结合语言和视觉的方法使得模型能够在没有大量标注数据的情况下,学习丰富的视觉概念。

  • 为什么说神经网络是通用的计算设备?

    -神经网络被视作通用的计算设备,因为它们可以接受任何形式的输入,通过学习内部表示来处理复杂的任务。无论是视觉识别、语言理解还是策略游戏,神经网络都可以通过适当的训练来适应这些任务,显示出它们的通用性和灵活性。

  • 如何提高神经网络的可靠性和可控性?

    -提高神经网络的可靠性和可控性可以通过多种方式实现,包括使用更大的数据集进行训练、设计更复杂的网络结构、以及通过强化学习从人类反馈中学习。此外,还可以通过精心设计的提示(prompting)来引导模型产生期望的输出。

Outlines

00:00

🤖 神经网络在计算机视觉领域的突破

在这段对话中,讨论了神经网络如何在计算机视觉领域取得突破性进展,超越了以往的所有方法。关键的转折点是James Martin的一篇论文,首次展示了可以通过监督数据端到端训练深度网络。这一发现颠覆了当时普遍的看法,即深度网络无法训练。此外,还讨论了神经网络的潜力,它们本质上是小型并行计算机,可以通过反向传播算法进行编程。这一认识促使人们开始探索神经网络在计算机视觉等领域的应用,最终导致了在ImageNet竞赛中的突破。

05:00

🧠 神经网络的快速发展和硬件的利用

这段对话讲述了神经网络如何迅速发展并利用当时的硬件。提到了人类视觉的快速反应和神经元的慢速特性,指出不需要太多层次就能实现有效的视觉识别。随着大型数据集的出现和GPU的普及,神经网络迎来了发展的黄金时期。特别是Alex Krizhevsky在ImageNet上使用GPU训练小型卷积网络取得显著成果,这进一步推动了神经网络在计算机视觉领域的应用。

10:01

🌐 神经网络在语言处理和游戏领域的应用

在这部分对话中,讨论了神经网络在语言处理和游戏领域的应用。提到了神经网络在语言翻译和围棋等游戏中的潜力。强调了神经网络的直觉力,它们能够像人类专家一样快速做出决策。此外,还探讨了神经网络如何通过深度学习来解决需要大量思考的问题,以及它们在语言模型和游戏策略中的应用。

15:01

🚀 神经网络的创新和未来发展

这段对话探讨了神经网络的创新过程和未来发展的可能性。讨论了如何通过不断的实验和探索来推动神经网络技术的进步。提到了在神经网络研究中,即使是简单的方法也可能带来意想不到的成果,如在DOTA游戏中的应用。同时,也强调了在神经网络研究中保持开放心态和勇于尝试的重要性。

20:04

🎓 从学术到实践:神经网络的商业化之路

在这部分对话中,讨论了神经网络从学术研究到商业应用的转变。提到了在Google的工作体验和DeepMind在AlphaGo项目中的成功,这些经历激发了对神经网络商业潜力的认识。讨论了如何将神经网络技术转化为实际产品,以及在这一过程中面临的挑战和机遇。

25:05

🤖 神经网络的自我学习和自我优化

这段对话探讨了神经网络的自我学习能力和自我优化潜力。讨论了如何通过训练神经网络来提高其在特定任务上的表现,以及如何利用神经网络的自我学习能力来解决更复杂的问题。提到了在神经网络训练过程中,如何通过调整网络结构和参数来优化其性能。

30:06

🧠 神经网络与人类智能的结合

在这部分对话中,讨论了如何将神经网络与人类智能相结合,以实现更高效的问题解决。提到了通过训练神经网络来模拟人类的认知过程,以及如何利用神经网络来增强人类的决策能力。强调了在神经网络研究中,理解和模拟人类智能的重要性。

35:07

🌐 神经网络在多模态学习中的应用

这段对话探讨了神经网络在处理多模态数据(如视觉和语言)中的应用。讨论了如何训练神经网络来理解和生成图像和文本,以及如何利用神经网络来实现跨模态的学习和理解。提到了CLIP和DALL-E等模型,它们能够将视觉和语言结合起来,实现更丰富的数据理解和生成。

40:08

🚀 神经网络的未来和对社会的影响

在这部分对话中,讨论了神经网络技术的未来发展方向,以及它可能对社会产生的深远影响。提到了随着神经网络能力的增强,它们将在各行各业中发挥越来越重要的作用。讨论了如何确保神经网络的可靠性和安全性,以及如何通过合理的政策和法规来引导神经网络技术的发展。

45:09

🎨 创造力与神经网络的结合

这段对话探讨了创造力在神经网络研究中的作用,以及如何通过结合艺术和科学来推动神经网络技术的发展。提到了个人如何在神经网络研究中保持创造力和创新精神,以及如何通过跨学科合作来探索新的研究方向。

Mindmap

Keywords

💡深度学习

深度学习是机器学习的一个分支,它使用多层神经网络来模拟人类学习的过程,从而让机器能够从数据中学习并做出决策。在视频中,深度学习是实现计算机视觉和其他人工智能应用的关键技术,如通过训练神经网络来识别图像和进行语言翻译。

💡神经网络

神经网络是深度学习的基础,它由许多相互连接的节点(或称为神经元)组成,模仿人脑的工作方式来处理信息。视频中提到,神经网络可以被训练来执行各种复杂的任务,如视觉识别和语言理解。

💡计算机视觉

计算机视觉是人工智能的一个领域,它使计算机能够理解和解释视觉信息。在视频中,计算机视觉是通过训练深度神经网络来实现的,使得机器能够识别和处理图像和视频中的内容。

💡卷积神经网络(CNN)

卷积神经网络是一种专门用于处理具有网格结构的数据(如图像)的神经网络。视频中提到,CNN在计算机视觉任务中表现出色,因为它们能够捕捉到图像的重要特征。

💡端到端学习

端到端学习是一种机器学习范式,其中模型直接从输入数据到最终输出结果进行训练,无需人为干预。视频中提到,端到端学习使得深度网络能够直接从监督数据中学习,这是深度学习成功的关键因素之一。

💡GPU

GPU(图形处理单元)是一种专门用于并行处理图形和复杂计算任务的硬件。视频中提到,GPU的计算能力对于训练大型神经网络至关重要,因为它们可以加速深度学习模型的训练过程。

💡ImageNet

ImageNet是一个大型的图像数据库,它在计算机视觉领域中用于训练和评估模型的性能。视频中提到,ImageNet挑战赛是深度学习在视觉识别任务上取得突破的一个重要里程碑。

💡Transformer

Transformer是一种深度学习模型架构,它在处理序列数据(如文本和时间序列)方面表现出色。视频中提到,Transformer模型在自然语言处理任务中取得了显著的成果,推动了语言模型的发展。

💡无监督学习

无监督学习是机器学习的一种类型,其中模型在没有标签或注释的数据上进行训练,以发现数据中的模式。视频中讨论了无监督学习的重要性,尤其是在训练大型语言模型时,如GPT系列。

💡强化学习

强化学习是机器学习的一个分支,它通过奖励和惩罚来训练模型,使其能够做出最优决策。视频中提到,强化学习在训练能够玩复杂游戏(如Dota和Go)的AI模型中发挥了关键作用。

💡多模态学习

多模态学习是指模型能够处理和理解多种类型的数据,如文本、图像和声音。视频中提到了CLIP和DALL-E模型,它们能够将语言和视觉信息结合起来,展示了多模态学习在AI领域的潜力。

Highlights

神经网络在计算机视觉领域取得了巨大突破,大幅超越了以往的方法。

深度学习的出现标志着可以从监督数据中端到端训练深度网络。

神经网络本质上是小型并行计算机,可以通过反向传播算法进行编程。

人类视觉快速,意味着不需要太多层次就能实现可观的视觉识别。

ImageNet数据集的出现和GPU的普及为训练大型神经网络提供了可能。

Alex Krizhevsky的GPU代码能够在60秒内训练小型卷积网络取得显著结果。

深度学习在计算机视觉领域的成功,引发了对其他领域应用的思考。

深度学习在语言处理领域的应用,尤其是机器翻译,取得了显著进展。

神经网络的并行计算能力使其在处理连续信号如语音和视觉信号方面表现出色。

AlphaGo的成功展示了深度学习在解决复杂策略问题上的巨大潜力。

深度学习的发展推动了AI领域从个体研究者向大规模工程项目的转变。

OpenAI的成立初衷是探索深度学习在工程领域的应用和潜力。

DOTA项目的挑战和成功展示了简单强化学习策略的有效性。

GPT系列模型的发展和成功,特别是在语言生成和理解方面。

CLIP模型的创新之处在于将自然语言和视觉理解结合在一起。

通过强化学习和人类反馈训练,提高了AI系统的可控性和可靠性。

未来AI的发展将更加注重效率和成本,以及在特定领域的专业化应用。

AI的最终目标是实现自动化生产力,使人们能够享受由AI创造的成果。

Transcripts

play00:00

a neural network beat all past

play00:02

approaches to computer vision by a very

play00:04

large margin and of course you were one

play00:06

of the people making that happen and so

play00:08

I'm really curious from your

play00:11

perspective H how did that come about um

play00:14

everybody else is working on different

play00:16

approaches to computer vision and there

play00:17

you are working on nets for computer

play00:19

vision and then you drastically

play00:21

outperform everyone um how do you even

play00:24

decide to do this yeah I'd say

play00:28

that what led me to this

play00:30

result

play00:32

was a set of

play00:36

realizations over over the time period

play00:38

of a number of years which I'll describe

play00:40

to you so I think the first really

play00:42

pivotal pivotal moment was

play00:45

when James Martins has written a paper

play00:49

called Deep learning by H in free

play00:51

optimization and that was the first time

play00:53

anyone has shown that you can train deep

play00:56

networks end to end from supervised data

play00:59

MH but for some context back in those

play01:01

days everybody knew that you cannot

play01:05

train deep

play01:06

networks it cannot be done back

play01:08

propagation is too weak you need to do

play01:10

some kind of pre-training of some sort

play01:11

and then maybe you'll get some kind of

play01:12

an NPH but if it is the case that you

play01:16

can train them end to

play01:19

end then what can they do and the thing

play01:22

is you know

play01:23

why there is one more piece of context

play01:26

that's really important so today we take

play01:28

deep learning for granted of course a

play01:31

large neural network is what you need

play01:33

and you get you you sh sh you shove data

play01:36

into it and you'll get amazing result

play01:37

everyone knows that every child knows

play01:39

that how can it be how can it be that we

play01:42

did not know that how could such an

play01:44

obvious thing was not known

play01:46

well people were really focused on

play01:51

machine learning models where they can

play01:54

prove that there is an algorithm which

play01:56

can perfectly train them but whenever

play01:59

you put this condition on yourself and

play02:01

you require to find a simple elegant

play02:03

mathematical proof you really end up

play02:05

restricting the power of your model in

play02:08

contrast neural networks like the

play02:11

fundamental thing about neural networks

play02:12

is that they are basically little

play02:14

computers little parall computers that

play02:16

are no longer so little anymore that

play02:19

definitely are they can be as little or

play02:21

as large as you want but basically it is

play02:23

a computer it is a parallel computer and

play02:25

when you train a neural network you

play02:27

program this computer with a back

play02:28

propagation algorithm

play02:30

and so the the thing that really clicked

play02:32

for me is when I saw this these results

play02:34

with the H and fre optimized I realized

play02:36

wait a second so we can actually program

play02:38

those things now it's no longer the case

play02:40

that you know maybe you could so the

play02:43

prevailing view was aspirationally maybe

play02:47

someone could train those things but

play02:48

it's obviously impossible local minimas

play02:50

will get you but no you can train a

play02:52

neural net then the second realization

play02:54

is human vision is fast it takes several

play02:58

hundred milliseconds at most to

play02:59

recognize something and yet our neurons

play03:00

are slow so that means that you don't

play03:02

even need that many layers to get

play03:04

respectable Vision so you put this so

play03:08

what does that mean it means that if you

play03:09

have a neural network which is pretty

play03:10

large then there exist some

play03:14

parameters which achieve good results on

play03:17

Vision now if only there was a data set

play03:20

which we could train from and then imet

play03:21

came up and then the gpus came up and

play03:24

then I was this has to happen and then

play03:28

at some point I had a with Alex kki

play03:30

where he said that he has GPU code which

play03:34

can train a small conet to get

play03:36

respectable results on Sear in 60

play03:38

seconds and I was like oh my God so let

play03:41

let's let's do this on imet it's gonna

play03:43

it's gonna Crush everything and that's

play03:44

how it happen that's how it came to

play03:47

be I I love the backstory here Ilia and

play03:50

how it reminds me a lot of our days at

play03:52

open ey where many things to you just

play03:55

look unavoidable and and just so clearly

play03:59

that they have to be that way I remember

play04:00

the first time you you articulated to me

play04:03

um that an net is just a computer

play04:06

program um and this is like several

play04:09

years before even karpathy started

play04:11

talking about software 2.0 being you

play04:14

know programming with neural Nets and

play04:16

it's just parallel and serial compute

play04:18

it's it's really it's really amazing

play04:20

that you saw this even before there was

play04:22

real success in neur Nets um when did

play04:27

you realize it was actually working on

play04:28

image net what was was that like I mean

play04:31

I I had very I had very little doubt

play04:33

that it would work but it was kind of

play04:35

you know at this point you know Alex was

play04:39

training the neural net and the results

play04:41

were getting better week after

play04:44

week and that's about it but I felt but

play04:47

I felt like the big risk for my

play04:49

perspective was can we have can we have

play04:53

that do we have the ability to utilize

play04:55

the gpus well enough MH train a big

play04:58

enough you know big enough there's no

play05:00

such thing it's more like an

play05:01

interestingly large neural network it

play05:03

has to be a neural network that is large

play05:05

enough to be interesting whereas all

play05:07

previous neural networks are small if

play05:09

you're just going to have something

play05:10

which is going to be way larger than

play05:12

anything before then it should do much

play05:16

better than anything anyone's ever seen

play05:18

of course we are far beyond that our

play05:20

computers are faster and your networks

play05:21

are larger but the goal was not the goal

play05:24

was just to go as far as possible with

play05:26

the hardware we had back then that was

play05:28

the risk and fortunately Alex had the

play05:32

kernels that eliminated that risk right

play05:35

that's a very good point I mean at the

play05:37

time it wasn't I mean today you put

play05:40

something in pytorch tensor flow

play05:42

whatever your favorite framework is and

play05:44

you can train in your network back then

play05:46

you actually had to build some pretty

play05:48

specialized tools yourself to to make

play05:51

this all

play05:52

run

play05:54

now as that breakthrough happens I'm

play05:58

curious what are you thinking next what

play06:01

do you think like okay we do this you

play06:02

probably knew this this breakthrough

play06:04

happened before everybody else in the

play06:06

world because I mean you you had the

play06:08

results before the public workshop and

play06:11

so before everybody else in the world

play06:12

even knew that neural Nets are going to

play06:14

be the new state-of thee art and a new

play06:15

way of doing computer vision you already

play06:17

knew that and so where was your mind

play06:19

going at that point so I think there

play06:22

were two things which I was thinking

play06:24

about so the the thing the the belief so

play06:27

my belief has been that we' proven that

play06:30

neural Nets can solve problems that

play06:33

human beings can solve in a short amount

play06:37

of

play06:37

time because with the risk we've proven

play06:40

that we can train neural Nets with

play06:42

modest numbers of layers and I thought

play06:45

we can make the neural networks wider

play06:48

but making and that will be pretty

play06:49

straightforward making them deeper is

play06:51

going to be harder and so I thought okay

play06:53

well depth is how you solve problems

play06:56

that require a lot of thinking so can we

play06:58

find some other interesting problems

play06:59

that don't require a lot of thinking and

play07:01

I actually was thinking a little bit

play07:02

about reinforcement learning but the

play07:04

other problem was problems in language

play07:07

that people can Sol can understand

play07:09

quickly as well so with language you

play07:10

also have the property that you don't

play07:11

need to spend a lot of time thinking

play07:12

know what did what did they say exactly

play07:14

you know sometimes you do but often you

play07:16

don't so problems in language and

play07:19

translation was

play07:21

the the preeminent problem in language

play07:24

at the time and so that's why I was

play07:26

wondering if he could do something there

play07:28

another thing which I was thinking about

play07:29

was was actually go as well I was

play07:31

thinking that using a

play07:33

convet could

play07:35

potentially provide very good intuition

play07:37

for the non neural network go plane

play07:40

system that existed back then can you

play07:42

say a bit more about the the go system

play07:45

um how how neur Network could and

play07:49

actually has changed then from there how

play07:52

that's

play07:53

done I

play07:54

mean basic the thing about neural

play07:57

networks is that okay

play08:00

so before deep

play08:03

learning anything you had to do with AI

play08:06

involved some kind of maybe search

play08:08

procedure with some kind of hardcoded

play08:10

heris sixs where you have really

play08:12

experienced Engineers spend a lot of

play08:15

time thinking really hard

play08:20

about how exactly under what conditions

play08:23

they should continue something or

play08:25

discontinue something or expand

play08:27

resources and they just spent all their

play08:28

time

play08:30

trying to figure out those fistic but a

play08:32

neural network is formalized

play08:35

intuition it is actually intuition it

play08:38

gives you the kind of expert gut feel

play08:40

because I read I read this thing that an

play08:43

expert player in any game you can just

play08:45

look at the situation and it instantly

play08:47

get a really strong gut feel it's either

play08:49

this or that and then I spend all their

play08:51

time thinking which one of those which

play08:52

which of those two it is it's say great

play08:54

the neural network should have

play08:55

absolutely no trouble if you buy the

play08:57

theory that we can replicate functions

play09:00

that humans can do in a short amount of

play09:01

time like less than a second and it felt

play09:04

like

play09:05

okay in case of something like go which

play09:08

was a big unsolved problem back

play09:11

then and neur should be able to do that

play09:14

back in the time Elia with the first

play09:16

time I heard that you know maybe use a

play09:18

confet for go my my naive reaction

play09:23

obviously because clearly it it

play09:25

succeeded my na naive reaction was

play09:28

confidence are famous for translation

play09:30

invariance and there's no way that we

play09:32

want to be translation invariant on on

play09:34

on the board of go because you know it

play09:36

really matters whether a pattern is you

play09:39

know in one place or another place um

play09:42

but obviously you know that that didn't

play09:45

stop The Confidence from succeeding and

play09:47

and just capturing the patterns

play09:48

nevertheless yeah I mean you know that's

play09:50

that's again the power of the parallel

play09:52

computer can you imagine programming a

play09:54

convet to do the right thing well it's a

play09:56

little bit hard to imagine that but

play10:00

it's it's true that that part may have

play10:02

been a small a small leap of faith and

play10:03

maybe to cl to close the loop on go so

play10:06

my my interesting

play10:10

go ended up in me participating on the

play10:13

alpha go paper as well in in a modest

play10:16

way you know like I I I got I had an

play10:18

intern Chris Madison and we wanted to

play10:21

apply super continents to go and at the

play10:23

same time Google acquired Deep Mind and

play10:25

all the Deep Mind folks have visited

play10:27

Google and so we spoke with David silver

play10:28

and Aang and be thought it would be a

play10:31

cool project to try out but then Deep

play10:33

Mind really they put a lot of effort

play10:34

behind it and they really had a

play10:36

fantastic execution in this

play10:38

project yeah I think while the imet

play10:42

moment is the moment most AI researchers

play10:45

saw the coming of age of deep learning

play10:48

and a whole new era starting alphao is

play10:51

probably the moment most of the world

play10:53

saw that AI is now capable of something

play10:56

very different from what was possible

play10:58

before

play11:00

um it's interesting though because While

play11:02

most of the world's focused on that

play11:04

around the same time um actually a New

play11:07

York Times article comes out saying that

play11:09

actually something very fundamental has

play11:11

been happening in national language

play11:13

processing which you alluded to and that

play11:15

actually the whole Google translate

play11:17

system was had been revamped with neural

play11:20

networks um even though a lot of people

play11:22

think of neur net as at the time as

play11:25

pattern recognition and patterns should

play11:27

be signals like speech or or visual

play11:30

signals and language is discret and so

play11:32

I'm really curious about that um how how

play11:36

do you make the leap from these

play11:38

continuous signals where NE Nets to many

play11:41

people seemed a natural fit to language

play11:45

which most people would look at as

play11:47

discrete symbols and very different yeah

play11:50

so I think that leap is very natural if

play11:53

you believe relatively strongly that

play11:57

biological neurons and artificial

play11:58

neurons are not the different because

play12:00

then you can say Okay human beings let's

play12:04

let's find let's think of the single

play12:06

best professional translator in the

play12:08

world someone who is extremely fluent in

play12:10

both languages that person could

play12:12

probably translate language almost

play12:15

instantly so there exists some neural

play12:18

network with a relatively small number

play12:19

of layers in that person's

play12:21

mind that can do this task okay so if

play12:25

you have a neural network in outside our

play12:27

computer which might be a little bit

play12:28

smaller and it's trained on a lot of

play12:30

input output examples we already know

play12:32

that we will succeed in finding the

play12:34

neural net that will solve the problem

play12:35

so therefore the existence of that that

play12:39

single really good instantaneous

play12:41

translator

play12:43

or the existence of such one such person

play12:46

is proof that the neural network can do

play12:48

it now it's a large neural network our

play12:49

brains are quite large but maybe you can

play12:51

take a leap of faith and say well maybe

play12:53

our digital neurons we can train them a

play12:56

little bit more and maybe they're a

play12:58

little bit less noisy and maybe it will

play12:59

still work out now of course the neural

play13:00

networks are still not at the level of a

play13:02

really amazing human translator so

play13:04

there's a gap but that was the chain of

play13:07

reasoning that humans can do it quickly

play13:10

biological neurons are not unlike

play13:12

artificial neurons so why can't the

play13:13

neural network do it let's find out with

play13:16

your collaborators at Google you

play13:18

invented the modern way of of doing a

play13:20

machine translation with neural networks

play13:23

which is uh really amazing can you say a

play13:25

little bit more about how that works all

play13:28

you need is a large neural network with

play13:30

some way of ingesting some

play13:33

representations of words and when the

play13:35

representation of words so what is it

play13:37

mean a representation it's a word that

play13:38

we use in AI often a representation is

play13:41

basically okay so you have the letter A

play13:44

how do you show it or the word cat how

play13:46

do you present it to the computer to the

play13:48

neural network and you basically just

play13:49

need to agree with yourself that hey

play13:52

we're going to create some kind of a

play13:53

mapping

play13:55

between the words or the letters into

play13:58

some kind of signals that happen to be

play14:01

in the format that the neuronet can

play14:02

accept so you have this one you you just

play14:05

say I'll just design this dictionary

play14:07

once and feed those signals to the

play14:09

neural net and now you need to have some

play14:11

way for the neural networ to ingest

play14:13

those signals one at a time and then it

play14:16

emits the words one at a time of the

play14:19

translation and that's literally it it's

play14:22

called the auto regressive modeling

play14:24

approach and it's quite popular right

play14:26

now but it's not because it's so but

play14:29

it's not because it's necessarily

play14:30

special it's just convenient the neural

play14:32

networks do all the work the neural

play14:34

networks figure out how to con build up

play14:38

their inner Machinery how to build up

play14:40

their neurons so that they will

play14:42

correctly interpret the words as they

play14:44

come in one at a time and then

play14:46

somehow you know break them into little

play14:49

pieces and transform them and then do

play14:52

exactly the right orchestrated dents to

play14:55

Output the correct words one at a time

play14:59

it's probably possible to design other

play15:01

neural networks that other ways of

play15:03

ingesting the words and people are

play15:05

exploring this right now you know you

play15:07

may have seen some you know if you

play15:09

follow ml Twitter you may have seen some

play15:10

words like phrases like diffusion

play15:13

models so maybe they will be able to

play15:15

ingest the words in parallel and then do

play15:17

some sequential work and then output

play15:18

them in parallel it doesn't actually

play15:19

matter what matters is that you just

play15:21

present the words to the neuronet

play15:22

somehow and you have some way that the

play15:24

neuronet can output the words of the

play15:27

Target and that's what matters yeah to

play15:29

me it was a very big surprise at the

play15:31

time that that it worked so well for

play15:32

language I I was 100% certain that it

play15:34

will work great for anything continuous

play15:36

and then all of a sudden the sequence to

play15:38

sequence models that you pioneered was

play15:41

like okay well I guess now it's going to

play15:43

work for everything was my conclusion

play15:46

because if it can work for for language

play15:48

what's what's left in terms of signals

play15:49

we we work with right um now you of

play15:54

course didn't start um working on neur

play15:57

Nets from from the day you're born and

play15:59

I'm really curious you know where did

play16:00

you grow up and how did that lead you to

play16:03

ending up you know being an AI

play16:05

researcher yeah so I was born in Russia

play16:09

I grew up in Israel and then I moved to

play16:11

Canada when I was

play16:14

16 according to my parents I've been

play16:17

talking about AI at a relatively early

play16:21

age and I definitely remember at some

play16:27

point thinking about a I and reading

play16:30

about this whole business with playing

play16:33

chess using brute force and it was

play16:35

totally clear it was it seemed that yeah

play16:38

you could do the chess stuff no problem

play16:40

but the learning stuff that's where the

play16:42

real meat of AI is that's why AI is so

play16:45

terrible because it doesn't learn and

play16:46

humans learn all the time so can we do

play16:48

any learning at all so my par so when my

play16:51

family moved to Canada to Toronto and I

play16:54

entered the University of Toronto I

play16:57

sought out the learning professors and

play17:00

that's how I found Jeff

play17:03

Hinton and then the other thing is that

play17:05

he had he had this he was into training

play17:08

neural networks and neural networks

play17:10

seemed

play17:11

like a much more promising Direction

play17:14

than the other approaches

play17:15

because they didn't have obvious

play17:18

computational limitations like things

play17:20

like decision trees which were those

play17:23

words were those that phrase was popular

play17:25

back in the day

play17:27

mhm now Jeff of course have a has a very

play17:31

long history working in Ai and

play17:35

especially neural networks deep

play17:37

learning um you know coming out of

play17:40

England coming to the US then moving to

play17:42

Canada and and his move to Canada in

play17:44

some sense helped spark the the AI the

play17:48

beginning of the new AI era in Canada of

play17:51

of all places right you're there at the

play17:54

same time which is really interesting

play17:56

kind of curious you know

play17:59

do you think there's any reason your

play18:00

parents decided to go to Toronto and

play18:02

that it is like the place where both you

play18:05

and Jeff ended up and Alex I mean the

play18:08

three of you were there together to make

play18:10

that happen I think it's a h a bit of a

play18:13

happy coincidence I think it has to do

play18:14

with the way immigration works it it is

play18:17

it is a fact that it

play18:19

is qu quite a bit easier to immigrate

play18:21

into Canada and if you immigrate into

play18:24

Canada Toronto is perhaps the most um

play18:26

appealing City to settle in

play18:30

now that coincidence brings you to

play18:33

University of

play18:34

Toronto and you find Jeff Hinton working

play18:36

on your networks but I gotta imagine

play18:38

when you you looked into his history you

play18:41

must have noticed he'd been working on

play18:42

it for 30 40 years and was there any

play18:47

moment you thought well maybe if it

play18:50

doesn't work after 30 40 years it's not

play18:53

going to work now either I see what

play18:54

you're saying but my motivation was

play18:56

different I had I had a very explicit

play18:58

motivation

play18:59

to

play19:01

make even a very very small but a

play19:04

meaningful contribution to AI to

play19:05

learning because I thought learning

play19:07

doesn't work at all completely and if it

play19:10

works just a little bit better because I

play19:11

was there I would declare it a success

play19:13

and so that was my goal and do you

play19:15

remember anything from your first

play19:16

meetings with jef how was that I mean so

play19:20

I was I was a thirdy year undergrad when

play19:23

I met him for the first time I mean I

play19:25

thought it was great so my major in

play19:29

undergrad was math but the thing about

play19:31

math is that math is very hard and lots

play19:34

of and all the really talented people

play19:36

would go into math and so one of the

play19:38

things which I thought was great about

play19:39

machine learning is that not only it is

play19:41

the thing but also all the really clever

play19:44

people going into math and

play19:46

physics so I was very pleased about that

play19:48

what I remember from actually reading

play19:50

Kade metz's book um is actually my my

play19:55

possibly my favorite anecdote from the

play19:57

book tell has Jeff telling the story

play20:00

about him meeting you Ilia and so here

play20:04

here's how the book tells the story

play20:06

maybe you've read it maybe not but

play20:07

essentially the book says yeah there's

play20:10

Jeff you know and this this young

play20:13

student comes in il it's of her

play20:15

undergrad still and Jeff gives you a

play20:19

paper and

play20:23

um

play20:24

you you go read

play20:27

it and you you you come back and um you

play20:33

tell him I don't understand it and

play20:36

Jeff's like oh that's okay you know

play20:38

you're still underground what don't you

play20:40

understand I can explain it to you and

play20:43

essentially you say actually I don't

play20:46

understand why it's why they don't

play20:48

automate the whole process of learning

play20:50

it's it's still too too much handholding

play20:53

um I understand the paper I just don't

play20:55

understand why they're doing it that way

play20:56

and Jeff's like okay wow this is this is

play20:58

is interesting it gives you another

play21:00

paper and um again you go read you come

play21:04

back so goes the story and you say oh I

play21:08

don't understand this one either and

play21:11

Jeff's like what do you understand don't

play21:12

you understand about this one I'm happy

play21:14

to explain and you go I don't understand

play21:15

why they train a separate neural network

play21:17

for every application why can't we train

play21:20

one gigantic Network for everything it

play21:22

should you know it should help to be

play21:24

trained jointly and to me that that

play21:27

that's really I mean that reminds me a

play21:29

lot of our times at open the eye where

play21:31

it always felt like you are you know

play21:34

already thinking you know several steps

play21:36

into the future of how things are going

play21:38

to shape up just from the evidence we

play21:40

have today you know how it really should

play21:42

be several years down the line uh that

play21:45

at least according to the book that's

play21:46

how Jeff remembers the first two

play21:49

meetings with you yeah I mean some some

play21:52

something like this did happen it's true

play21:55

so the field of AI back then when I was

play21:57

starting out was not not a hopeful field

play22:00

it was a field of desolation and despair

play22:03

no one was making any progress at all

play22:06

and it was not clear if progress was

play22:08

even

play22:09

possible and so that's why well H how

play22:11

what do you do when you are in this

play22:13

situation so you say you walking down

play22:14

this path this is the path the most

play22:17

important path but you have no idea how

play22:19

long it is you have no idea how hard

play22:20

it's going to be what would be a

play22:22

reasonable goal in this case well the

play22:25

goal which I chose was can I make a

play22:27

useful step one use useful step so that

play22:30

was

play22:30

my explicit motivation at least for

play22:33

quite a while before it became clear

play22:36

that actually the path is going to

play22:37

become a lot Ste a lot you know a lot

play22:39

sloper and a lot more rapid where

play22:41

Ambitions became grew very rapidly but

play22:44

at first when there was no no gradient

play22:47

the goal was just make any Step at all

play22:50

anything useful that would be meaningful

play22:52

progress towards Ai and I think that's

play22:53

really intriguing actually because I

play22:55

think that's what drives a lot of

play22:57

researchers is is

play22:59

to just find a way to make to make some

play23:02

progress knowing not knowing actually

play23:05

ahead of time how far you can get but

play23:08

just being so excited about the topic

play23:10

that you you just want to find a way to

play23:13

at least make some progress and and then

play23:15

keep going um and it's it of course very

play23:18

interesting in your case that you know

play23:20

then that whole thing switched from you

play23:23

know slow progress to ever faster

play23:26

progress all of a sudden thanks to the

play23:29

thing that you're like look you you're

play23:30

trying to make that bit of progress and

play23:32

it turns out to open up the floodgates

play23:34

for for massive

play23:36

progress now you start you start in

play23:39

Canada you your PhD research of course

play23:42

you know completely changes the field

play23:45

you start a company that gets acquired

play23:48

um by Google and you're at Google then

play23:51

the big thing and also the the moment

play23:53

actually our paths start start Crossing

play23:55

or about to cross is that you you know

play23:58

you're on this role at Google you're

play24:00

doing some of the most amazing

play24:02

pioneering work in AI you're clearly in

play24:05

an amazing situation where you are you

play24:09

know doing some some of the best work

play24:11

that's happening in the world and you

play24:14

you decide to change your

play24:17

situation how did that come about I

play24:20

remember being at Google and feeling

play24:23

really

play24:24

comfortable and also really restless

play24:28

I think two two factors contributed to

play24:31

that one is that I I somehow I could

play24:33

look 10 years into the future and I had

play24:36

a little bit too much clarity about how

play24:38

things will look like and I didn't enjoy

play24:41

that very much but there was another

play24:42

thing as well and that's

play24:44

the the experience of seeing Deep Mind

play24:47

bu work on alphao and I and it was it

play24:51

was very inspiring and I thought

play24:54

that it's a sign of things to come that

play24:57

the field is starting to sure up until

play24:59

that point all progress in AI has been

play25:02

driven by individual researchers working

play25:04

on small projects maybe small groups of

play25:07

researchers with some advice by their

play25:10

professors and maybe some other

play25:11

collaborators but usually it would be

play25:12

small groups it would most it would most

play25:15

of the work would be idea heavy and then

play25:17

it would be some kind of a some effort

play25:20

on an engineering on the engineering

play25:22

execution to prove that dat is valid but

play25:26

I felt that alphago was a little

play25:28

different

play25:28

it showed that in

play25:31

fact it showed to me that the

play25:34

engineering is critical and in fact the

play25:37

field will change and you become the

play25:39

engineering field that it is

play25:41

today because the tools were getting

play25:44

very solid and the question then becomes

play25:46

okay how do you really Trin those

play25:48

networks how do you debug them how do

play25:50

you set up the distributed training and

play25:52

it's a lot of work and the stack is

play25:54

quite deep

play25:56

and I felt that the culture at Google

play26:01

was very similar to the Academia culture

play26:05

which is really good for generating

play26:07

radical novel ideas and in fact Google

play26:09

has generated a lot of radical and

play26:11

revolutionary ideas in AI over the years

play26:14

and most most notably the Transformer

play26:17

from from the past few

play26:19

years but I felt that that's not going

play26:21

to be the whole of progress in AI I felt

play26:25

that it's not now only a part of

play26:27

progress in AI so if you think of it as

play26:28

of of the body you can say you need both

play26:31

the the muscles and the skeleton and the

play26:34

nervous system and if you only have one

play26:36

it's amazing but the whole thing won't

play26:39

won't really move you need all things

play26:41

together and so I felt that I had a

play26:43

vague feeling that it would be really

play26:45

nice if there

play26:49

was some kind of a company which

play26:52

would have these elements together but I

play26:55

didn't know how to do it I didn't have

play26:56

any any path to it I was kind of I just

play26:58

daydreaming about it and then at some

play27:00

point I got an email from Sam Altman

play27:02

saying hey let's get dinner with some

play27:03

cool people and I said sure and and I

play27:06

showed up and and Greg Greg Brookman was

play27:09

there and Elon Musk was there and a few

play27:12

others and we just chatted about

play27:16

wouldn't it be nice to start an a new AI

play27:18

lab and I found that really the time was

play27:21

right because I was thinking about the

play27:22

same thoughts

play27:23

independently and I really wanted it to

play27:25

be engineering heavy and you know no no

play27:29

seeing that Elon was going to be

play27:31

involved I thought well who better to

play27:33

who would be

play27:35

better can't imagine a better person

play27:37

from whom to learn the you know big

play27:40

engineering project side of

play27:43

things so I think this was the Genesis

play27:45

there is kind of there there is more to

play27:47

it but I think that was the real the

play27:49

real Genesis of open AI from my

play27:51

perspective that yeah like I was

play27:53

thinking about something and then it

play27:55

just one day I woke up with this email

play27:57

hey the thing from my perspective it was

play27:59

like I was daydreaming about something

play28:00

and then my Daydream come came true

play28:02

almost like this the the dream Daydream

play28:05

becomes true what you're really saying

play28:07

there is that you know there is a group

play28:09

of

play28:11

people very highly accomplished and

play28:13

ambitious people who are in some sense

play28:17

aligned with your dream and want to want

play28:19

to make this happen together but all

play28:21

that gets you is essentially you know

play28:23

some some paperwork that a new company

play28:26

now exists and um maybe some money to to

play28:29

get going but you actually still need to

play28:31

decide what to do with those resources

play28:33

and with your time I'm kind of curious

play28:36

at the beginning of of open eye what

play28:38

what was going on in your mind in terms

play28:40

of how to shape this up um I mean

play28:43

obviously it's been been a massive

play28:44

success but I'm really curious about you

play28:46

know that the beginning part and how how

play28:48

that played out for you so the beginning

play28:50

part I would describe it as a whole lot

play28:53

of

play28:55

stress and it wasn't exactly clear how

play28:59

to get going right

play29:01

away there was only clarity about a few

play29:03

things which is there need to be some

play29:06

kind of a large

play29:07

project

play29:10

and I also was excited about the idea

play29:12

that maybe if you can predict really

play29:14

well you make progress on un supervised

play29:16

learning but beyond that it wasn't clear

play29:18

what to do so we tried a whole lot of

play29:20

different things and then we decided

play29:22

that maybe it would be good to solve a

play29:25

difficult computer game Dota and if this

play29:27

is this and this is where Greg just

play29:30

showed his strength and he just took on

play29:32

this project even though it seemed

play29:34

really impossible genuinely impossible

play29:36

and just went for it and

play29:38

somehow it worked in the

play29:42

most stereotypical deep learning way

play29:44

where the simplest method that he tried

play29:47

just ended up working the simplest

play29:49

policy gradient method as we kept

play29:51

scaling it up just never never stopped

play29:55

improving with more scale and more

play29:57

training just to double click on that

play29:58

for for a moment I don't think everybody

play30:00

knows what DOTA is can you say a bit

play30:03

about that and I mean I fully agree why

play30:06

it's so surprising that the simplest

play30:08

approach ultimately work is a very hard

play30:11

problem so for some context the state of

play30:13

the field back then was okay so if you

play30:17

look at reinforcement learning in

play30:19

particular Deep Mind has made some very

play30:22

exciting progress first by training rein

play30:26

a neural net with reinfor ment learning

play30:28

to play Simple computer

play30:30

games and then and then the reaction was

play30:33

okay that's exciting and interesting and

play30:35

kind of cool but what else can it do and

play30:38

then alphago happened and then the

play30:40

opinion is shifted okay reinforcement

play30:42

learning maybe can do some things but

play30:44

you know go it's funny by the way go

play30:47

used to seem this used to to be this

play30:49

impossible game and now everyone says oh

play30:51

such a single game the board is so small

play30:53

our

play30:54

perceptions change quickly but then

play30:58

Deep Mind we're talking about how

play31:00

Starcraft is the next logical step after

play31:02

to go and it made a lot of sense to me

play31:05

as well it seemed like a much harder

play31:08

game not necessarily in its not

play31:10

necessarily from the for for a if a

play31:14

person to play but for our tools it

play31:16

seemed harder because it had much

play31:19

more you had a lot more moving Parts

play31:22

it's much more chaotic it's a realtime

play31:24

strategy game and we thought that it

play31:27

would be nice to have our own twist on

play31:29

it and to try to make a bot which can

play31:32

play DotA and DOTA is another real-time

play31:35

strategy game that's really popular it's

play31:37

been the it had I believe it definitely

play31:43

had I don't know if it still has the

play31:45

largest prize

play31:47

pool the largest annual prize pool of

play31:49

any professional esport

play31:51

game so it was very it has a very

play31:54

vibrant very strong professional scenes

play31:56

people dedicate their lives play in this

play31:58

game

play32:00

they it's it's a game of of reflex and

play32:03

strategy and Instinct and a lot of

play32:05

things happen you don't get to see the

play32:07

whole game the the point is that it

play32:09

definitely felt like a grand challenge

play32:11

for reinforcement learning at that time

play32:14

and our opinion about the tools of

play32:17

reinforcement learning was so let's put

play32:19

it this way so the Grand Challenge felt

play32:21

like it's here and The Field's opinion

play32:24

about the tools and their ability to

play32:26

solve a problem like this was like here

play32:29

there was a huge mismatch and so when we

play32:31

started working on it we we thought oh

play32:32

yeah we're going to need to develop all

play32:34

kinds of crazy planning methods and

play32:36

hierarchical reinforcement learning

play32:37

methods and whatnot but let's just get a

play32:39

baseline let's just see when the

play32:41

Baseline breaks

play32:45

and that's when the Baseline just didn't

play32:50

break it just kind of kept improving all

play32:51

the time and it's interesting with

play32:53

each with each so what would happen on

play32:56

over the course of this project we would

play32:58

have these public demonstrations of our

play32:59

progress as we reach different

play33:01

Milestones to Performance we would have

play33:03

some kind of a public exhibition game

play33:05

against a professional of different

play33:07

level of accomplishment so at first we

play33:09

had a a public exhibition game

play33:13

against retired professionals then we

play33:15

had them against active professionals

play33:17

and then finally we had a game against

play33:19

the strongest professionals and we

play33:20

defeated them but the interesting thing

play33:23

is that at each step you'd have people

play33:25

who you'd have very knowledgeable

play33:26

experts in a

play33:28

who would come out on Twitter and say

play33:32

well this was really cool great

play33:33

successful reinforcement learning but

play33:35

obviously the next step would require

play33:37

the plan the explicit planning thing or

play33:39

the hierarchy thing and somehow it's did

play33:43

not so that was that was a very

play33:47

important

play33:48

result for us I felt like it really it

play33:52

really proved to us that we can do large

play33:54

projects I remember I was not part of

play33:56

this project uh just to be clear um but

play33:59

but I was there at open ey when it was

play34:00

all happening working on on other

play34:02

projects and I remember being very very

play34:05

surprised and that no explicit structure

play34:11

was

play34:12

needed though OB though well in my mind

play34:15

obviously but maybe you know it's not

play34:16

even true but in my mind there is this

play34:20

large lstm model neural network

play34:25

that maybe somehow through back

play34:28

propagation actually internalize the

play34:29

structure that we all at least not all

play34:31

of us but maybe me I thought we would

play34:33

have to put in explicitly and maybe the

play34:35

ne network was able to just absorb that

play34:38

intuition through back propagation

play34:39

without the need to to hardcode it which

play34:42

was really intriguing to me because it

play34:45

just seemed like wow um you know a lot

play34:48

of intuitions might be better uh

play34:50

provided through through data than

play34:52

through hard coding which seems a very

play34:54

common Trend in all of deep learning but

play34:56

maybe in reinfor wining at the time

play34:58

wasn't that strongly believed yet till

play35:01

till that result came out yeah I mean I

play35:03

I agree I agree with your assessment I

play35:05

feel

play35:07

like yeah

play35:09

I I I like to think that this result had

play35:12

changed the fields view at least a

play35:15

little bit about the capability of

play35:18

simple reinforcement learning now to be

play35:21

fair you still need quite a hefty amount

play35:23

of experience to get a very strong

play35:25

result on such a game

play35:28

and then we also use the similar

play35:31

approach so I would say if you if you

play35:34

have the ability to generate a very

play35:35

large amount of experience against some

play35:37

kind of a simulator then this style of

play35:40

reinforcement learning can be extremely

play35:43

successful and in fact we have

play35:46

also another important

play35:49

result in open a history was to use the

play35:53

same exact approach to train a robot to

play35:56

solve the Rubik Cube So Physical robot a

play35:59

physical robot hand actually soled the

play36:01

physical Rubik's Cube and it was a quite

play36:04

challenging project the training was

play36:06

done entirely in

play36:09

simulation and the simulation was

play36:11

designed in such a way so that it's

play36:13

extra hard and it requires the the

play36:17

neural net to be very adaptive so that

play36:20

when you give it the real robot the the

play36:22

real physical robot it will still

play36:23

succeed but at core it was the same

play36:26

exact approach as the one we used with

play36:29

the DOTA project which was very large

play36:32

scale reinforcement learning in fact it

play36:34

was even the same

play36:36

code

play36:38

so that was a case where we had this

play36:40

General technique these General powerful

play36:42

results which we were able to use in

play36:44

more than one place and that was what

play36:47

you've done on reinforcement L now I

play36:50

know that right now there's other

play36:51

reinforcement learning happening at

play36:53

openi in the context of of language

play36:56

actually before we we get to and I'm

play36:57

really curious about about that but

play36:59

before we we get to

play37:02

that

play37:03

um language modeling

play37:06

GPT is probably you

play37:10

know the the most visible thing in

play37:13

recent

play37:14

years in the public eye of what AI is

play37:18

capable of and you know opening ey

play37:21

generated these GPD generations of

play37:23

models that can complete articles in

play37:26

very credible always um and it's been

play37:31

very surprising how capable it is and so

play37:34

what I'm really curious about again in

play37:37

some sense is you know you decided that

play37:40

I mean not alone but together with

play37:41

collaborators at open ey you decided

play37:43

that you know it was it was time was

play37:47

right to to go down this path of you

play37:51

know building language models and I'm

play37:52

really curious what what was it for you

play37:55

that made you believe that you know this

play37:57

was the thing to start doing yeah so

play37:59

from my

play38:01

side a really important thing that

play38:03

happened to me is that I was really

play38:05

interested in unsupervised

play38:08

learning and for

play38:11

context the the the results that we

play38:13

spoke about earlier on about vision and

play38:16

even about you know go and DOTA all

play38:19

these results translation they are all

play38:22

cases where you have somehow you train a

play38:25

neural network by presenting it with

play38:27

inputs

play38:28

and desired

play38:29

outputs you have your random input not

play38:32

random you have a typical input a

play38:35

sentence an image something and you have

play38:38

the desired output and the neural

play38:41

network you run it and you compare the

play38:44

predicted output with the desired output

play38:47

and then you change the neural network

play38:49

to reduce this error and you just do it

play38:51

a lot you do it a lot and that's how

play38:53

learning works and it's completely

play38:55

intuitive that if you will do this the

play38:57

neural network will succeed I should say

play39:00

maybe not completely intuitive but

play39:02

definitely pretty intuitive today

play39:04

because you say hey here is my input

play39:06

here's my desired output don't make the

play39:08

mistakes eventually the mistakes will go

play39:11

away and it is something where you can

play39:14

at least have a reasonably strong

play39:15

intuition about why it should work why

play39:18

supervised learning works and why

play39:19

reinforcement learning works in contrast

play39:23

at least in my mind unsupervised

play39:25

learning is much more mysterious now

play39:27

what is unsupervised going exactly it's

play39:29

the idea that you can understand the

play39:32

world whatever that means you can

play39:34

understand the World by simply observing

play39:37

it without there being a teacher that

play39:40

will tell you what the desired Behavior

play39:43

should

play39:44

be so there is pretty obvious question

play39:46

which is okay so

play39:48

like why would like how could that

play39:51

possibly work how could it possibly be

play39:54

that you have okay so what would you do

play39:56

then what was the typical prevailing

play39:58

thinking the prevailing thinking has

play40:00

been that maybe you have some kind of

play40:02

task

play40:04

like you take your input your

play40:07

observation an image let's say and then

play40:10

you you ask the neural network to

play40:12

somehow transform it in some way and

play40:14

then to reproduce the same image back

play40:17

but why would that be a good thing for

play40:18

the test you care about is there some

play40:20

mathematical reason for it I found it

play40:22

very unsatisfying in my mind it felt to

play40:26

me like the there is no good

play40:28

mathematical basis for unsupervised

play40:29

learning at all

play40:30

whatsoever and I was really bothered by

play40:33

it and after a lot of thinking I had I

play40:37

had the I had the I developed the belief

play40:39

that actually if you predict the next

play40:42

bit really well you should have a really

play40:44

good on supervis

play40:46

level the idea is that if you can

play40:48

predict the next bit really well then

play40:50

you have extracted all the meaningful

play40:53

information that somehow the model knows

play40:55

about all the meaningful information

play40:57

exist in the signal and therefore it

play40:59

should have a representation of all the

play41:01

concepts and it's the idea in the

play41:03

context of language modeling it's very

play41:05

intuitive you know if you can predict

play41:07

the next word moderately accurately

play41:09

maybe the model will know that words

play41:11

that just clusters of characters

play41:13

separated by space if you predict better

play41:16

you might know that there is a

play41:17

vocabulary but you won't be good at

play41:19

syntax if you improve your prediction

play41:21

even further you'll get better at the

play41:22

syntax as well and suddenly we'll be

play41:24

producing syntactical mambo jumbo but if

play41:26

you improve prediction even further

play41:29

necessarily the semantics has to start

play41:31

kicking in I felt that the the same the

play41:33

same argument can be made about

play41:36

predicting pixels as well so at some

play41:40

point I started to believe that maybe

play41:42

doing a really good job on

play41:44

prediction they'll give us on supervis

play41:47

larel which back then felt like a Grand

play41:50

Challenge and other interesting thing

play41:51

that now everyone knows that

play41:53

unsupervised learning just

play41:55

works but not that long ago it seemed

play41:58

like this completely intractable

play42:00

thing so anyway to come back to the

play42:03

story of how the gpts were created so

play42:07

then you know I'd say the first project

play42:10

that really uh was a step in this

play42:12

direction was led by Alec Ratford who is

play42:15

an important hero of the GPD Saga where

play42:18

we trained a neur an lsdm to predict the

play42:21

next character on reviews on on on um on

play42:25

Amazon reviews of products

play42:28

and we discovered that this lsdm has a

play42:30

neuron which corresponds to

play42:33

sen in other words if you are reading a

play42:36

review which is positive the sentiment

play42:38

neuron will fire and if you're reading a

play42:41

review which is negative the sentiment

play42:43

NE will not

play42:44

F so that's interesting and that felt to

play42:48

us like it validated the conjecture of

play42:50

yeah of course eventually if you want to

play42:52

predict what comes next really

play42:55

well you need

play43:01

need you need to discover the truth

play43:03

about the

play43:06

data and so then what happened is that

play43:08

the Transformer came

play43:10

out and then we saw the Transformer and

play43:12

I think it was it was

play43:15

pretty like it got us really excited

play43:18

because we were really struggling we

play43:19

believe that long-term dependencies were

play43:21

really

play43:21

important and the Transformer had a very

play43:24

clean elegant and compute efficient

play43:27

answer to long-term dependency and for

play43:29

context the Transformer is this neural

play43:31

network

play43:32

architecture and in some sense it's just

play43:35

really

play43:36

good but a little bit more technically

play43:40

so we discussed that these neural

play43:41

networks are deep in some way and we

play43:43

know and it's been the case until

play43:45

relatively recently that it was pretty

play43:47

hard to train deep neural networks and

play43:50

previous neural networks for training

play43:52

models on sequences of language the

play43:55

longer the sequence was the deeper the

play43:57

Network would get the harder it would be

play43:58

to

play44:00

train but the Transformer decoupled the

play44:04

depth of the Transformer from the length

play44:06

of the sequence so you could have a

play44:08

Transformer of manageable depth with

play44:09

very long sequences and that was

play44:12

exciting and this investigation led to

play44:16

gpt1 and then I would say further then

play44:20

we we we contined to believe in scale

play44:22

and that led and that led to gpt2 and

play44:24

three and here it's really I want to I

play44:25

want to call out dark mod who really

play44:28

believed that if he were to scale up the

play44:30

gpts it would be the most amazing thing

play44:31

ever and that's how we got

play44:36

gpt3 in GB3 I mean when it came out it

play44:41

wasn't just I think was so exciting to

play44:44

the entire Community it wasn't just

play44:45

something that could

play44:47

complete text when you start with a

play44:49

prompt it could maybe say oh this is

play44:51

likely your next

play44:53

sentence you I mean it could complete

play44:56

all kind kinds of things people would

play44:58

write web pages even write some very

play45:01

basic code that gets completed with gpt3

play45:05

they they would

play45:06

um and and they would be able to prompt

play45:09

it and and that really intrigued me this

play45:11

notion of of prompting where you have

play45:13

this gigantic model that's trained on I

play45:15

don't know how much text out there but

play45:18

that somehow when you then briefly feed

play45:21

it a little bit of extra text in in the

play45:25

moment you can actually prime it to

play45:29

start doing something that you wanted to

play45:30

do can you say a bit more about that

play45:32

where did that come from and how how how

play45:34

how does that work you think so what is

play45:36

a language model

play45:38

exactly you just have a neural network

play45:40

that takes takes some text and tries to

play45:43

Output an educated guess of what the

play45:45

next word might be and it outputs an

play45:48

educated guess it might

play45:50

say you know it's 30% the word the some

play45:55

kind of a prob guess of probabilities

play45:57

what the words might be then you can

play45:59

pick a word according to this

play46:00

probability that the neural net outputs

play46:02

and then commit to it and then ask the

play46:04

neural to predict the next word again

play46:06

and again and

play46:07

again now we know that real the we know

play46:11

that real text in some sense is very

play46:13

responsive to its beginning like we know

play46:16

that text has a lot of very complicated

play46:17

structure and if you read a document

play46:19

which says this document below will'll

play46:21

describe a list of questions that were

play46:23

given in the um uh

play46:27

MIT entrance exam in the 1900 I just

play46:29

made it up then I I strongly expect that

play46:32

in fact there will be 10 or so questions

play46:36

in math of the kind of math that was

play46:38

usually in math exams in the

play46:40

1900 if the model is good enough it

play46:43

should actually do

play46:44

that now how good enough is good enough

play46:47

well this is a little bit of a

play46:48

qualitative statement but if it is

play46:50

definitely good enough it should be able

play46:52

to do it so then you train a gpt3 and

play46:54

you see can it actually do it and

play46:56

sometimes cannot but very often indeed

play46:58

it is responsive it is very responsive

play47:01

to whatever you whatever text you give

play47:04

it because to predict what comes

play47:06

next correct well enough you need to

play47:09

really understand the text you're

play47:12

given and I think this is kind of in

play47:14

some way the centrality of

play47:16

prediction good enough prediction gives

play47:19

you everything you could ever dream

play47:21

about now one of the things that I think

play47:24

also stood out to me with GPT is

play47:29

that it's it's a research breakthrough

play47:32

it's a major research

play47:34

breakthrough but it also feels very

play47:36

practical like I mean whenever I'm

play47:39

typing

play47:40

something I mean I know what I want to

play47:43

type

play47:44

next it's already in my head but I still

play47:47

have to type it but with a

play47:50

GPT you know gbt2 onwards probably it it

play47:54

could complete it fairly accurately and

play47:56

so it it seemed like very different in

play47:59

that sense from for example the Rubik's

play48:01

Cube breakthrough or the um DOTA

play48:04

breakthroughs which

play48:06

were fundamental research breakthroughs

play48:08

but it was hard to dream of the direct

play48:12

applications and here with

play48:14

GPT it was so easy to to dream of so

play48:17

many applications and I'm curious if

play48:19

that you know in in your own uh kind of

play48:23

evolution on things when GPT started

play48:26

working did you start thinking about

play48:28

applications or did you know more

play48:30

generally people around you at open a

play48:32

start thinking about applications what

play48:33

was going on yeah we were

play48:36

definitely excited about the potential

play48:38

applications I mean we were so excited

play48:40

about them that we built a whole API

play48:43

product around gpt3 so that people could

play48:45

go and build their new and convenient

play48:50

and sometimes unprecedented applications

play48:53

in

play48:54

language I mean I think it's it's it's a

play48:59

general so maybe another a way of

play49:02

looking at what's happening is that AI

play49:04

is just continues to continuing to get

play49:06

more and more

play49:09

capable

play49:12

and it can sometimes be tricky to tell

play49:16

if a particular research Advance is real

play49:19

or not

play49:20

real suppose you have some cool demo of

play49:23

something like what do you make of it it

play49:26

can be hard to understand the magnitude

play49:28

of the advance especially if you don't

play49:29

know how similar the demo is to their

play49:32

training data for

play49:33

example but if you have a product that's

play49:36

useful then the advance is real and I

play49:38

feel that maybe in a sense we have moved

play49:42

away

play49:44

from the field has matured so much that

play49:47

we no longer need to rely on demos and

play49:50

even benchmarks as indicators of as the

play49:53

only indicators of progress but

play49:55

usefulness as the truest indicator of

play49:59

progress and so that's why I and and so

play50:01

I think this

play50:03

is a good sign for gpt3 for sure and

play50:06

yeah the applications we were excited

play50:08

about them and people are using gpt3 all

play50:11

the time right now are there any uses

play50:13

that you've seen that you're able to

play50:14

share the applications being built

play50:17

there's plenty of applications I

play50:18

remember seeing something that helps you

play50:20

write a

play50:21

resume and cify it something that helps

play50:24

improve your emails I think I've seen

play50:26

something like this I don't remember but

play50:27

they all have this kind of

play50:30

flavor I know that there is a lot of

play50:33

users unfortunately I don't remember

play50:35

specific examples of the top of my head

play50:37

this is jumping ahead a little bit in

play50:38

the progression of of of the research

play50:41

trajectory you've gone through with open

play50:43

eye but maybe the biggest application of

play50:45

course and maybe it's not called GPT

play50:47

anymore it's called codex but it it's

play50:49

very similar it's a system that can help

play50:51

you write

play50:52

programs can you say a bit about that

play50:54

and how is it I'm curious is it just

play50:57

like GPT but trained on on GitHub code

play51:00

instead of text or are there some

play51:02

differences so the the the system that

play51:05

we described in the paper is essentially

play51:08

a GPT train on code it's that

play51:11

simple the thing that's the thing that's

play51:15

interesting about it is that it works as

play51:16

well as it

play51:18

does because you can say like what what

play51:21

have you even done you've done nothing

play51:22

you just took a large neural net and you

play51:24

train it and code from GitHub

play51:28

but the result is not bad at

play51:33

all the abil its ability to it can solve

play51:37

real coding

play51:39

problems much better than I think most

play51:42

people would have

play51:43

expected and again this comes back to

play51:46

the the power of deep learning the power

play51:49

of these neural Nets they don't care

play51:52

what problem to

play51:53

solve and you can all kind of say well

play51:56

you know people can code so why can't

play51:59

neuron if you believe that in a

play52:01

biological neuron is not very different

play52:03

from an artificial one then it's not an

play52:05

unreasonable belief at all so then the

play52:08

question becomes what's the training

play52:09

data you know predicting giab is not

play52:12

exactly the same as coding so maybe it

play52:15

won't quite do the right

play52:17

thing but it turned out to be good

play52:19

enough and it turned out to be very

play52:21

useful especially in situations you have

play52:23

a library which you don't know because

play52:25

it's right all of GitHub it has such

play52:26

familiarity with all the major libraries

play52:29

and if you don't know it but you kind of

play52:31

just write a comment use this library to

play52:33

do X you come up with code which is

play52:35

going to often be correct or pretty

play52:38

close and then you have something to

play52:39

work with and you edit it a little bit

play52:41

and you have something

play52:42

working but yeah it's just it's just the

play52:45

GPT training to predict code pretty

play52:48

well I think in many ways it's really

play52:50

mind-blowing in terms of potential

play52:52

societal impact because if I think

play52:55

about a lot of the the way we create

play52:59

impact in the world as people we're

play53:01

often sitting behind a computer right

play53:04

and we're we're typing things and

play53:07

whether it's typing emails or or writing

play53:10

up documents on on work we've been doing

play53:13

or writing code

play53:17

um this could really accelerate any

play53:20

anybody's work and and the kind of

play53:22

things we could do in one day I don't

play53:24

know if we're already seeing metrics for

play53:25

this but I would imagine that you know

play53:27

if it's not now in the Next Generation

play53:29

and I'm curious about your thinking you

play53:31

know what kind of productivity we can

play53:33

expect from from people thanks to these

play53:35

tools so I'd say that in the near

play53:38

term productivity will continue to

play53:40

increase

play53:42

gradually I think that as time goes by

play53:45

and the capability of AI systems

play53:47

increases productivity will increase

play53:49

absolutely dramatically I feel very

play53:52

confident in that we will have we will

play53:54

witness dramatic increases in

play53:56

productivity

play53:57

eventually in the long term a day will

play53:59

come when the systems will in fact just

play54:01

the world will be kind of like the AI is

play54:04

doing all the work and then that work is

play54:05

given to people to enjoy that what I

play54:08

think is the long-term

play54:10

future will hopefully be

play54:13

life so in the medium term it's going to

play54:15

be amazing productivity increases and

play54:18

then in the longterm future it's going

play54:19

to be like infinite productivity or

play54:22

fully automated productivity now one of

play54:24

the things that of course people think

play54:26

about a lot in that context when you're

play54:28

giv AI a lot of productivity it better

play54:32

be productive doing the right thing and

play54:35

better not be productive I don't know

play54:37

you know blowing something up by mistake

play54:39

and so forth or just misunderstanding

play54:41

what it's supposed to be doing

play54:45

and in that sense I've been really

play54:47

curious about this project at open ey

play54:50

where reinforcement learning is combined

play54:52

with

play54:53

GPT can you say a bit more about that

play54:56

step back so we have these AI systems

play54:58

that are becoming more and more

play55:01

powerful and a great deal of their power

play55:04

is coming from us training them on very

play55:07

large data sets we don't understand for

play55:10

which we have an intuitive understanding

play55:12

of what they do so they learn all kinds

play55:14

of

play55:15

things and then they act in ways which

play55:19

we can inspect but perhaps

play55:21

not we can inspect but it might be but

play55:25

and we do have

play55:27

for these large language models for

play55:28

example we do have some ability to

play55:29

control them through the prompts and in

play55:33

fact the better the language model will

play55:34

get the more controllable it will become

play55:37

through the prompt but we want more we

play55:40

want our models to do exactly what we

play55:42

want

play55:44

or act closer to what we want as much as

play55:48

possible so we had this project indeed

play55:51

that you alluded to of training these

play55:54

language models with reinforcement

play55:55

learning from feedback where now you do

play55:59

reinforcement learning not against a

play56:01

simulator but against human judges that

play56:05

tell you whether the output was

play56:06

desirable or undesirable and if you

play56:08

think about it this this reinforcement

play56:10

learning environment is really exciting

play56:12

you could even argue that reinforcement

play56:13

learning has kind of maybe slowed down a

play56:16

little bit because there weren't really

play56:18

cool environments in which you could do

play56:20

it but doing reinforcement learning with

play56:22

language models and with people that

play56:25

feels like such it's such a it opens

play56:27

such a par Vista such a you can do so

play56:30

many things there and what we've

play56:33

shown is that these large neural

play56:37

networks these large GPT models when you

play56:40

do reinforcement learning from

play56:45

these from these teachers essentially

play56:48

and I should also say there is a small

play56:49

technicality which again this is a

play56:50

technical thing for the ml focused

play56:53

subset of the audience in reinforcement

play56:56

learning you're usually providing reward

play56:58

good or bad but the way we do it with

play57:00

reinforcement learning from Human

play57:02

feedback is

play57:04

that the the teacher needs to look at

play57:08

two outputs by the model and to say

play57:11

which one is better because it's an

play57:13

easier

play57:14

task it's an easier task to compare two

play57:17

things than to say whether one thing is

play57:18

good or bad in absolute and then we do a

play57:20

little bit of machine learning in order

play57:23

to then create a reward out a reward out

play57:26

of it a reward model and then use this

play57:28

reward model to train the neural net and

play57:30

this is a pretty sample efficient thing

play57:31

to do and you you you obtain a very fine

play57:34

grained way of controlling the behavior

play57:38

of these neural networks of these

play57:40

language models and we've been using it

play57:42

quite a bit like recently we've trained

play57:44

we've been training these instruction

play57:45

following models which actually people

play57:47

can use through the API through the open

play57:50

AI API where in

play57:53

gpt3 the model is just train of the

play57:55

instrument so you need to be quite

play57:56

clever about specifying your prompt

play57:59

specifying your prompt into design into

play58:01

kind of and getting the model to do what

play58:04

you want providing some examples whereas

play58:07

the instruction following model has been

play58:09

trained in this way to literally do what

play58:11

we tell it to

play58:13

so there's a word which I think is known

play58:16

in some subsets of the machine Learning

play58:18

Community but not in all of it and it's

play58:19

called the model is is this this is an

play58:22

attempt to align the model so that the

play58:24

model with its power and with great

play58:27

power and unclear capabilities will in

play58:30

fact be trained and incentivized to

play58:32

literally do what you want and with the

play58:33

instruction following model you just

play58:35

tell it what you want do X write y

play58:38

modify Z and it will do it so it's

play58:41

really convenient to use and this is an

play58:43

example of the technique of

play58:46

reinforcement learning from Human

play58:48

feedback in practice but moving forward

play58:52

you of course you want to learn from

play58:54

teachers in all kinds of ways

play58:56

and you want to use machine learning to

play58:58

not not just have people you know

play59:02

provide supervised examples or provide

play59:04

rewards but you would really want to

play59:05

have a conversation where you ask

play59:07

exactly the right question to learn the

play59:09

information that you need to understand

play59:10

the concept so that's how things will be

play59:12

in the future but right now this

play59:15

approach has been used fairly

play59:17

successfully to all to make our GPT

play59:20

models more aligned than than they are

play59:23

naturally I'm going to say aligned as I

play59:26

understand it you can also align them in

play59:28

a personalized way so align to a

play59:30

specific person's preferences like I

play59:32

could teach you to follow my preferences

play59:34

and you could have a different one I

play59:36

mean So Def the answer is definitely yes

play59:40

so the specific model that I mentioned

play59:41

to you the instruction following model

play59:43

this model it's a single model and it's

play59:47

been

play59:48

aligned you know it's been you know we

play59:50

say it's aligned which is which is

play59:52

another which is a way to say that it's

play59:54

been trained and incentivized to follow

play59:57

the instruction you

play59:58

g so it it's an interface and it's a

play60:03

very convenient interface of course it

play60:05

is possible with these neural Nets they

play60:06

can do whatever you want you can train

play60:08

them in literally any way you want you

play60:10

can personalize them in arbitrary ways

play60:13

you could say okay for this user you do

play60:14

this for that user you do that and the

play60:16

user can be specified with some with the

play60:19

paragraph or maybe with some of the past

play60:22

actions

play60:24

so almost anything is

play60:28

possible now when you say almost

play60:30

anything is possible that also reminds

play60:32

me of a lot of our past conversations

play60:36

always seems like you know no no limits

play60:38

to your imagination of what might be

play60:40

possible and and you know angles to to

play60:42

try to get there and maybe maybe one of

play60:45

the other most surprising recent results

play60:48

is um you know traditionally a lot of

play60:50

work in computer vision in language

play60:53

processing in reinforcement learning

play60:56

kind of SE separate research Arenas

play61:00

almost but then uh recently you together

play61:04

with collaborators at op released the

play61:06

clip and Dolly models that bring

play61:09

language and vision in some sense

play61:12

together in into the same network to to

play61:14

Really

play61:15

somehow have a single Network that can

play61:18

handle both at the same time I'm kind of

play61:21

again I'm curious about you know H how

play61:24

how do you come

play61:26

to conclude okay this is the direction

play61:30

that maybe we should push now maybe it

play61:31

becomes possible now to have this

play61:34

combined model that can handle both

play61:36

vision and language in the same model

play61:39

and effectively translate between them

play61:42

as desired well I think the the

play61:44

underlying motivation here is that it

play61:47

seems

play61:49

implausible that the neural networks of

play61:51

the future will not have both vision and

play61:55

language

play61:57

and that was the motivation to begin

play61:59

thinking in this

play62:02

direction and as to

play62:05

whether this should be possible I mean I

play62:08

think I think at least in my view there

play62:10

was plenty of evidence

play62:12

that neural networks who just succeeded

play62:15

this task if you make it large and you

play62:17

have an appropriate data set if they can

play62:21

generate language like they do why can't

play62:23

they generate the language of images or

play62:26

going in the other direction as well so

play62:28

it was

play62:30

more maybe it's maybe it's good to think

play62:32

of it as of an exploration of training

play62:34

neural networks in both images and text

play62:36

and with di for context di is literally

play62:39

a

play62:40

gpt3 that is strained on text followed

play62:46

by almost like a textual representation

play62:48

of an image so we use those tokens to

play62:51

represent an image so that from the

play62:53

perspective of the model it's just some

play62:55

kind of a fun

play62:57

language but it's kind of like you know

play62:59

you could you can train gpt2 on on on

play63:01

English text on French text it doesn't

play63:04

care so what if you just had a different

play63:06

language which had some human language

play63:09

and the language of images and that's Di

play63:12

and it worked exactly as you'd

play63:14

expect and it was still a lot of fun to

play63:17

see a neural network generate images

play63:19

like it did and with clip it was an

play63:21

exploration in the opposite direction

play63:23

which is can a neural network learn to

play63:26

see using a lot of loose natural

play63:29

language supervision can it learn a huge

play63:32

variety of visual context Concepts and

play63:34

can it do so in a way that's very robust

play63:38

so that you know and I think the

play63:39

robustness point is something which I

play63:41

think is you know it's also very

play63:43

flexible but I I think the robustness

play63:46

point is is especially important in my

play63:48

eyes and let me explain what I mean by

play63:51

robustness so there is one thing which I

play63:53

think is especially notable and un is

play63:56

fying in neural networks provision is

play63:59

that they make these mistakes that a

play64:01

human being would never make so we we

play64:04

spoke earlier about the image net data

play64:06

set and about training neural networks

play64:08

to recognize the images in this data set

play64:11

and you'd have neural Nets which achieve

play64:13

super human performance in this data set

play64:15

then you put it on your phone and start

play64:16

taking photos and you would make all

play64:18

these disappointing mistakes what's

play64:19

going on and then it turns out that

play64:22

what's really going on is that there are

play64:23

all kinds of peculiarities data set

play64:26

which are hard to notice if you don't

play64:28

pay close attention and so people have

play64:30

built all kinds of test sets with the

play64:33

same objects but from maybe unusual

play64:35

angles or in a different presentation

play64:38

for which the image of neural is just

play64:40

fil but the clip neural network it was

play64:43

trained on this vast and Loosely labeled

play64:49

data from the in of his text this neural

play64:51

network

play64:53

was able to do well

play64:56

on all these variants of image was much

play64:59

more robust to the presentation of the

play65:01

visual concept and I think this kind of

play65:03

robustness is very important because

play65:05

human beings are in in when it comes to

play65:08

our vision you know a third of our brain

play65:10

is dedicated to Vision our vision is

play65:12

unbelievably

play65:14

good

play65:18

and and I feel like this is a step

play65:21

towards making neural Nets a little bit

play65:23

more robust a little bit more

play65:27

neural Net who's capability is a little

play65:28

bit more in line with the capability of

play65:31

of our own Vision now you say image Net

play65:35

versus the clip data set um the clip

play65:39

data set is a lot larger how much larger

play65:41

is it that I mean what's the difference

play65:42

in size between those like hundreds of

play65:44

times larger it has it has open-ended

play65:46

categories because the categories are

play65:48

just free form text but it's really kind

play65:51

of the size but also the coverage and

play65:53

the variety you need the data set needs

play65:55

to be verse it needs to have a lot of

play65:57

stuff in if data set is narrow it will

play66:01

hurt the neural netwk when I look back

play66:03

at the last 10 well nineish years right

play66:07

since um since the image net

play66:09

breakthrough it seems

play66:12

like year after year there are new

play66:14

breakthroughs new capabilities that

play66:16

didn't exist before many of them thanks

play66:19

to you IIA and your

play66:21

collaborators and I'm kind of curious

play66:24

how do you kind of from looking back at

play66:27

the last nine years and then as you

play66:29

project forward know are there some

play66:33

things that you are particularly excited

play66:35

about that we can't get to today but

play66:36

you're hopeful that you know maybe

play66:38

become feasible in the next few years

play66:41

yeah so I'd say that there is a sense in

play66:44

which the Deep learning Saga is actually

play66:46

a lot older than the past nine years you

play66:49

know it's funny if you read if you read

play66:51

some of the statements made by Rosen

play66:53

blad I think in the 60s so the Rosen

play66:56

blad invented the perceptron which was

play66:59

the one of the first neural networks

play67:02

that could learn something interesting

play67:04

on a real computer it could learn some

play67:06

image classification and then Rosen blat

play67:08

went to onto the New York Times and he

play67:10

said you know one day a neural network

play67:12

will see and hear and translate and be

play67:15

conscious of itself and be your friend

play67:17

something something like

play67:19

this and he had he he was trying to

play67:24

raise money to build in increasingly

play67:25

larger computers and he had academic

play67:28

detractors who didn't like the way

play67:30

funding was misallocated in their mind

play67:32

and that let to the you know to the

play67:35

first major neural network

play67:37

winter and then I think now these ideas

play67:40

were kind of always there in the

play67:41

background just that the environment

play67:43

wasn't ready because you needed both the

play67:45

data and the compute then as soon as the

play67:47

data and the compute became ready you

play67:49

were able to jump on this opportunity

play67:51

and materialize the progress and I I

play67:53

fully expect that progress will continue

play67:55

I think that we will have far more

play67:57

capable neural networks I think that you

play68:00

know I don't want to be too specific

play68:01

about what I think like about what

play68:05

exactly may happen because it's hard to

play68:07

predict those things but I would say one

play68:09

thing which would be nice if is to see

play68:11

our new oics being even more reliable

play68:13

than they are being so reliable that you

play68:15

can really trust their output and when

play68:18

they don't know something they'll just

play68:19

tell you and maybe ask for

play68:22

clarification I think that would be

play68:24

quite impactful

play68:26

I think they'll be they will be taking a

play68:27

lot more action than they are right now

play68:29

I think our neuron networks is still

play68:30

quite inert and passive and they'll be

play68:33

much more useful the usefulness will

play68:34

continue to

play68:36

grow

play68:37

and I mean for sure I I'm totally

play68:40

certain that we will need some kind of

play68:43

new ideas even if those new ideas may

play68:45

have the form of looking at things

play68:48

differently from the way looking at them

play68:49

right now and I would argue that a lot

play68:52

of the major progress in deep learning

play68:54

has this formed well for example the

play68:56

most recent progress with on supervised

play68:57

learning like what what was what was

play68:59

done what what's different we just train

play69:01

larger language models but they existed

play69:03

in the past it just we realized that

play69:06

language models were were the right

play69:08

thing all along so I think there will be

play69:10

more realizations like this where things

play69:12

that are right in front of our

play69:14

noses are actually far more powerful and

play69:17

far more capable than we

play69:19

expected and yeah I do expect that the

play69:22

capability of these systems will

play69:23

continue to increase they will become

play69:26

increasingly more impactful in the world

play69:28

it will become a much greater topic of

play69:32

conversation I think that the product we

play69:35

will see unbelievable truly unbelievable

play69:37

applications incredible applications

play69:39

positive

play69:42

very like even transformative

play69:44

applications I think you know we could

play69:46

we could imagine lots of them with very

play69:48

powerful Ai and eventually I really do

play69:52

think that you'll be in a world where

play69:54

the AI does the work

play69:56

and we the

play69:58

people enjoy enjoy this work and we we

play70:01

use that work to to our to our benefit

play70:03

and enjoyment you know

play70:05

this part part of the reason open AI is

play70:08

a cap profit company where after we

play70:10

return our obligations to our investors

play70:13

we turn back into a nonprofit so that we

play70:15

could help materialize this Future

play70:17

Vision where you have this useful AI

play70:21

That's doing all the work and all the

play70:24

people get to enjoy it

play70:27

and and that's really beautiful I I like

play70:30

the model you have there um because it

play70:33

essentially I mean it reflects

play70:37

the in some sense the vision that the

play70:39

benefits of you know really capable AR

play70:42

could be unlimited and it's not great to

play70:45

concentrate an unlimited benefit into a

play70:47

very small group of people because I

play70:51

mean that's just not not great for the

play70:52

rest of the world um so love the model

play70:56

you have there one of the things that

play70:58

ties into this Zia is um that maybe AI

play71:03

is also becoming more expensive a lot of

play71:05

people talk about it that you know

play71:07

training

play71:08

models um you want a bigger model is

play71:11

going to be more capable but then you

play71:13

know you need the resources to train

play71:16

those bigger models and I'm really

play71:17

curious about your thinking on that you

play71:20

know is is it just going to be you know

play71:22

the more money the bigger the model the

play71:25

more capable or is it possible that the

play71:27

future is different so so there is there

play71:30

is a huge amount of incentive to

play71:32

increase the efficiency of our models

play71:34

and to find ways to do more with less

play71:37

and this incentive is very strong and it

play71:40

affects everyone in the field and I

play71:43

fully expect that in the future we'll be

play71:45

able to do much more using a fraction of

play71:49

the cost that we do right now I think

play71:52

that's just going to happen for sure I

play71:53

think costing hard will drop I think

play71:56

methods will become more efficient in

play71:57

all sorts of ways there are multiple

play72:00

dimensions of efficiency that a models

play72:02

could utilize that they

play72:04

aren't at the same time I also think

play72:07

that it is true that bigger models will

play72:09

always be better

play72:11

and I think it's just a fact of life and

play72:15

I expect there should be almost like a

play72:17

kind of a power law of different models

play72:19

doing different things I think you'll

play72:21

have very powerful models in small

play72:23

numbers that are us for certain tasks

play72:26

then you'd have many more smaller models

play72:28

that are still hugely useful but and

play72:30

then you have even more models which are

play72:31

smaller and more specialized so you have

play72:33

this kind of Continuum of size

play72:36

specialization and it's going to be an

play72:39

ecosystem it's going to be not unlike

play72:42

how in nature there are animals that

play72:45

occupy any Niche and so I expect that

play72:46

the same thing will happen with comput

play72:48

that for every level of compute there

play72:50

will be some optimal way of using it and

play72:53

people will find that way and

play72:57

create very interesting applications

play73:00

love your vision IIA

play73:03

um I think we we actually covered a

play73:07

tremendous amount already and I'm really

play73:10

intrigued by everything we cover but

play73:11

there there is there's one question

play73:13

that's really still on my mind that I'm

play73:15

I'm hoping we can uh we can get through

play73:18

which is um

play73:21

Helia you've been behind a lot of the

play73:24

the breakthroughs in AI in the last 10

play73:27

years even actually even a bit before

play73:29

that um

play73:31

and I'm just kind of

play73:34

curious what what does your day look

play73:37

like what do you think are some habits

play73:40

and things in your schedule or or things

play73:42

you do that help you be creative and

play73:46

productive it's hard to give useful

play73:48

blanket advice like this but maybe two

play73:51

two answers consist of protecting my

play73:54

time and just trying really

play73:57

hard you know I don't think I don't

play73:59

think there is an easy way you need to

play74:01

just just got to embrace the

play74:04

suffering and and push through it and

play74:07

that's and and push through those walls

play74:09

and that's where the good stuff is

play74:11

found now when you say protecting your

play74:13

time which which really resonates of

play74:16

course

play74:18

um then you get to choose how you fill

play74:21

it in and I'm kind of curious if you

play74:22

just look at let's say maybe you know

play74:24

the last week or or the week before and

play74:27

they're like protected time you know

play74:30

what are you doing are you going on

play74:31

walks are you reading papers are you

play74:33

brainstorming with people what what's

play74:35

going on yeah I'd say I'd say mostly in

play74:39

my case it would

play74:40

be not necessarily going in works but

play74:43

lots of solitary work and yeah there are

play74:45

people with whom I have very intense

play74:47

research conversations which are very

play74:51

important and I think those are those

play74:54

are the main things I do

play74:56

I do know that you're also an artist or

play74:58

you know aspiring artist whatever we

play75:01

want to call it at the same time do you

play75:03

think that plays a role at all in in

play75:04

boosting your

play75:06

creativity I mean I'm sure it doesn't

play75:10

hurt so now it's hard hard to know with

play75:12

these things obviously

play75:16

but yeah I I think it can only help

Rate This

5.0 / 5 (0 votes)

Étiquettes Connexes
人工智能深度学习神经网络技术创新机器学习自然语言处理图像识别AI研究算法突破未来趋势
Besoin d'un résumé en anglais ?