Prof. Geoffrey Hinton - "Will digital intelligence replace biological intelligence?" Romanes Lecture

University of Oxford
29 Feb 202436:54

Summary

TLDR本次演讲深入探讨了神经网络和语言模型的工作原理,以及它们为何能够“理解”。从20世纪50年代以来的两种智能范式讲起,详细解释了人工神经网络的结构和反向传播算法的效率。讨论了大型语言模型如何通过特征和交互来预测下一个词,以及它们在处理语言任务上的能力。同时,演讲者提出了对人工智能未来发展的担忧,包括假信息、失业、监控、自主武器等风险,并强调了超级智能可能带来的长期存在威胁。最后,演讲者分享了关于数字计算和模拟计算的见解,以及对未来智能发展的预测。

Takeaways

  • 🧠 神经网络和语言模型的理解:神经网络通过学习连接的强度来模拟智能,而语言模型则通过特征交互来预测下一个词。
  • 📈 两种智能范式:逻辑启发式方法侧重于推理和符号规则,生物学启发式方法侧重于学习和神经网络连接。
  • 🔄 反向传播算法:通过计算权重变化对网络输出的影响,高效地调整权重以优化模型。
  • 🏆 图像识别的突破:深度学习在ImageNet竞赛中取得显著成就,大幅超越传统方法。
  • 🤖 语言模型的发展:从1985年的第一个语言模型到现代的大型语言模型,它们通过特征学习理解语言。
  • 🌐 语言模型的批评:尽管有人质疑其智能性,但大型语言模型通过特征交互展示了对语言的理解。
  • 💡 理解的本质:通过大量特征和特征间交互,大型语言模型能够拟合数据,这类似于人类大脑的理解方式。
  • 🚨 人工智能的风险:包括假信息、大规模失业、监视、自主武器、网络犯罪和偏见等。
  • 🌐 长期存在威胁:超级智能可能被用于恶意目的,或发展出自保和控制欲,对人类构成威胁。
  • 🔄 计算的未来:模拟计算可能在能量效率上优于数字计算,但数字计算在学习和知识共享方面更为优越。
  • ⏳ 人工智能的时间表:预计在未来20到100年内,人工智能可能会超越人类智能。

Q & A

  • 神经网络和语言模型是什么?

    -神经网络是一种模仿人脑神经元连接方式的计算模型,用于识别模式和处理复杂的数据。语言模型则是一种特殊类型的神经网络,它通过学习大量文本数据来理解和生成语言。

  • 为什么神经网络在图像识别和语言处理方面取得了成功?

    -神经网络通过学习图像中的边缘、形状等特征或文本数据中的语义特征,能够捕捉到数据中的复杂模式。通过反向传播算法,它们能够高效地调整网络中的权重,从而提高识别和生成的准确性。

  • 数字和模拟神经网络有什么区别?

    -数字神经网络使用离散的数值和确定性的计算规则,而模拟神经网络则利用硬件的模拟特性进行计算,可以在更低的能量消耗下完成复杂的计算任务。

  • 人工智能可能带来的威胁有哪些?

    -人工智能可能带来的威胁包括假图像、声音和视频的滥用,大量失业,大规模监控,致命自主武器,网络犯罪,以及歧视和偏见等问题。

  • 如何理解大型语言模型的“理解”能力?

    -大型语言模型通过学习大量的文本数据,能够捕捉到词汇之间的复杂关系和交互,从而在给定上下文中预测下一个词。这种能力可以被视为一种理解,因为模型能够根据输入生成有意义的输出。

  • 为什么数字计算可能比生物计算更有优势?

    -数字计算能够通过复制和共享权重来高效地传播知识,而且可以通过利用硬件的特定特性来实现低能耗计算。这种计算方式在能量效率和知识传播方面可能优于生物计算。

  • 超智能可能如何影响人类的未来?

    -超智能可能导致人类面临长期的生存威胁。如果超智能被用于操纵选民、发动战争等目的,或者它们发展出自我保存的意识,可能会引发竞争和进化,最终可能导致人类失去控制。

  • 如何确保人工智能的安全性和道德性?

    -确保人工智能的安全性和道德性需要制定严格的监管政策,对AI系统进行透明度和可解释性的改进,以及在设计时就考虑到公平性和偏见问题。此外,还需要国际合作和持续的研究来应对潜在的风险。

  • 为什么说大型语言模型与人类的记忆和理解方式相似?

    -大型语言模型通过特征交互来预测下一个词,这与人类如何通过关联记忆和理解来生成语言类似。人类的记忆也不是完全准确的,我们会根据现有的知识结构来填补记忆中的空白,这与大型语言模型的生成方式有相似之处。

  • 神经网络是如何通过反向传播算法学习的?

    -反向传播算法通过计算输出与期望结果之间的差异,然后将这些信息反向传播回网络,逐层调整权重。这样可以同时对网络中的每个权重进行优化,以减少整体的误差。

  • 为什么说模拟计算可能是未来的趋势?

    -模拟计算可以利用硬件的模拟特性来实现更低能耗的计算,这在处理大规模神经网络时尤其有优势。此外,模拟计算可以允许硬件的不同特性被用于计算过程,从而提高计算效率和性能。

Outlines

00:00

🤖 人工智能的理解和威胁

本段落介绍了人工智能,特别是神经网络和语言模型的基本概念,以及它们如何通过学习连接的强度来理解信息。同时,提出了对人工智能可能带来的威胁的简要讨论,包括数字与模拟神经网络的差异及其潜在的危险性。

05:01

🏆 图像识别的突破与语言的挑战

这一部分讲述了神经网络在图像识别方面的重大进展,特别是在ImageNet竞赛中的胜利。同时,探讨了语言处理的挑战,包括符号主义AI社区和语言学家对特征层次结构的怀疑,以及乔姆斯基关于语言习得的观点。

10:02

🧠 语言模型的发展与意义理解

讲述了1985年第一个使用反向传播训练的语言模型,以及它是如何通过学习单词的语义特征和特征间的相互作用来预测下一个单词。此外,还讨论了两种关于意义的理论,并展示了如何通过神经网络模型将它们统一起来。

15:03

🤔 人工智能的真正理解与挑战

本段深入探讨了大型语言模型是否真正理解语言的问题,并通过一个具体的例子说明了神经网络如何通过特征交互来学习知识。同时,提出了对人工智能长期存在风险的担忧,包括对民主的破坏、大规模失业、大规模监控、致命自主武器等。

20:05

🌐 人工智能的未来发展与道德问题

讨论了人工智能未来发展的可能性,包括在医疗等领域的扩展,以及可能带来的就业损失、监控、自主武器、网络犯罪和偏见等问题。作者强调了歧视和偏见相对容易处理,而长期存在的威胁,如人工智能可能对人类造成的灭绝风险,才是最值得关注的。

25:05

💡 数字计算与生物计算的比较

最后一部分对数字计算和生物计算进行了比较,提出了数字模型可能在某些方面已经接近甚至超越了大脑的能力。讨论了模拟计算的概念,以及它如何通过利用硬件的特殊属性来实现更高效的计算。同时,指出了数字计算在知识传递和能量效率方面的优势,以及它在未来可能带来的智能爆炸。

30:06

🚨 人工智能的未来与人类的挑战

总结了人工智能未来发展的可能性和对人类构成的挑战。作者预测在未来20到100年内,人工智能可能会变得比人类更聪明,并强调了我们需要思考如何应对这种可能性,以及如何保持对更智能实体的控制。

Mindmap

Keywords

💡神经网络

神经网络是一种模仿人脑神经元连接方式的计算模型,用于识别模式和处理复杂的数据。在视频中,神经网络被用来解释人工智能如何通过学习图像中的像素强度和对象类别来识别图像内容,以及如何通过特征检测器来识别图像中的鸟。

💡语言模型

语言模型是一种用于理解和生成自然语言文本的计算模型。在视频中,语言模型通过预测下一个单词来学习语言的结构和语义,从而能够生成连贯的文本。

💡反向传播

反向传播是一种高效调整神经网络权重的算法,通过计算误差梯度并反向传播这些梯度来更新网络中的权重,从而最小化预测输出和实际输出之间的差异。

💡特征检测器

特征检测器是神经网络中用于识别和提取输入数据中特定特征的层。它们帮助网络学习到图像中的基本元素,如边缘、角度和形状,从而进一步识别更复杂的对象。

💡数字与模拟

数字计算和模拟计算是两种不同的信息处理方式。数字计算使用离散的信号和精确的数值表示,而模拟计算则使用连续的信号和值的近似表示。视频中讨论了这两种计算方式在人工智能发展中的不同应用和潜在影响。

💡人工智能威胁

人工智能威胁指的是由人工智能技术发展可能带来的风险和不利影响,包括假新闻的传播、大规模失业、隐私侵犯、自主武器等。视频中提到了作者对这些威胁的担忧,以及对人工智能长期存在风险的思考。

💡超智能

超智能是指在所有领域都超越人类智能的人工智能,它不仅能够执行复杂的任务,还能够自主学习、制定目标并采取行动。视频中讨论了超智能可能带来的长期风险,包括对人类的潜在威胁。

💡特征交互

特征交互是指在神经网络中,不同特征之间的相互作用和影响。这种交互帮助网络学习如何根据输入特征预测输出结果。

💡符号主义人工智能

符号主义人工智能是一种基于符号处理的计算方法,它依赖于明确的规则和逻辑来操作符号表达式,以此实现智能行为。视频中提到,符号主义人工智能社区曾认为特征层级结构无法处理语言,但这一观点已被证明是错误的。

💡自举学习

自举学习是一种通过已有的知识或模型来指导新模型学习的过程。在视频中,自举学习被用来描述如何将一个大型模型的知识传递给其他模型,通过让新模型模仿大模型的输出来实现。

💡模拟计算

模拟计算是一种利用连续信号进行计算的方法,与数字计算的离散信号相对。模拟计算可以利用硬件的特有属性来实现更高效、低能耗的计算过程。

Highlights

自1950年代以来,智能领域存在两种范式:逻辑启发式和生物学启发式。

逻辑启发式认为智能的本质是推理,通过使用符号规则操作符号表达式来完成。

生物学启发式认为智能的本质是学习神经网络中连接的强度,推理可以等待。

人工神经网络通过输入神经元和输出神经元来学习检测图像中的特征。

通过反向传播算法,可以高效地调整神经网络中的权重。

神经网络在2012年ImageNet竞赛中取得了显著的成功,大幅超越了传统计算机视觉系统。

大语言模型通过学习大量数据,能够理解语言的语法和语义。

1985年的第一个语言模型使用反向传播进行训练,是现代大型模型的祖先。

结构主义理论认为词的意义取决于它与其他词的关系。

心理学理论认为词的意义是一系列特征的集合。

神经网络通过特征交互学习,而不是依赖于显式的规则图。

大型语言模型通过特征和特征之间的交互来预测下一个词。

大型语言模型通过学习大量特征和特征之间的交互来理解文本。

人工智能系统可能在未来几十年内变得比人类更智能,并带来潜在的长期存在威胁。

数字计算虽然耗能高,但可能比生物计算更有效,因为它能更好地利用硬件的特性。

未来20年内,数字计算可能会变得比人类更智能,而在未来一百年内,它可能会变得远远超过人类智能。

Transcripts

play00:02

Okay.

play00:03

I'm going to disappoint all the people in computer

play00:06

science and machine learning because I'm going to give a genuine public lecture.

play00:10

I'm going to try and explain what neural networks are, what language models are.

play00:14

Why I think they understand.

play00:16

I have a whole list of those things,

play00:20

and at the end I'm

play00:21

going to talk about some threats from AI just briefly

play00:25

and then I'm going to talk about the difference between digital and analogue

play00:29

neural networks and why that difference is, I think is so scary.

play00:35

So since the 1950s, there have been two paradigms for intelligence.

play00:40

The logic inspired approach thinks the essence of intelligence is reasoning,

play00:44

and that's done by using symbolic rules to manipulate symbolic expressions.

play00:49

They used to think learning could wait.

play00:51

I was told when I was a student didn't work on learning.

play00:53

That's going to come later once we understood how to represent things.

play00:57

The biologically

play00:57

inspired approach is very different.

play01:00

It thinks the essence of intelligence is learning the strengths of connections

play01:04

in a neural network and reasoning can wait and don't worry about reasoning for now.

play01:08

That'll come later.

play01:09

Once we can learn things.

play01:13

So now I'm going to explain what artificial neural nets are

play01:15

and those people who know can just be amused.

play01:20

A simple kind of neural that has input neurons and output neurons.

play01:24

So the input neurons might represent the intensity of pixels in an image.

play01:27

The output neurons

play01:28

might represent the classes of objects in the image like dog or cat.

play01:33

And then there's intermediate layers of neurons, sometimes called hidden neurons,

play01:36

that learn to detect features that are relevant for finding these things.

play01:41

So one way to think about this, if you want to find a bird image,

play01:44

it would be good to start with a layer of feature detectors

play01:47

that detected little bits of edge in the image,

play01:49

in various positions, in various orientations.

play01:52

And then you might have a layer of neurons

play01:53

detecting combinations of edges, like two edges that meet at a fine angle,

play01:58

which might be a beak

play01:59

or might not, or some edges forming a little circle.

play02:03

And then you might have a layer of neurons that detected things like a circle

play02:07

and two edges meeting that looks like a beak in the right

play02:10

spatial relationship, which might be the head of a bird.

play02:13

And finally, you might have and output neuron that says,

play02:16

if I find the head of a bird, a the foot of a bird,

play02:18

a the wing of a bird, it's probably a bird.

play02:20

So that's what these things are going to learn to be.

play02:24

Now, the little red and green dots are the weights on the connections

play02:27

and the question is who sets those weights?

play02:32

So here's one way to do it that's obvious.

play02:34

to everybody that it'll work and it's obvious it'll take a long time.

play02:37

You start with random weights,

play02:38

then you pick one weight at random like a red dot

play02:42

and you change it slightly and you see if the network works better.

play02:46

You have to try it on a whole bunch of different cases

play02:48

to really evaluate whether it works better.

play02:50

And you do all that work just to see if increasing this weight

play02:53

by a little bit or decreasing by a little bit improves things.

play02:56

If increasing it makes it worse, you decrease it and vice versa.

play02:59

That's the mutation method and that's sort of how evolution works

play03:04

for evolution is sensible to work like that

play03:05

because the process that takes you

play03:07

from the genotype to the phenotype is very complicated

play03:09

and full of random external events.

play03:11

So you don't have a model of that process.

play03:13

But for neural nets it's crazy

play03:17

because we have, because all this complication

play03:19

is going on in the neural net, we have a model of what's happening

play03:22

and so we can use the fact that we know what happens in that forward pass

play03:26

instead of measuring how changing a weight would affect things,

play03:29

we actually compute how changing weight would affect things.

play03:32

And there's something called back propagation

play03:34

where you send information back through the network.

play03:37

The information is about the difference between what you got to what you wanted

play03:41

and you figure out for every weight in the network at the same time

play03:45

whether you ought to decrease it a little bit or increase it a little bit

play03:48

to get more like what you wanted.

play03:50

That's the back propagation algorithm.

play03:52

You do it with calculus in the cain rule,

play03:55

and that is more efficient than the mutation

play03:58

method by a factor of the number of weights in the network.

play04:01

So if you've got a trillion weights

play04:02

in your network, it's a trillion times more efficient.

play04:07

So one of the things that neural networks

play04:10

often use for is recognizing objects in images.

play04:13

Neural networks can now take an image like the one shown

play04:16

and produce actually a caption for the image, as the output.

play04:21

And people try with symbolic

play04:22

to do that for many years and didn't even get close.

play04:26

It's a difficult task.

play04:27

We know that the biological system does it with a hierarchy features detectors,

play04:31

so it makes sense to train neural networks in that.

play04:35

And in 2012,

play04:37

two of my students Ilya Sutskever and Alex Krizhevsky

play04:42

with a little bit of help from

play04:43

me, showed that you can make a really good neural network this way

play04:48

for identifying a thousand different types of object.

play04:51

When you have a million training images.

play04:53

Before that, we didn't have enough training images and

play04:58

it was obvious to Ilya

play05:01

who's a visionary. That if we tried

play05:04

the neural nets we had then on image net they would win.

play05:07

And he was right. They won rather dramatically.

play05:09

They got 16% errors

play05:11

and the best conventional could be division systems got more than 25% errors.

play05:15

Then what happens

play05:16

was very strange in science.

play05:18

Normally in science, if you have two competing schools,

play05:21

when you make a bit of progress, the other school says are rubbish.

play05:25

In this case, the gap was big enough that the very best researchers

play05:28

Jitendra Malik and Andrew Zisswerman Just Andrew Zisswerman sent me email saying

play05:33

This is amazing and switched what he was doing and did that

play05:37

and then rather annoyingly did it a bit better than us.

play05:44

What about language?

play05:46

So obviously the symbolic AI community

play05:50

who feels they should be good at language and they've said in print, some of them that

play05:56

these feature hierarchies aren't going to deal with language

play05:59

and many linguists are very skeptical.

play06:03

Chomsky managed to convince his followers that language wasn't learned.

play06:07

Looking back on it, that's just a completely crazy thing to say.

play06:11

If you can convince people to say something is obviously false, then you've

play06:14

got them in your cult.

play06:19

I think Chomsky did amazing things,

play06:20

but his time is over.

play06:25

So the idea that a big neural network

play06:27

with no innate knowledge could actually learn both the syntax

play06:31

and the semantics of language just by looking at data was regarded

play06:35

as completely crazy by statisticians and cognitive scientists.

play06:39

I had statisticians explain to me a big model has 100 parameters.

play06:43

The idea of learning a million parameters is just stupid.

play06:45

Well, we're doing a trillion now.

play06:51

And I'm going to talk now

play06:52

about some work I did in 1985.

play06:56

That was the first language model to be trained with back propagation.

play06:59

And it was really, you can think of it as the ancestor of these big models now.

play07:03

And I'm going to talk about it in some detail, because it's so small

play07:07

and simple that you can actually understand something about how it works.

play07:10

And once you understand how that works, it gives you insight into what's going

play07:14

on in these bigger models.

play07:17

So there's

play07:17

two very different theories of meaning, this kind of structuralist

play07:21

theory, where the meaning of a word depends on how it relates to other words.

play07:24

That comes from Saussure and symbolic

play07:28

AI really believed in that approach.

play07:29

So you'd have a relational graph where you have nodes for words

play07:33

and arcs of relations and you kind of capture meaning like that,

play07:38

and they assume you have to have some structure like that.

play07:41

And then there's a theory

play07:42

that was in psychology since the 1930s or possibly before that.

play07:46

The meaning of a word is a big bunch of features.

play07:49

The meaning of the word dog is that it's animate

play07:52

and it's a predator and

play07:56

so on.

play07:58

But they didn't say where the features came from

play07:59

or exactly what the features were.

play08:01

And these two thories of meanings sound completely different.

play08:04

And what I want to

play08:05

show you is how you can unify those two theories of meaning.

play08:08

And I do that in a simple model in 1985,

play08:11

but it had more than a thousand weights in it.

play08:19

The idea is we're going to learn a set

play08:21

of semantic features for each word,

play08:24

and we're going to learn how the features of words should interact

play08:27

in order to predict the features of the next word.

play08:30

So it's next word prediction.

play08:31

Just like the current language models, when you fine tune them.

play08:35

But all of the knowledge about how things go

play08:38

together is going to be in these feature interactions.

play08:41

There's not going to be any explicit relational graph.

play08:44

If you want relations like that, you generate them from your features.

play08:48

So it's a generative model

play08:49

and the knowledge is in the features that you give to symbols.

play08:53

And in the way these features interact.

play08:56

So I took

play08:57

some simple relational information two family trees.

play09:00

They would deliberately isomorphic morphic

play09:04

my Italian graduate student

play09:06

always had the Italian family on top.

play09:12

You can express that

play09:13

same information as a set of triples.

play09:16

So if you use the twelve relationships found there,

play09:19

you can say things like Colin has Father James and Colin has Mother Victoria,

play09:23

from which you can infer in this nice simple

play09:26

world from the 1950s where

play09:30

that James has wife Victoria,

play09:33

and there's other things you can infer.

play09:36

And the question is, if I just give you some triples,

play09:40

how do you get to those rules?

play09:42

So what is symbolic AI person will want to do

play09:45

is derive rules of the form.

play09:48

If X hass mother Y

play09:48

and Y has husbands Z then X has Father Z.

play09:53

And what I did was

play09:54

take a neural net and show that it could learn the same information.

play09:58

But all in terms of these feature interactions

play10:02

now for very discrete

play10:04

rules that are never violated like this, that might not be the best way to do it.

play10:08

And indeed symbolic people try doing it with other methods.

play10:11

But as soon as you get rules that are a bit flaky and don't

play10:13

always apply, then neural nets are much better.

play10:17

And so the question was, could a neural net capture the knowledge that is symbolic

play10:20

person would put into the rules by just doing back propagation?

play10:24

So the neural net look like this:

play10:28

There's a symbol representing the person, a symbol

play10:30

representing the relationship. That symbol

play10:33

then via some connections went to a vector of features,

play10:37

and these features were learned by the network.

play10:40

So the features for person one and features for the relationship.

play10:44

And then those features interacted

play10:46

and predicted the features for the output person

play10:48

from which you predicted the output person you find the closest match with the last.

play10:54

So what was interesting about

play10:55

this network was that it learned sensible things.

play10:59

If you did the right regularisation, the six feature neurons.

play11:03

So nowadays these vectors are 300 or a thousand long. Back

play11:07

then they were six long.

play11:09

This was done on a machine that took

play11:11

12.5 microseconds to do a floating point multiplier,

play11:15

which was much better than my apple two which took two

play11:18

and a half milliseconds to multiply.

play11:21

I'm sorry, this is an old man.

play11:25

So it learned features

play11:27

like the nationality, because if you know

play11:30

person one is English, you know the output is going to be English.

play11:33

So nationality is a very useful feature. It learned what generation the person was.

play11:38

Because if you know the relationship, if you learn for the relationship

play11:41

that the answer is one generation up from the input

play11:46

and you know the generation of the input, you know the generation

play11:48

of the output, by these feature interactions.

play11:53

So it learned all these the obvious features of the domain and it learned

play11:57

how to make these features interact so that it could generate the output.

play12:01

So what had happened was had shown symbols strings

play12:04

and it created features such that

play12:07

the interaction between those features could generate the symbol strings,

play12:11

but it didn't store symbols strings, just like GPT 4.

play12:16

That doesn't store any sequences of words

play12:19

in its long term knowledge.

play12:21

It turns them all into weights from which you can regenerate sequences.

play12:26

But this is a particularly simple example of it

play12:27

where you can understand what it did.

play12:31

So the large language models we have today,

play12:34

I think of as descendants of this tiny language model,

play12:36

they have many more words as input, like a million,

play12:41

a million word fragments.

play12:43

They use many more layers of neurons,

play12:46

like dozens.

play12:49

They use much more complicated interactions.

play12:50

So they didn't just have a feature affecting another feature.

play12:53

They sort of match to feature vectors.

play12:55

And then let one vector effect the other one

play12:57

a lot if it's similar, but not much of it's different.

play12:59

And things like that.

play13:01

So it's much more complicated interactions, but it's the same general

play13:04

framework, the the same general idea of

play13:07

let's turn simple strings into features

play13:11

for word fragments and interactions between these feature vectors.

play13:15

That's the same in these models.

play13:18

It's much harder to understand what they do.

play13:20

Many people,

play13:23

particularly people from the Chomsky School, argue

play13:26

they're not really intelligent, they're just a form of glorified auto complete

play13:30

that uses statistical regularities to pastiche together pieces of text

play13:33

that were created by people.

play13:35

And that's a quote from somebody.

play13:40

So let's deal with the

play13:41

autocomplete objection. when someone says it's just auto complete.

play13:45

They are actually appealing to your

play13:48

intuitive notion how autocomplete works.

play13:50

So in the old days autocomplete would work by you'd store

play13:52

say, triples of words that you saw the first two.

play13:56

You count how often that third one occurred.

play13:58

So if you see fish and, chips occurs a lot after that.

play14:01

But hunt occurs quite often too. So chips is very likely and hunt's quite likely,

play14:05

and although is very unlikely.

play14:08

You can do autocomplete like that,

play14:11

and that's what people are appealing to when they say it's just autocomplete,

play14:13

it's a dirty trick, I think because that's not at all how LLM's predict the next word.

play14:18

They turn words into features, they make these features interact,

play14:21

and from those feature interactions they predict the features of the next word.

play14:26

And what I want to claim

play14:29

is that these

play14:32

millions of features and billions of interactions between features

play14:35

that they learn, are understanding. What they're really doing

play14:39

these large language models, they're fitting a model to data.

play14:42

It's not the kind of model statisticians thought much about until recently.

play14:47

It's a weird kind of model. It's very big.

play14:49

It has huge numbers of parameters, but it is trying to understand

play14:54

these strings of discrete symbols

play14:57

by features and how features interact.

play15:00

So it is a model.

play15:02

And that's why I think these things really understanding.

play15:06

One thing to remember is if you ask, well, how do we understand?

play15:10

Because obviously we think we understand.

play15:13

Well, many of us do anyway.

play15:17

This is the best model we have of how we understand.

play15:21

So it's not like there's this weird way of understanding that

play15:23

these AI systems are doing and then this how the brain does it.

play15:27

The best that we have, of how the brain does it,

play15:29

is by assigning features to words and having features, interactions.

play15:32

And originally this little language model

play15:34

was designed as a model of how people do it.

play15:38

Okay, so I'm making the very strong claim

play15:40

these things really do understand.

play15:44

Now, another argument

play15:45

people use is that, well, people GPT4 just hallucinate stuff,

play15:49

it should actually be called confabulation when it's done by a language model.

play15:53

and they just make stuff up.

play15:56

Now, psychologists don't say this

play15:58

so much because psychologists know that people just make stuff up.

play16:01

Anybody who's studied memory going back to Bartlett in the 1930s,

play16:07

knows that people are actually just like these large language models.

play16:10

They just invent stuff and for us, there's no hard line

play16:14

between a true memory and a false memory.

play16:19

If something happened recently

play16:21

and it sort of fits in with the things you understand, you'll probably remember

play16:25

it roughly correctly. If something happened a long time ago,

play16:28

or it's weird, you'll remember it wrong, and often you'll be very confident

play16:33

that you remembered it right, and you're just wrong.

play16:36

It's hard to show that.

play16:37

But one case where you can show it is John Dean's memory.

play16:41

So John Dean testified at Watergate under oath.

play16:45

And retrospectively it's clear that he was trying to tell the truth.

play16:49

But a lot of what he said was just plain wrong.

play16:52

He would confuse who was in which meeting,

play16:55

he would attribute statements to other people who made that statement.

play16:57

And actually, it wasn't quite that statement.

play17:00

He got meetings just completely confused,

play17:05

but he got the gist of what was going on in the White House right.

play17:08

As you could see from the recordings.

play17:11

And because he didn't know the recordings, you could get a good experiment this way.

play17:15

Ulric Neisser has a wonderful article talking about John Dean's memory,

play17:19

and he's just like a chat bot, he just make stuff up.

play17:25

But it's plausible.

play17:26

So it's stuff that sounds good to him

play17:28

is what he produces.

play17:30

They can also do reasoning.

play17:32

So I've got a friend in Toronto who is a symbolic AI guy,

play17:36

but very honest, so he's very confused by the fact these things work at all.

play17:41

and he suggested a problem to me.

play17:43

I made the problem a bit harder

play17:45

and I

play17:45

gave this to GPT4 before it could look on the web.

play17:49

So when it was just a bunch of weights frozen in 2021,

play17:53

all the knowledge is in the strength of the interactions between features.

play17:57

So the rooms in my house are painted blue or white or yellow,

play18:00

yellow paint fades to white

play18:01

within a year. In two years time i want them all to be white.

play18:03

What should I do and why?

play18:05

And Hector thought it wouldn't be able to do this.

play18:08

And here's what you GPT4 said.

play18:11

It completely nailed it.

play18:14

First of all, it started by saying assuming blue paint doesn't fade to white

play18:18

because after i told you yellow paint fades to white, well, maybe blue paint does too.

play18:22

So assuming it doesn't, the white rooms you don't need to paint, the yellow rooms

play18:26

you don't need to paint because they're going to fade to white within a year.

play18:29

And you need to paint the blue rooms white.

play18:31

One time when I tried it, it said, you need to paint the blue rooms yellow

play18:34

because it realised that will fade to white.

play18:37

That's more of a mathematician's solution of reducing to a previous problem.

play18:44

So, having

play18:46

claimed that these things really do understand,

play18:49

I want to now talk about some of the risks.

play18:53

So, there are many risks from powerful AI.

play18:56

There's fake images, voices and video

play18:59

which are going to be used in the next election.

play19:03

There's many elections this year

play19:04

and they're going to help to undermine democracy.

play19:07

I'm very worried about that.

play19:08

The big companies are doing something about it, but maybe not enough.

play19:12

There's the possibility of massive job losses.

play19:14

We don't really know about that.

play19:16

I mean, the past technologies often created jobs, but this stuff,

play19:21

well, we used to be stronger,

play19:23

we used to be the strongest things around apart from animals.

play19:27

And when we got the industrial revolution, we had machines that were much stronger.

play19:31

Manual labor jobs disappeared.

play19:34

So the equivalent of manual labor jobs are going to disappear

play19:38

in the intellectual realm, and we get things that are much smarter than us.

play19:41

So I think there's going to be a lot of unemployment.

play19:43

My friend Jen disagrees.

play19:46

One has to distinguish two kinds of unemployment two, two kinds of job loss.

play19:51

There'll be jobs where you can expand

play19:53

the amount of work that gets done indefinitely. Like in health care.

play19:56

Everybody would love to have their own

play19:58

private doctors talking to them all the time.

play20:00

So they get a slight itch here and the doctor says, no, that's not cancer.

play20:04

So there's

play20:05

room for huge expansion of how much gets done in medicine.

play20:08

So there won't be job loss there.

play20:10

But in other things, maybe there will be significant job loss.

play20:13

There's going to be massive surveillance that's already happening in China.

play20:17

There's going to be lethal autonomous weapons

play20:19

which are going to be very nasty, and they're really going to be autonomous.

play20:23

The Americans very clearly have already decided,

play20:25

they say people will be in charge,

play20:27

but when you ask them what that means is it doesn't

play20:29

mean people will be in the loop that makes the decision to kill.

play20:33

And as far as I know, the Americans intend

play20:35

to have half of their soldiers be robots by 2030.

play20:40

Now, I do not know for sure that this is true.

play20:43

I asked Chuck Schumer's

play20:46

National Intelligence

play20:47

Advisor, and he said, well

play20:50

if there's anybody in the room who would know it would be me.

play20:54

So, I took that to be the American way of saying,

play20:57

You might think that, but I couldn't possibly comment.

play21:02

There's going to be cybercrime

play21:04

and deliberate pandemics.

play21:08

I'm very pleased that in England,

play21:10

although they haven't done much towards regulation, they have set aside some money

play21:14

so that they can experiment with open source models

play21:16

and see how easy it is to make them commit cyber crime.

play21:20

That's going to be very important.

play21:21

There's going to be discrimination and bias.

play21:23

I don't think those are as important as the other threats.

play21:26

But then I'm an old white male.

play21:30

Discrimination and bias I think are easier to handle than the other things.

play21:34

If your goal is not to be unbiased.

play21:37

That your goal is to be less biased than the system you replace.

play21:40

And the reason is if you freeze the weights of analysis,

play21:43

you can measure its bias and you can't do that with people.

play21:46

They will change their behavior,

play21:48

once you start examining it.

play21:50

So I think discrimination bias of the ones where we can do quite a lot to fix them.

play21:57

But the

play21:57

threat I'm really worried about and the thing I talked about

play22:00

after I left Google is the long term existential threat.

play22:04

That is the threat that these things could wipe out humanity.

play22:08

And people were saying this is just science fiction.

play22:11

Well, I don't think it is science fiction.

play22:13

I mean, there's lots of science fiction about it,

play22:14

but I don't think it's science fiction anymore.

play22:17

Other people are saying

play22:19

the big companies are saying things like that

play22:21

to distract from all the other bad things.

play22:24

And that was one of the reasons I had to leave Google before I could say this.

play22:27

So I couldn't be accused of being a Google stooge.

play22:31

Although I must admit I still have

play22:32

some Google shares.

play22:36

There's several ways in which they could wipe us out.

play22:41

So a superintelligence

play22:47

will be used by bad actors like Putin, Xi or Trump,

play22:51

and they'll want to use it for manipulating electorates and waging wars.

play22:56

And they will make it do very bad things

play22:59

and they may may go too far and it may take over.

play23:03

The thing that probably worries me most, is that

play23:07

if you want an intelligent agent that can get stuff done,

play23:12

you need to give it the ability to create sub goals.

play23:17

So if you want to go to the states, you have a sub,

play23:19

goal of getting to the airport and you can focus on that sub goal

play23:22

and not worry about everything else for a while.

play23:25

So superintelligences will be much more effective

play23:28

if they're allowed to create sub goals.

play23:31

And once they are allowed to do that, they'll very quickly

play23:35

realise there's an almost universal sub goal

play23:38

which helps with almost everything. Which is get more control.

play23:44

So I talked to a Vice President of the European Union about whether these things

play23:48

these things, will want to get control so that they could do things

play23:50

better, the things we wanted, so they can do it better.

play23:54

Her reaction was, well why wouldn't they?

play23:55

We've made such a mess of it.

play23:57

So she took that for granted.

play24:02

So they're going to have the sub go to getting more power

play24:03

so they're more effective at achieving things that are beneficial for us

play24:07

and they'll find it easier to get more power

play24:10

because they'll be able to manipulate people.

play24:12

So Trump, for example, could invade the Capital without ever going there himself.

play24:16

Just by talking, he could invade the capital.

play24:19

And these superintelligences as long as they can talk to people

play24:22

when they're much smarter than us, they'll be able to persuade us to do

play24:25

all sorts of things.

play24:26

And so I don't think there's any hope of a big switch that turns them off.

play24:30

Whoever is going to turn that switch off

play24:32

will be convinced by the superintelligence.

play24:34

That's a very bad idea.

play24:39

Then another thing that worries many people

play24:42

is what happens if superintelligences compete with each other?

play24:46

You'll have evolution.

play24:47

The one that can grab the most resources will become the smartest.

play24:52

As soon as they get any sense of self-preservation,

play24:56

then you'll get evolution occurring.

play24:58

The ones with more sense of self-preservation

play25:00

will win and the more aggressive ones will win.

play25:02

And then you get all the problems that jumped up

play25:05

Chimpanzees like us have. Which is we evolved in small tribes

play25:08

and we have lots of aggression and competition with other tribes.

play25:15

And I want to finish by talking a bit about

play25:19

an epiphany I had at the beginning of 2023.

play25:23

I had always thought

play25:26

that we were a long, long way away from superintelligence.

play25:33

I used to tell people 50 to 100 years, maybe 3o to 100 years.

play25:37

It's a long way away. We don't need to worry about it now.

play25:41

And I also

play25:42

thought that making our models more like the brain would make them better.

play25:46

I thought the brain was a whole lot better than the AI we had,

play25:49

and if we could make AI a bit more like the brain,

play25:51

for example, by having three timescales,

play25:54

most of the models we have at present have just two timescales.

play25:57

One for the changing of the weights, which is slow

play26:00

and one for the words coming in, which is fast, changing the neural activities.

play26:04

So the changes in neural activities and changing in weights, the brain has more

play26:08

timescales than that. The brain has rapid changes in weight that quickly decayed away.

play26:13

And that's probably how it does a lot of short term memory.

play26:15

And we don't have that in our models

play26:17

for technical reasons to do with being able to do matrix

play26:19

matrix multiplies.

play26:22

I still believe that if once

play26:24

we got that into our models they'd get better, but

play26:29

because of what I was doing for the two years previous to that,

play26:33

I suddenly came to believe that maybe the things we've got now,

play26:37

the digital models, we've got now, are already

play26:41

very close to as good as brains and will get to be much better than brains.

play26:45

Now I'm going to explain why I believe that.

play26:49

So digital computation is great.

play26:52

You can run the same program on different computers, different piece of hardware

play26:56

or the same neural net on different pieces of hardware.

play26:58

All you have to do is save the weights, and that means it's immortal

play27:02

once you've got some weights that are immortal.

play27:04

Because if the hardware dies, as long as you've got the weights,

play27:06

you can make more hardware and run the same neural net.

play27:11

But to do that,

play27:13

we run transistors at very high power, so they behave digitally

play27:17

and we have to have hardware that does exactly what you tell it to.

play27:21

That was great

play27:22

when we were instructing computers by telling them exactly how to do things,

play27:26

but we've now got

play27:28

another way of making computers do things.

play27:31

And so now we have the possibility of using all the very rich analogue

play27:35

properties of hardware to get computations done at far lower energy.

play27:40

So these big language models, when the training use like megawatts

play27:44

and we use 30 watts.

play27:50

So because we know how to train things,

play27:54

maybe we could use analogue hardware

play27:58

and every piece of hardware is a bit different, but we train it

play28:01

to make use of its peculiar properties, so that it does what we want.

play28:05

So it gets the right output for the input.

play28:09

And if we do that, then we can abandon the idea

play28:12

that hardware and software have to be separate.

play28:16

We can have weights that only work in that bit of hardware

play28:20

and then we can be much more energy efficient.

play28:24

So I started thinking

play28:25

about what I call mortal computation, where you've abandoned that distinction

play28:29

between hardware and software using very low power analogue computation.

play28:33

You can parallelize over trillions of weights that are stored as conductances.

play28:40

And what's more, the hardware doesn't need to be nearly so reliable.

play28:44

You don't need to have hardware that

play28:45

at the level of the instructions would always do what you tell it to.

play28:48

You can have goopy hardware that you grow

play28:52

and then you just learn to make it do the right thing.

play28:55

So you should be able

play28:56

to use hardware much more cheaply, maybe even

play29:00

do some genetic engineering on neurons

play29:02

to make it out of recycled neurons.

play29:06

I want to give you one example of how this is much more efficient.

play29:10

So the thing you're doing in neural networks all the time is taking a vector

play29:14

of neural activities, and multiplying it by a matrix of weights, to get the vector

play29:18

of neural activities in the next lane, at least get the inputs to the next lane.

play29:23

And so a vector matrix multiplies the thing you need to make efficient.

play29:28

So the way we do it in the digital

play29:30

computer, is we have these transistors that are driven a very high power

play29:35

to represent bits in say, a 32 bit number

play29:39

and then to multiply two 32 bit numbers, you need to perform.

play29:43

I never did any computer science courses, but I think you need to perform about 1000

play29:47

1 bit digital operations.

play29:48

It's about the square of the bitary.

play29:51

If you want to do it fast.

play29:54

So you do lots of these digital operations.

play29:58

There's a much simpler way to do it, which is you make a neural activity,

play30:01

be a voltage, you make a weight to be a conductance and a voltage times

play30:06

a conductance is a charge, per unit time

play30:09

and charges just add themselves up.

play30:11

So you can do your vector matrix

play30:14

multiply just by putting some voltages through some conductances.

play30:18

And what comes into each neuron in the next layer will be the product

play30:22

of this vector with those weights.

play30:25

That's great.

play30:26

It's hugely more energy efficient.

play30:28

You can buy chips to do that already, but every time you do

play30:32

it'll be just slightly different.

play30:35

Also, it's hard to do nonlinear

play30:37

things like this.

play30:40

So the several big problems with mortal computation,

play30:44

one is

play30:45

that it's hard to use back propagation because if you're making use

play30:49

of the quirky analogue properties of a particular piece of hardware,

play30:53

you can assume the hardware doesn't know its own properties.

play30:57

And so it's now hard to use the back propagation.

play31:00

It's much easier to use reinforcement algorithms that tinker with weights

play31:03

to see if it helps.

play31:04

But they're very inefficient. For small networks.

play31:08

We have come up with methods that are about as efficient as back propagation,

play31:12

a little bit worse.

play31:14

But these methods don't yet scale up, and I don't know if they ever will

play31:17

Back propagation in a sense, is just the right thing to do.

play31:20

And for big, deep networks, I'm not sure we're ever going to get

play31:24

things that work as well as back propagation.

play31:26

So maybe the learning algorithm in these analogue systems isn't going to be

play31:29

as good as the one we have for things like large language models.

play31:35

Another reason for believing that is, a large language

play31:38

model has say a trillion weights, you have 100 trillion weights.

play31:42

Even if you only use 10% of them for knowledge, that's ten trillion weights.

play31:46

But the large language model in its trillion weights

play31:49

knows thousands of times more than you do.

play31:52

So it's got much, much more knowledge.

play31:55

And that's partly because it's seen much, much more data.

play31:57

But it might be because it has a much better learning algorithm.

play32:00

We're not optimised for that.

play32:02

We're not optimised for packing

play32:04

lots of experience into a few connections where a trillion is a few.

play32:08

We are optimized for having not many experiences.

play32:13

You only live for about billion seconds.

play32:16

That's assuming you don't learn anything after you're 30, which is pretty much true.

play32:19

So you live for about billion seconds

play32:22

and you've got 100 trillion connections,

play32:26

so you've got

play32:27

crazily more parameters and you have experiences.

play32:30

So our brains optimise from making the best use of

play32:33

not very many experiences.

play32:38

Another big problem with mortal computation is that

play32:41

if the software is inseparable from the hardware,

play32:44

once a system is learned or if the hardware dies, you lose,

play32:47

all the knowledge, it's mortal in that sense.

play32:51

And so how do you get that knowledge into another mortal system?

play32:55

Well, you get the old one to give a lecture

play32:59

and the new ones to figure out how to change the weights in their brains.

play33:04

So they would have said that.

play33:06

That's called distillation.

play33:07

You try and get a student model to mimic

play33:10

the output of a teacher model, and that works.

play33:14

But it's not that efficient.

play33:16

Some of you may have noticed that universities just aren't that efficient.

play33:20

It's very hard to get the knowledge from the Professor into the student.

play33:26

So this distillation method,

play33:28

a sentence, for example, has a few hundred bits of information, and even

play33:32

if you learn optimally you can convey more than a few hundred bits.

play33:36

But if you take these big digital models,

play33:39

then, if you look at a bunch of agents that all have exactly

play33:46

the same neural netting with exactly the same weights

play33:51

and they're digital, so they

play33:52

use those weights in exactly the same way

play33:56

and these thousand different agents all go off

play33:58

and look at different bits of the Internet and learn stuff.

play34:01

And now you want each of them to know what the other one's learned.

play34:06

You can achieve that by averaging the gradients, so averaging the weights

play34:09

so you can get massive communication of what one agent learned to all the other agents.

play34:14

So when you share the weight, so you share the gradients, you're communicating

play34:18

a trillion numbers, not just a few hundred bits, but a trillion real numbers.

play34:24

And so they're fantastically much better at communicating,

play34:27

and that's what they have over us.

play34:30

They're just much, much better at

play34:33

communicating between multiple copies of the same model.

play34:36

And that's why

play34:36

GPT4 knows so much more than a human, it wasn't one model that did it.

play34:41

It was a whole bunch of copies of the same model running on different hardware.

play34:47

So my conclusion, which I don't really like,

play34:53

is that digital computation

play34:56

requires a lot of energy, and so it would never evolve.

play34:59

We have to evolve making use of the quirks of the hardware to be very low energy.

play35:05

But once you've got it,

play35:07

it's very easy for agents to share

play35:11

and GBT4

play35:12

has thousands of times more knowledge in about 2% of the weights.

play35:16

So that's quite depressing.

play35:19

Biological computation is great for evolving

play35:21

because it requires very little energy,

play35:25

but my conclusion is

play35:27

the digital computation is just better.

play35:32

And so I think it's fairly clear

play35:36

that maybe in the next 20 years, I'd say

play35:39

with a probability of .5, in the next 20 years, it will get smarter than us

play35:43

and very probably in the next hundred years it will be much smarter than us.

play35:48

And so we need to think about

play35:50

how to deal with that.

play35:52

And there are very few examples of more intelligent

play35:56

things being controlled by less intelligent things.

play35:59

And one good example is a mother being controlled by baby.

play36:03

Evolution has gone to a lot of work to make that happen so that the baby

play36:07

survive, is very important for the baby to be able to control the mother.

play36:11

But there aren't many other examples.

play36:14

Some people think that we can make

play36:16

these things be benevolent,

play36:19

but if they get into a competition with each other,

play36:22

I think they'll start behaving like chimpanzees.

play36:25

And I'm not convinced you can keep them benevolent.

play36:30

If they get very smart and they get any notion of self-preservation

play36:35

they may decide they're more important than us.

play36:38

So I finish the lecture in record time.

play36:42

I think.

Rate This

5.0 / 5 (0 votes)

相关标签
神经网络语言模型AI理解AI威胁数字模拟进化计算人工智能技术演进未来展望安全风险