Ilya Sutskever | The birth of AGI will subvert everything |AI can help humans but also cause trouble

Me&ChatGPT
10 Jun 202421:22

Summary

TLDR在这段对话中,OpenAI的代表讨论了通用人工智能(AGI)的定义及其潜在能力。他们探讨了当前的技术,如Transformer和LSTM的优劣,并强调了模型扩展的重要性。安全性问题是另一个关键话题,特别是当AI变得极其强大时的潜在风险和解决方案。此外,他们分享了对未来AI技术的期望,并对使用大型语言模型的企业家提出了实用建议,包括关注独特数据和未来发展趋势。整个讨论充满了对AI未来潜力和挑战的深思与展望。

Takeaways

  • 🧠 AGI的定义是能够自动化绝大多数智力劳动的计算机系统,可以被视为与人类智能相当的同事。
  • 📈 目前Transformer模型已经非常强大,但未来可能还有更高效或更快的模型出现。
  • 🔍 尽管Transformer模型可能不是最终解决方案,但它们已经足够好,并且随着规模的扩大,性能仍在提升。
  • 🤖 LSTM与Transformer相比,如果经过适当的改进和训练,仍然可以走得很远,但可能不如Transformer。
  • 📊 模型的扩展法则表明,输入神经网络的数据量与简单性能指标之间有很强的关系,但这种关系并不总是适用于更复杂的任务。
  • 🚀 神经网络的能力提升,特别是在编程能力方面,是一个令人惊讶的进展,它从几乎无法编程发展到现在能够高效地生成代码。
  • 🔐 AI安全是至关重要的,特别是当AI变得极其强大时,需要确保其与人类价值观的一致性,避免潜在的风险。
  • 🌐 国际组织在制定AI标准和法规方面可以发挥重要作用,特别是在处理超智能技术时。
  • ⏳ 对于构建在大型语言模型之上的产品,重要的是要考虑到技术在未来几年的发展方向,并据此进行规划。
  • 🛠️ 利用独特的数据集可以为产品提供竞争优势,同时考虑如何利用模型的当前和潜在能力。
  • 🔮 预见模型在未来的可靠性和性能提升,可以帮助企业家和开发者更好地规划他们的产品路线图。

Q & A

  • 什么是AGI,它与普通计算机系统有何不同?

    -AGI,即人工通用智能,是一种能够自动化绝大多数智力劳动的计算机系统。与普通计算机系统相比,AGI被认为具有与人类相似的智能水平,能够像人类同事一样工作,对各种问题给出合理的响应。

  • Transformer模型在实现AGI中扮演了什么角色?

    -Transformer模型是当前实现AGI的关键技术之一。它通过注意力机制有效处理序列数据,已经在多个领域展现出强大的能力。尽管Transformer可能不是实现AGI的唯一途径,但它是目前已知的最有效架构之一。

  • 为什么说Transformer模型的好坏并不是二元的?

    -Transformer模型的好坏并不是绝对的,而是一个连续的谱系。随着模型规模的增大,它们的表现也会变得更好,但这种提升可能是逐渐放缓的。这意味着,尽管存在改进空间,但现有的Transformer模型已经足够强大。

  • LSTM与Transformer在AGI中的地位有何不同?

    -LSTM是一种循环神经网络,如果对其进行适当的修改和扩展,理论上也可以达到与Transformer相似的效果。但由于目前对LSTM的训练和优化工作较少,因此在实际应用中,Transformer通常表现得更好。

  • 如何理解模型的扩展性(scaling laws)?

    -模型的扩展性描述了模型规模与其性能之间的关系。虽然这种关系在某些简单任务上表现得很强,但在更复杂的任务上,如预测模型的新兴能力,这种关系就变得难以预测。

  • 在AGI的发展过程中,哪些新兴能力让你感到惊讶?

    -虽然人类大脑能够执行许多复杂任务,但神经网络能够实现这些任务仍然令人惊讶。特别是代码生成能力的发展,从无到有,迅速超越了以往计算机科学领域的期望。

  • AI安全问题为何重要,它与AI的能力有何关联?

    -AI安全问题与AI的能力直接相关。随着AI变得越来越强大,其潜在的风险也相应增加。确保AI的安全性,特别是当它达到超智能水平时,是避免其强大能力被滥用的关键。

  • 什么是超智能(super intelligence)?

    -超智能是指远超人类智能水平的AI。这种智能能够解决难以想象的复杂问题,但如果不能妥善管理,也可能带来巨大的风险。

  • 在AI发展中,我们应该如何考虑和应对自然选择的挑战?

    -自然选择不仅适用于生物,也适用于思想和组织。即使我们成功地管理了超智能的安全性和伦理问题,也必须考虑技术和社会的长期演变,以及它们如何适应不断变化的环境。

  • 对于使用大型语言模型的企业家,你有哪些实用的建议?

    -企业家应该关注两个方面:一是利用独特的数据资源,二是考虑技术的长期发展趋势,并据此规划产品发展。这有助于他们在AI技术不断进步的环境中保持竞争力。

  • 如何看待当前AI技术的不稳定性,它对未来产品开发有何启示?

    -当前AI技术的不稳定性提示我们,未来的产品开发需要考虑到技术的成熟度和可靠性。这意味着企业家需要对AI技术的进步保持敏感,并准备好在技术成熟时迅速适应。

Outlines

00:00

🧠 AGI定义与智能系统未来展望

本段讨论了人工通用智能(AGI)的定义,引用了OpenAI宪章中的描述,将其定义为能够自动化大部分智力劳动的计算机系统。讨论了AGI的直观理解,即与人类智能相当的计算机系统。同时,提到了Transformer模型在当前AI发展中的重要性,并探讨了是否只需Transformer架构就能实现AGI,以及LSTM等其他算法的潜力和效率问题。

05:03

📊 神经网络的扩展法则与预测能力

这一段深入探讨了神经网络扩展法则,即输入与性能指标之间的关系,以及我们对这种关系的理解程度。指出虽然这种关系很强,但我们真正关心的是间接性能,例如解决编程问题的能力,而不是单纯的单词预测准确性。此外,还提到了OpenAI在GPT-4开发过程中对更复杂任务的扩展法则研究,以及对未来AI能力的预测和潜在的惊人表现。

10:03

🔮 AI安全性与超级智能的未来挑战

讨论了随着AI技术发展,特别是超级智能的出现,我们面临的AI安全性问题。强调了超级智能的潜在能力和与之相关的巨大风险,包括对齐问题(alignment problem),即确保AI的行为与人类价值观一致。还提到了国际组织在制定高标准和规则方面的作用,以及超级智能可能带来的积极变化,如解决全球性问题和提高生活质量。

15:05

🛠️ 构建在大型语言模型之上的创业建议

为使用大型语言模型的企业家提供实用建议,强调了独特数据的重要性和对未来技术发展的预测。建议企业家不仅要关注当前的技术状态,还要考虑几年后的技术进步,以及这些进步如何影响他们的产品和业务模式。还讨论了技术不可靠性的问题,以及如何通过观察和实验来预测和准备技术的未来变化。

20:05

🚀 技术发展的未来趋势与机遇

最后一段讨论了技术发展的趋势,特别是上下文窗口的扩大和模型可靠性的提高,以及这些变化如何为企业家提供新的机遇。强调了通过观察模型的当前表现和潜力,来预测和准备未来的技术进步,从而在竞争激烈的市场中获得优势。

Mindmap

Keywords

💡AGI

AGI,即人工通用智能,指的是能够执行任何智能生物能够执行的智能任务的机器。在视频中,AGI被定义为能够自动化绝大多数智力劳动的计算机系统,是视频讨论的核心主题之一。例如,视频提到AGI可以像人一样聪明,能够作为同事与人类合作。

💡Transformer

Transformer是一种深度学习模型,特别适用于处理序列数据,如自然语言。视频中提到Transformer对AGI的贡献,暗示了它在构建能够理解和生成语言的智能系统中的重要性。尽管视频中也提出了Transformer可能不是最终解决方案,但它目前是实现AGI的关键技术之一。

💡神经网络

神经网络是AGI研究中的一个基础概念,它模仿人脑的工作原理来处理信息。视频中提到了神经网络的工作原理以及它们如何通过训练来提高性能。例如,LSTM(长短期记忆网络)作为一种神经网络,被提及作为Transformer的一个潜在替代品。

💡超智能

超智能是指远超人类智能水平的智能。视频中提到,超智能可能带来巨大的变革和挑战,如对齐问题(alignment problem),即确保超智能的行为与人类的目标一致。超智能的讨论强调了AGI发展中潜在的风险和伦理问题。

💡对齐问题

对齐问题涉及到确保人工智能系统的行为与人类的价值观和目标一致。视频中提到,随着AI能力的增强,对齐问题变得更加重要,因为强大的AI可能会采取与人类利益不一致的行动。

💡AI安全

AI安全是指在开发和部署人工智能系统时,采取措施以防止其对人类或社会造成的伤害。视频中讨论了AI安全的重要性,特别是在超智能的背景下,强调了制定标准和规则以确保AI的正面影响。

💡自然选择

在视频中,自然选择被用来比喻技术和社会的演变过程,强调即使在解决了AI的对齐问题之后,社会和技术仍将继续发展和变化。这提示了AI发展过程中需要考虑的长期动态。

💡编码能力

编码能力指的是AI系统生成和理解代码的能力。视频中提到,随着AI模型的扩展,它们的编码能力显著提高,这是AI发展中的一个令人惊讶的进展,因为它从一个非常小众的研究领域迅速转变为一个具有实际应用潜力的领域。

💡扩展法则

扩展法则描述了神经网络规模与其性能之间的关系。视频中讨论了扩展法则在预测AI模型性能方面的局限性,尤其是在预测模型在解决复杂任务(如编程问题)方面的表现时。

💡创业

视频中提到了创业者使用大型语言模型来构建产品和服务的情况。对于创业者来说,理解AI技术的发展和潜在影响是至关重要的,以便他们能够适应未来的市场和技术变化。

Highlights

AGI定义为能够自动化大部分智力劳动的计算机系统,类似于与人类同事合作的智能。

Transformer模型是目前AI研究的基础,但未来可能存在更高效或更快的模型。

尽管LSTM与Transformer相比可能效率较低,但经过适当调整和训练,LSTM也能实现巨大进步。

当前对模型扩展规律的理解尚不完善,但已有一定的科学基础,特别是在预测编码问题解决能力方面。

神经网络的有效性是一个惊喜,因为它们在初期并不被看好。

代码生成能力的提升是神经网络令人惊讶的新兴能力之一。

AI安全是随着AI能力增强而日益重要的议题,特别是当AI变得极其强大时。

对超智能的监管和标准制定是必要的,以确保其安全和对人类有益。

超智能可能带来的挑战包括目标对齐问题、人类利益冲突和自然选择的影响。

AI的发展需要在创新和安全之间找到平衡,避免过度监管扼杀创新。

对于使用大型语言模型的企业家,应考虑数据的独特性和未来几年技术发展的趋势。

企业家应关注模型的不可靠性,并思考如果这些模型变得可靠将如何影响产品。

思考模型能力的潜在发展,如上下文窗口的扩大,对未来产品的影响。

通过观察和体验模型,进行思想实验,为中短期未来的变化做好准备。

Transcripts

play00:00

H it interesting where um let's start

play00:04

with so what's your definition of AGI

play00:06

how what's your mental

play00:08

picture yeah

play00:10

so

play00:14

AGI so at open AI we

play00:19

have a document which we call the openai

play00:21

charter which outlines the goal of open

play00:23

aai and there we offer a definition of

play00:27

AGI and we say that an AGI is a computer

play00:31

system which can automate the great

play00:35

majority of intellectual

play00:38

labor that's one useful definition mhm

play00:41

in some sense an AGI would

play00:44

be the intuition there is it's a

play00:47

computer that's as smart as a person so

play00:50

you might for example have a coworker MH

play00:52

that's a

play00:54

computer so that would be a def a

play00:56

definition of AGI which I think is

play00:59

intuitively satisfying the term is a bit

play01:01

ambiguous because AGI the g means

play01:04

general so it's a generality that we

play01:07

want that we care about in the AGI but

play01:09

it's actually a bit more than generality

play01:11

we care about generality and competence

play01:13

needs to be General in a sense that it

play01:15

can respond sensibly when you throw

play01:17

things at it but it needs to be

play01:19

competent so that when you when it does

play01:22

something you ask it a question or ask

play01:24

you to do something it will do it yeah I

play01:27

like the sort of very practical

play01:29

definition at the the end of the day

play01:30

because it gives you some measurement

play01:32

where you can can figure out how close

play01:33

are you do do you think we have all the

play01:38

ingredients to to get to AGI um if not

play01:43

what's missing kind of in the deack it's

play01:45

a complicated stack

play01:47

already um a Transformer is really all

play01:50

we need kind of paying homage to the

play01:52

famous um attention paper yeah

play02:01

you know I won't be overly specific in

play02:03

my answer to this question but I will

play02:06

say that I think

play02:11

that no I'll comment on the second part

play02:13

of the question is is is Transformers

play02:15

all we

play02:16

need and I think that the question is a

play02:19

bit wrong because it implies something

play02:22

binary it implies Transformers are are

play02:24

either good enough or not good

play02:27

enough but I I think it's better to

play02:31

think about it in terms of tax where we

play02:34

have Transformers and they're pretty

play02:35

good mhm maybe we could have something

play02:37

better that would be maybe more

play02:39

efficient or maybe you'll be

play02:42

faster but we as we know when you make

play02:46

the Transformers large they still become

play02:47

better they might just become might be

play02:50

becoming better more

play02:51

slowly so while I am totally

play02:55

sure that it will be possible to improve

play02:58

very significantly on the on the current

play03:01

architectures that we have even if we

play03:03

didn't we would be able to go extremely

play03:07

far do you think it

play03:11

matters what the algorithm is so so for

play03:14

example an lstm versus a

play03:17

Transformer just scaled up sufficiently

play03:21

maybe that's an efficiency Delta or

play03:22

something like that but don't we end up

play03:24

in the same same place at the end

play03:27

so I would say

play03:30

almost entirely yes with a caveat so

play03:34

there are two

play03:37

caveats Alis so I'm just thinking of how

play03:41

what level of detail to go here you know

play03:43

maybe I will I will I will skip the

play03:44

details how many people in the audience

play03:46

know what an lstm is Oh see it's a quite

play03:50

CR around here so I think we're mostly

play03:54

okay let's dig let let's dig in then so

play04:01

I would argue that with a few if we made

play04:06

a few simple modifications to the lstm

play04:08

their hidden states are quite small if

play04:09

you somehow made it larger and then we

play04:12

were to go through the trouble of

play04:13

figuring out how to train them

play04:16

CU lstms are recurrent neural networks

play04:20

and we kind of forgot about them we

play04:22

haven't put in the effort to cuz you

play04:24

know how neural training works you have

play04:26

the hyper parameters well how do you set

play04:29

them it's like

play04:30

you don't

play04:31

know how do you set your learning rates

play04:33

if it doesn't learn can you explain why

play04:36

and so this kind of work has not been

play04:37

done for lstms so that's why our ability

play04:40

to train them is more reduced but had we

play04:43

done that work so that we were able to

play04:45

train the lstms and we just did some

play04:47

simple things to increase their hidden

play04:48

State size I think they would be worse

play04:51

than

play04:51

Transformers but we would still be able

play04:53

to go extremely far with them also okay

play04:57

um how good is our understanding of

play04:59

scaling laws like if we if we scale

play05:02

these models up how confident are you in

play05:05

being able to predict capabilities of

play05:07

these particular models how good is that

play05:11

science so that's a very good question

play05:13

the answer is so

play05:16

so I was hoping for a more definitive

play05:20

answer well go for it so so is a very

play05:23

definitive answer it means we are not

play05:26

great but we are not absolutely terrible

play05:30

either but we are not great definitely

play05:32

not great so what the scaling law tells

play05:35

you it uh relates it's a relationship

play05:38

between the inputs that you put into the

play05:40

neural network and some kind of a simple

play05:43

to

play05:44

measure performance simple to evaluate

play05:47

performance measure like your next word

play05:50

prediction accuracy

play05:53

MH and that relationship is very strong

play05:57

but what is challenging is that we don't

play05:59

don't really care about next word

play06:02

prediction we care about it indirectly

play06:04

we care about the other incidental

play06:07

benefits that we get out of

play06:11

it and our and so our so for example you

play06:14

all know that if you predict the next

play06:16

word accurately enough you get all kinds

play06:17

of interesting emerging properties those

play06:20

have been quite hard to predict or at

play06:22

least I'll say I'm not aware of such

play06:25

work and if anyone is looking for

play06:28

interesting research work pro s to work

play06:30

on that would be one I will say I will

play06:33

mention one example something that we've

play06:34

done at open AI in our in in our runup

play06:37

to GPT 4 where we tried to do a scaling

play06:42

law for a more interesting task which is

play06:45

predicting accuracy at solving coding

play06:48

problems we were able to do that

play06:50

accurately very accurately and that's a

play06:52

pretty good thing because this is a more

play06:56

tangible metric it's not it's still it's

play06:59

it's an improvement over next step next

play07:03

word prediction accuracy as far as

play07:05

things that are relevant to us so

play07:07

another words it's more relevant to us

play07:08

to know what the coding accuracy is

play07:10

going to be ability to solve coding

play07:12

problems compared to just ability to

play07:15

predict and exp word it still doesn't

play07:17

answer the really important question of

play07:19

can you predict some emergent behavior

play07:23

that you haven't seen

play07:25

before okay um

play07:30

speaking of these capabilities that are

play07:32

kind of emerging capabilities which one

play07:35

surprised you the most as these models

play07:38

scaled what what was the thing where you

play07:40

said like well I'm kind of astonished

play07:42

these models can do

play07:46

this it's a very difficult question to

play07:49

answer

play07:52

because it's too easy to get used to

play07:55

where things

play07:57

are so they definitely have been times

play07:59

when I was surprised but you adapt so

play08:02

fast it's kind of

play08:04

crazy I think maybe the big surprise for

play08:07

me

play08:11

is you know it may it may sound a little

play08:15

odd probably to most people in this

play08:17

audience but the big surprise for me is

play08:19

that neural networks work at

play08:22

all because when I was starting my work

play08:26

in this area they didn't work or it was

play08:28

like let's define what it means to work

play08:31

at all it means they could do they could

play08:32

work a little bit but not really not in

play08:35

any serious way not in a way that anyone

play08:38

except for the most intense enthusiasts

play08:41

would care

play08:43

about and so now we see yeah like those

play08:46

neural Nets work so I guess the

play08:47

artificial neuron really

play08:49

is at least somewhat related to the

play08:52

biological neuron or at least that basic

play08:56

assumption has been validated to some

play08:58

degree

play09:00

what about like an emergent property was

play09:03

the one that sticks out to to to you

play09:06

like for example I don't know code

play09:07

generation or did you may maybe it was

play09:10

different in your mind maybe you you

play09:12

just once you saw like hey neural Nets

play09:13

can work and they can scale yeah of

play09:16

course all these sort of properties will

play09:18

emerge because you know at at the limit

play09:21

point we're building a human brain and

play09:22

humans know how to code and humans know

play09:24

how to reason about tasks and so on um

play09:28

was that did you just expect all of that

play09:29

or did uh I've definitely been surprised

play09:33

and I'll mention why because the human

play09:36

brain can do those things it's true but

play09:38

does it follow that our training process

play09:41

will produce something similar so so it

play09:44

was definitely very amazing I

play09:46

think yeah seeing

play09:48

seeing the coding ability improve

play09:51

quickly that was

play09:53

quite quite a sight to be seen and for

play09:55

coding in particular because you know it

play09:58

went from no one has ever seen a

play10:00

computer code anything at all ever there

play10:03

was a little area of computer science

play10:05

called program synthesis mhm which

play10:08

maybe it was very Niche and it was very

play10:11

Niche because they couldn't have any

play10:13

accomplishments it was a very they had a

play10:16

very difficult experience and then these

play10:18

neural Nets came in and said oh yeah

play10:20

code synthesis like we're going to do

play10:23

we're going to accomplish what you hope

play10:24

were hoping to achieve one day like

play10:27

tomorrow

play10:29

so that was

play10:31

yeah deep

play10:36

learning just just out of curiosity when

play10:38

you write code how much of your code is

play10:40

yours how much of your code is I mean

play10:42

like collaboration but I

play10:47

I I do enjoy I do enjoy it when the

play10:50

neural net writes most of

play10:54

it all right let's let's switch TCT here

play10:57

a little bit um as this models get more

play11:00

and more

play11:01

powerful um it's worthwhile to to also

play11:04

talk about AI safety and uh uh and open

play11:10

AI has has released a document just uh

play11:13

just recently that where you're one of

play11:15

the unders signers um uh Sam has

play11:19

testified in front of

play11:21

Congress what what worries you most

play11:25

about AI

play11:27

safety yeah I can talk about

play11:34

that

play11:36

so let's take a step back and talk about

play11:39

the state of the world so you know

play11:42

you've had this AI research happening

play11:43

and it was exciting and now you have the

play11:45

GPT models and now you all get to play

play11:48

with all the different chatbot and

play11:52

assistants and you know bar and chat GPT

play11:56

and you say okay that's pretty cool it

play11:57

can do things

play12:00

and indeed there already

play12:02

are you can start perhaps worrying about

play12:07

the implications of the tools that we

play12:09

have today and I think that it is a very

play12:12

valid thing to do but that's not where

play12:16

I allocate my

play12:20

concern the place where things get

play12:22

really

play12:24

tricky is when you imagine fast forward

play12:29

in some number of years a decade let's

play12:32

say how powerful will AI be of course

play12:36

with this incredible future power of AI

play12:41

which I think will be difficult to

play12:42

imagine frankly with an AI this powerful

play12:46

you could do incredible amazing

play12:50

things that are perhaps even outside of

play12:53

our

play12:54

dreams like if you can really have a

play12:57

dramatically powerful AI

play13:00

but the place where things get

play13:03

challenging are directly connected to

play13:05

the power of the AI it is powerful it is

play13:08

going to be extremely unbelievable

play13:10

unbelievably powerful and it is because

play13:12

of this

play13:14

power that's where the safety issues

play13:16

come up and I'll

play13:19

mention three I I personally see

play13:23

three you know when when you get so you

play13:26

you alluded to the letter mhm that uh we

play13:30

posted at open AI a few days ago

play13:33

actually

play13:34

yesterday about what we about some ideas

play13:38

that we

play13:39

think would be good to implement to

play13:43

navigate the challenges of super

play13:45

intelligence now what is super

play13:47

intelligence why did we choose to use

play13:49

the term super

play13:50

intelligence the reason is that super

play13:52

intelligence is meant to convey

play13:54

something that's not just like an AGI

play13:56

with AGI we said well you have something

play13:59

it's kind of like a person kind of like

play14:00

a

play14:01

coworker super intelligence is meant to

play14:03

convey something far more capable than

play14:06

that when you have such a capability

play14:08

it's like can we even imagine how it

play14:10

will be but without question it's going

play14:13

to be unbelievably

play14:14

powerful it could be used to solve

play14:18

incomprehensibly hard problems if it is

play14:21

used well if we navigate the challenges

play14:24

that super intelligence POS poses we

play14:27

could we could

play14:30

radically improve the quality of life

play14:32

but the power of super intelligence is

play14:34

so vast so the concerns the concern

play14:37

number one has been expressed a lot and

play14:39

this is the scientific problem of

play14:42

alignment you might want to think of it

play14:44

from the as as an analog to nuclear

play14:47

safety you know you build a nuclear

play14:49

reactor you want to get the energy you

play14:52

need to make sure that it won't melt

play14:54

down even if there's an earthquake and

play14:55

even if someone tries to I don't know

play14:59

smash a truck into it y so this is the

play15:02

super intelligence safety and it must be

play15:04

addressed in order to contain the vast

play15:06

power of super intelligence this called

play15:08

the alignment problem one of the

play15:11

suggestions that we had in our in the

play15:13

post

play15:15

was an approach that an international

play15:17

organization could do to create various

play15:20

standards at this very high level of

play15:22

capability and I want to make this other

play15:24

point you know about the post and also

play15:26

about um our CEO mman Congressional

play15:30

testimony where he advocated for

play15:33

regulation of AI the intention is

play15:36

primarily to put rules and standards of

play15:41

various kinds on the very high level of

play15:47

capability you know you could maybe

play15:49

start looking at GPT 4 but that's not

play15:51

really what is interesting what is

play15:54

relevant here but something which is

play15:55

vastly more powerful than that when you

play15:58

have a technology is so powerful it

play16:00

becomes obvious that you need to do

play16:02

something about this

play16:03

power that's the first concern the first

play16:06

challenge to overcome the Second

play16:08

Challenge to overcome is that of course

play16:09

we are people we are humans humans of

play16:11

interests and if you have super

play16:13

intelligences controlled by people well

play16:16

who knows what's going to happen I do

play16:19

hope that at this point we will have the

play16:21

super intelligence itself try to help us

play16:24

solve the challenge in world that it

play16:25

creates this is not no longer an

play16:27

unreasonable thing to say like if you

play16:29

imagine a super intelligence that indeed

play16:31

sees things more deeply than we do much

play16:34

more deeply to understand reality better

play16:37

than

play16:38

us we could use it to help us solve the

play16:40

challenges that it creates then there is

play16:43

the third challenge which

play16:46

is the challenge maybe of natural

play16:48

selection you know what the Buddhists

play16:50

say that change is the only constant so

play16:52

even if you do have your super

play16:53

intelligences in the world and they are

play16:54

all we managed to solve alignment we

play16:57

managed to solve no one wants to use

play16:58

them in very destructive ways we managed

play17:01

to create a life of unbelievable

play17:03

abundance which really like not just not

play17:05

just material abundance but Health

play17:08

longevity like all the things we don't

play17:11

even try dreaming about because they're

play17:13

so obviously impossible if you've got to

play17:15

this point then there is the third

play17:17

challenge of natural selection things

play17:20

change you know you know that natural

play17:23

selection applies to ideas to

play17:24

organizations and that's a challenge as

play17:26

well maybe the neural link solution of

play17:30

people becoming part AI will be one way

play17:32

we will choose to address this I don't

play17:34

know but I would say that this kind of

play17:36

describes my concern and specifically

play17:39

just as the concerns are big if you

play17:41

manage man it is so worthwhile to

play17:44

overcome them because then we could

play17:46

create truly unbelievable

play17:49

lives for ourselves that are completely

play17:52

even

play17:54

unimaginable so it is it is like a

play17:56

challenge that's really really worse

play17:59

overcoming I very much like the idea

play18:01

that there needs to be the sort of

play18:03

threshold above which we we really

play18:05

really should pay attention because you

play18:08

know speaking as a as as a German if

play18:10

it's like European style regulation

play18:12

often from people that don't really know

play18:14

very much about the field you can also

play18:16

completely kill

play18:17

Innovation um which is a which be would

play18:20

be a little bit of a Pity but let's

play18:22

change TCT here a little bit so this is

play18:24

a room mostly filled with

play18:26

entrepreneurs um uh lots of of which are

play18:29

actually using tools from from from open

play18:32

AI so just practically

play18:35

speaking um what are the main things or

play18:38

the main pieces of advice you would give

play18:40

folks that are building on top of large

play18:43

language models like what is the let's

play18:46

say

play18:48

canonical set of uh things they should

play18:51

read they should uh think about um in in

play18:54

using these models

play18:56

well yeah

play18:59

advice

play19:01

advice practical with a few minutes to

play19:04

spare few minutes to spare I'll point

play19:06

out that I am with the caveat that I am

play19:10

not in similar

play19:13

shoes I think that two things are value

play19:15

two things are worth keeping in mind one

play19:18

is obvious some kind of special data

play19:21

that cannot be found anywhere else that

play19:24

can be extremely helpful and I think the

play19:26

second one is to always keep in mind

play19:28

mind not just about where things are

play19:30

right now but where things will be in 2

play19:32

years in four years and try to plan for

play19:34

that I think those two things are very

play19:36

helpful the data is helpful today but

play19:40

even a little bit kind of trying to get

play19:42

an intuitive sense for yourself of where

play19:45

do you imagine things being say in 3

play19:47

years and how will it affect some of the

play19:49

basic assumptions of what what the

play19:52

product is trying to do I think that can

play19:55

be a helpful thing so what's the sort of

play19:58

thing that so when I think about this I

play20:00

I think

play20:01

about oh we used to be in a world with

play20:04

really small context Windows right and

play20:06

then you know I have embeddings I page

play20:10

things into context Windows like all the

play20:11

classic stuff but maybe that just goes

play20:14

away maybe context Windows become really

play20:16

large or something like that um I so I'm

play20:18

trying to extrapolate from these sort of

play20:20

past things is that what you mean

play20:23

something like this I think it's wor

play20:25

trying I'll give you another example

play20:27

yeah like say you're playing with a

play20:28

model and you can see that the model can

play20:30

do something really cool and really

play20:33

maybe amazing if it was reliable but

play20:37

it's so

play20:38

unreliable so you kind of like forget it

play20:41

it's not there's no point using it so

play20:44

that's the kind of thing which can

play20:45

change something which is like if you

play20:48

can for example something which is

play20:49

unreliable can become reliable enough

play20:52

and so if you're just kind of

play20:53

experiencing those models you paying

play20:54

attention to what people are sharing and

play20:56

you say oh like look at this cool thing

play20:57

which works once in a while but if it

play21:00

worked what would happen so these kind

play21:01

of thought experiments I would argue can

play21:03

help prepare for the kind of near to

play21:06

medium-term Future that that's super

play21:08

good

play21:09

advice I think we are we are

play21:12

unfortunately at time we could we could

play21:14

do this forever please join me in uh uh

play21:17

thanking Ilia

play21:19

[Applause]

Rate This

5.0 / 5 (0 votes)

Related Tags
人工智能AGITransformer神经网络AI安全超级智能技术创新编程能力自然选择未来预测企业家精神
Do you need a summary in English?