Inside OpenAI [Entire Talk]

Stanford eCorner
26 Apr 202350:23

Summary

TLDR本次访谈中,OpenAI联合创始人兼首席科学家Ilya Sutskever讨论了人工智能的发展,特别是大型语言模型GPT-3的创新及其对社会的影响。Ilya分享了他对AI未来的看法,包括对深度学习、专业训练与通用训练的见解,以及OpenAI从非营利组织转变为有限利润公司的背后逻辑。他还谈到了AI的伦理问题,以及如何平衡技术进步与全球公民责任。Ilya鼓励对AI和创业感兴趣的学生,要信任直觉并利用自己独特的才能和视角。

Takeaways

  • 🌟 伊利亚·苏茨凯弗是OpenAI的联合创始人兼首席科学家,对大型语言模型GPT-3及其前身产品ChatGPT的发展做出了重要贡献。
  • 🎓 伊利亚出生于俄罗斯,成长于以色列,并在多伦多大学完成了数学学士和计算机科学硕士及博士学位。
  • 🧠 伊利亚认为,人工智能的发展受到人类大脑神经元工作原理的启发,人工神经网络通过学习数据来提高预测能力。
  • 📈 伊利亚讨论了人工智能的潜力,包括在特定领域如法律或医学中通过专业数据集训练来提高性能。
  • 🚀 伊利亚提到,尽管目前人工智能在学习和适应方面与人类存在差异,但未来可能会达到或超越人类的学习速度。
  • 🌐 OpenAI的使命是确保人工智能的发展惠及全人类,公司从非营利组织转变为有上限利润的公司,并与微软建立了合作关系。
  • 💡 伊利亚强调,随着人工智能能力的增长,未来可能会出于安全考虑而不再开源这些模型。
  • 📊 伊利亚认为,衡量OpenAI成功的关键是技术进步、研究执行情况以及对模型的控制和引导。
  • 🔍 伊利亚提到,尽管OpenAI的某些方面是封闭的,但公司仍为学术研究提供了不同形式的模型访问权限。
  • 🎭 伊利亚个人非常喜欢GPT-3的诗歌创作能力,他认为这是人工智能一个有趣且引人注目的特点。
  • 🤖 伊利亚讨论了人工智能在团队中的集成如何提高生产力,尽管目前尚未对团队动态产生显著影响。
  • 🌍 伊利亚表达了对全球人工智能监管和合理进步的看法,认为未来可能会有更多的政府监管来确保人工智能的安全和负责任的使用。

Q & A

  • Ilya的背景和教育经历是怎样的?

    -Ilya出生于俄罗斯,五岁时移居以色列,在以色列度过了他的成长岁月。他在以色列完成了大学前半部分的学习,之后转学到多伦多大学,获得了数学学士学位。之后,他在多伦多大学继续深造,获得了计算机科学的硕士和博士学位。

  • Ilya在OpenAI的角色是什么?

    -Ilya是OpenAI的联合创始人和首席科学家,致力于构建对全人类有益的人工通用智能。

  • Ilya对于大型语言模型的看法是什么?

    -Ilya认为大型语言模型是一种强大的技术,通过训练大型神经网络来预测文本中的下一个词,从而实现对人类语言的深入理解和生成。

  • Ilya如何看待人工智能的未来,特别是在学习速度和适应性方面?

    -Ilya认为虽然目前人工智能在学习和适应方面与人类有很大不同,但他不会押注反对人工智能在未来某个时刻达到或超过人类的学习速度和适应性。

  • Ilya对于意识或自我意识的看法是什么?

    -Ilya对意识问题持开放态度,他认为意识可能是一个程度问题,并且对于未来的人工智能系统,如果它们能够展现出与人类类似的意识特征,那将是非常值得思考的。

  • OpenAI的使命和转变是怎样的?

    -OpenAI的使命是确保人工通用智能的利益普惠全人类。它最初是一个非营利和开源组织,后来转变为有限责任公司,并与微软建立了紧密的合作关系。

  • Ilya如何看待OpenAI的开源与闭源的转变?

    -Ilya认为在人工智能的能力较低时,开源是有益的,但随着能力的提升,出于安全考虑,闭源可能变得必要。OpenAI的闭源决定是出于对人工智能能力增长的预期和对安全性的考虑。

  • Ilya对于OpenAI的盈利模式有何看法?

    -Ilya解释说OpenAI是一个有利润上限的公司,这意味着一旦对投资者的义务得到履行,OpenAI将再次成为一个非营利组织。他认为这种结构是独特的,并且有助于平衡公司的使命和经济需求。

  • Ilya对于人工智能在全球范围内的监管和伦理问题有何看法?

    -Ilya认为未来的世界需要明智的政府监管来确保人工智能的发展是可控和负责任的。同时,他也认为每个国家都有其在全球人工智能发展中的角色和责任。

  • Ilya对于想要进入AI领域的斯坦福学生有什么建议?

    -Ilya建议学生们应该发挥自己的独特优势和直觉,深入研究自己感兴趣的领域。他认为在研究和创业方面,独特的视角和创新思维都是非常重要的。

  • Ilya如何看待深度学习在未来5到10年的发展?

    -Ilya预计深度学习将继续取得进步,尽管通过扩大模型规模来获得进步的时代可能已经结束,但深度学习在许多层面上仍有改进的空间,这些改进将共同推动领域的发展。

Outlines

00:00

🌟 欢迎与介绍

本段介绍了斯坦福大学与YouTube社区合作的创业思想领袖研讨会,特别邀请了OpenAI的联合创始人兼首席科学家Ilia Sutskever。Ilia在人工智能领域有着杰出的贡献,是GPT-3等大型语言模型的基础思想者。他的个人背景和职业经历也被简要介绍,包括他如何从俄罗斯移居以色列,并在多伦多大学完成学业,最终成为OpenAI的关键人物。

05:01

🤖 大型语言模型技术

Ilia解释了大型语言模型的工作原理,包括人工神经网络的学习过程和反向传播算法。他强调了这些模型如何通过预测文本中的下一个词来训练,并指出这些模型的成功依赖于它们对“接下来会发生什么”的预测能力。他还提到了人类与机器学习方式的差异,以及如何通过大量数据训练来提高模型的性能。

10:02

🧠 人工智能与意识

Ilia探讨了人工智能与意识之间的关系,分享了他个人对意识的好奇和探索。他提出了一个未来可能进行的实验,即在不提及意识的情况下训练AI,然后观察AI是否能自行理解和表达意识这一概念。他还讨论了意识的度量问题,认为意识可能是一个程度问题,而不是简单的二元对立。

15:03

📈 OpenAI的使命与伦理

Ilia讨论了OpenAI的使命,即确保人工智能的利益普惠全人类,以及公司从非营利组织转变为有限责任公司的过程。他解释了与微软的合作关系,以及如何平衡公司的商业目标与对AI安全性的考虑。他还谈到了自己作为首席科学官的角色,以及他对公司发展方向的看法。

20:04

🌐 全球视角与AI的传播

Ilia讨论了AI技术在全球范围内传播的重要性,以及他作为世界公民的责任。他提到了与其他国家和地区的合作,以及如何在全球范围内推动AI技术的合理使用和监管。他还分享了自己对于AI技术未来发展的看法,包括深度学习的演进和专业领域内的应用。

25:06

🎯 OpenAI的未来方向

Ilia对OpenAI的未来方向提供了见解,包括公司可能成为人们直接使用的平台,或者作为其他应用背后的支持技术。他还讨论了ChatGPT在团队工作中的应用,以及它如何提高生产力和效率。最后,他分享了一些关于ChatGPT的有趣用途,如写诗和说唱,以及它如何被整合到OpenAI的团队工作中。

Mindmap

Keywords

💡人工智能

人工智能是指由人造系统所表现出来的智能行为。在视频中,Ilya 讨论了人工智能的发展,特别是大型语言模型和通用人工智能的研究。人工智能技术的进步对社会和经济有着深远的影响。

💡深度学习

深度学习是机器学习的一个子领域,它通过构建和训练多层神经网络来模拟人脑处理数据的方式。视频中讨论了深度学习的进步,以及它如何推动了像 GPT-3 这样的大型语言模型的发展。

💡神经网络

神经网络是一种模仿人脑神经元结构的计算模型,用于识别模式和处理复杂的数据。在视频中,Ilya 解释了神经网络是如何通过学习来预测文本中的下一个单词,从而实现对语言的理解和生成。

💡GPT-3

GPT-3(Generative Pre-trained Transformer 3)是一个由 OpenAI 开发的大型语言模型,它能够生成连贯的文本、回答问题、翻译语言等。视频中提到了 GPT-3 的发布以及它对世界产生的影响。

💡OpenAI

OpenAI 是一个致力于确保人工智能(AI)技术惠及全人类的研究实验室。视频中,Ilya 作为 OpenAI 的联合创始人和首席科学家,分享了该公司的使命和他在其中的角色。

💡技术创新

技术创新指的是创造或改进产品、服务或流程的新颖方法。视频中,Ilya 强调了他在人工智能领域的创新工作,包括他在深度学习和大型语言模型方面的贡献。

💡意识

意识通常指的是个体对自己思想、感觉和周围环境的感知和体验。视频中,Ilya 探讨了意识的本质以及它是否可能在人工智能系统中出现。

💡伦理

伦理涉及道德原则和行为准则,它指导个人和社会的行为。在视频中,Ilya 讨论了作为 OpenAI 首席科学家,他如何看待公司在人工智能发展中的伦理责任。

💡创业

创业是指创建新的业务或企业的过程,通常涉及创新、风险承担和资源整合。视频中,Ilya 作为一个成功的创业者,分享了他对有志于在人工智能和创业领域发展的斯坦福学生的建议。

💡数据集

数据集是用于训练机器学习模型的信息集合。在视频中,Ilya 讨论了使用特定数据集进行训练以提高人工智能在特定领域(如法律或医学)的性能的重要性。

Highlights

Ilya Sutskever 是 OpenAI 的联合创始人兼首席科学家,致力于构建人工通用智能(AGI)以造福全人类。

Ilya 在俄罗斯出生,五岁时移居以色列,并在多伦多大学完成数学学士学位。

Ilya 认为人工神经网络与生物神经网络在某种程度上是相似的,这种假设推动了深度学习的发展。

Ilya 强调了反向传播算法的重要性,这是一种让人工神经网络通过经验学习的数学方程。

大型语言模型(如GPT-3)通过训练神经网络猜测文本中的下一个单词,从而提高其预测能力。

Ilya 认为,尽管人工智能在某些方面超越了人类,但在学习方式上,人类和机器仍然存在显著差异。

OpenAI 的使命是确保人工通用智能(AGI)造福全人类,从非营利和开源开始,现已转变为有利润和闭源的公司。

Ilya 讨论了 OpenAI 与微软的合作关系,以及如何确保投资者的财务责任与 OpenAI 使命的一致性。

Ilya 认为,随着 AI 能力的增长,未来可能会有一天出于安全考虑而不再开源这些模型。

Ilya 讨论了意识问题,以及 AI 是否可能发展出意识,他认为这是一个复杂且难以定义的概念。

Ilya 认为深度学习将继续取得进展,不仅仅是通过扩大模型规模,还包括在多个层面上改进深度学习堆栈。

Ilya 强调了在 AI 研究中,信任自己的直觉和独特见解的重要性。

OpenAI 的成功关键绩效指标(KPI)包括技术进步、研究执行情况以及对 AI 系统的控制和引导。

Ilya 认为,尽管 AI 技术已经整合到团队工作中,提高了生产力,但目前对团队动态的影响并不显著。

Ilya 讨论了 OpenAI 企业结构的独特性,它是一种有利润上限的公司,最终将转变为非营利组织。

Ilya 认为,未来的 AI 系统将在多个小的改进和一些大的改进的基础上,变得更加强大和复杂。

Ilya 讨论了在 AI 领域中,通用训练与专门训练的优劣,以及如何结合两者以取得最佳效果。

Ilya 认为,尽管 AI 模型如 GPT-3 在诗歌创作方面表现出色,但在团队工作中,它们目前还没有显著改变团队动态。

Transcripts

play00:00

who you are defines how you build

play00:05

welcome YouTube and Stanford communities

play00:08

to the entrepreneurial thought leaders

play00:10

seminar

play00:12

um brought to you by stvp the

play00:13

entrepreneurship Center in the School of

play00:15

Engineering at Stanford and basis The

play00:17

Business Association of Stanford

play00:18

entrepreneurial students today we are so

play00:21

honored to have Ilia suitskiver here at

play00:25

ETL Ilya is the co-founder and chief

play00:28

scientist of open AI which aims to build

play00:31

artificial general intelligence for the

play00:34

benefit of all Humanity Elon Musk and

play00:38

others have cited that Ilya is the

play00:40

foundational mind behind the large

play00:41

language model generative pre-trained

play00:43

Transformer 3 or gpt3 and its

play00:47

public-facing product chat gbt a few

play00:50

product releases have created as much

play00:52

excitement Intrigue and fear as the

play00:56

release of chat gbt in November of 2022.

play00:59

Ilia was Ilia is another example of how

play01:03

the U.S and the world has been the

play01:05

beneficiary of amazing talent from

play01:07

Israel and Russia is Elia was born in

play01:10

Russia and then when he was five he

play01:12

moved to Israel where he grew up and he

play01:14

spent um at the first half of undergrad

play01:17

even in Israel and then he transferred

play01:19

and went to the University of Toronto to

play01:21

complete his bachelor's degree in

play01:23

mathematics he went on to get a master's

play01:25

in PhD in computer science from the

play01:27

University of Toronto and then came over

play01:29

here to the farm and did a short stint

play01:31

with Andrew ing before returning back

play01:34

to Toronto to work under his advisor

play01:37

Jeffrey Hintz research company DNN

play01:39

research Google then acquired DNN

play01:42

research shortly thereafter in 2013 and

play01:45

Ilya became a research scientist as part

play01:48

of Google brain and in 2015 he left

play01:51

Google to become a director of the then

play01:53

newly formed open AI it's hard to

play01:57

overestimate the impact that chat gbt

play02:00

has had on the world since its release

play02:01

in November of last year and while it

play02:04

feels like chat gbt came out of nowhere

play02:06

to turn the world on its head the truth

play02:07

is there's a deep history of innovation

play02:09

that has led to that moment and as

play02:12

profound as chat gbt is Ilia is No

play02:14

Stranger in uttering in discontinuous

play02:16

leaps of innovation and AI Jeff Hinton

play02:19

has said that Ilya was the main impetus

play02:20

for Alex net which was the convolutional

play02:22

neural network in 2012 that is

play02:25

attributed to setting off the deep

play02:26

learning Revolution that has led to the

play02:28

moment that we are now in and of course

play02:31

it was seven years since the founding of

play02:32

open AI that chat GB T was finally

play02:35

Unleashed to the world Ilyas was elected

play02:37

a fellow of the Royal Society in 2022

play02:40

he's been named to the MIT tech review

play02:42

35 under 35 list in 2015. he's received

play02:45

the University of Toronto's innovator of

play02:46

the Year award in 2014 and the Google

play02:48

graduate Fellowship from 2010 to 2012.

play02:51

so with that everybody please give a

play02:53

virtual warm Round of Applause and

play02:56

welcome for Ilia to the entrepreneurial

play02:57

thought leader seminar so Ilya imagine

play02:59

lots of Applause and you're always

play03:01

invited back onto the farm physically

play03:03

whenever you are able so Ilya there's so

play03:07

much to discuss and I know we're gonna

play03:08

have solo time and we have quite a broad

play03:10

range of fluency around the audience in

play03:13

terms of chat gbt and lot large language

play03:16

models I wanted to start off with just a

play03:18

quick question on the technology which

play03:20

is just the key technology underlying

play03:22

open Ai and generative AI more broadly

play03:25

is large language models can you

play03:28

describe the technology in simple terms

play03:30

and now that you're at the Forefront of

play03:32

the tech can you share would have

play03:34

surprised you the most about what the

play03:36

tech can do that you didn't anticipate

play03:39

yeah

play03:41

I I can't explain

play03:44

well this technology is and why it works

play03:49

I think the explanation for why it works

play03:51

is both simple and extremely beautiful

play03:56

and it works for the following reason

play04:01

so you know how

play04:03

the human brain is our best example of

play04:06

intelligence

play04:08

in in the world

play04:11

and we know that the human brain is made

play04:13

out of a large number of neurons a very

play04:16

very large number of neurons

play04:18

neuroscientists have studied neurons for

play04:20

many decades to try to understand how

play04:22

they work precisely and while the

play04:25

operation of our biological neurons are

play04:28

still mysterious

play04:29

there's been a pretty bold conjecture

play04:31

made by the earliest deep learning

play04:34

researchers in the 40s

play04:37

the idea that an artificial neuron

play04:42

the ones that we have in our artificial

play04:44

neural networks kind of sort of similar

play04:48

to a biological neuron if you squint so

play04:51

that's there's an assumption there

play04:54

and we can just run with this assumption

play04:56

now one of the nice things about these

play04:59

artificial neurons is that you can

play05:01

they are much simpler and you can study

play05:03

them mathematically

play05:05

and a very important breakthrough that

play05:06

was done by the very very early deep

play05:09

learning Pioneers before it was known as

play05:12

deep learning was the discovery of the

play05:14

back propagation algorithm

play05:16

which is a mathematical equation for how

play05:18

these artificial neural networks should

play05:20

learn

play05:22

it

play05:23

provides us with a way of

play05:27

taking a large computer

play05:31

and implementing this neural network in

play05:34

code and then there would be there is an

play05:36

equation that we can code up that tells

play05:38

us how this neural network should adapt

play05:40

its connections to learn from experience

play05:45

now a lot of additional further progress

play05:47

had to do with understanding just how

play05:50

good and how capable this learning

play05:53

procedure is

play05:54

and what are the exact conditions under

play05:57

which this learning procedure works well

play05:59

it's although this is although we do

play06:01

with computers it was a little bit of an

play06:03

experimental science a little bit like

play06:04

biology where you have something that is

play06:08

you know like like like a local

play06:10

biological experiment a little bit

play06:12

and

play06:15

so then a lot of the progress with deep

play06:17

learning basically boils down to this

play06:20

we can build these neural networks in

play06:23

our large computers and we can train

play06:25

them on some data we can train those

play06:29

large neural networks to do whatever it

play06:32

is that the data asks them to do

play06:35

now the idea of a large language model

play06:38

is that if you have a very large neural

play06:40

network

play06:41

perhaps one that's now not that far from

play06:45

like these neural networks are pretty

play06:47

large

play06:48

and we train them on the task

play06:51

to guess the next word from a bunch of

play06:54

previous words

play06:56

in text

play06:58

so this is the idea of a large language

play07:01

model you train a big neural network to

play07:04

guess the next word from a previous from

play07:06

the previous words in text and you want

play07:09

the neural network to guess the next

play07:11

word as accurately as possible

play07:15

now the thing that happens here is we

play07:18

need to come back to our original

play07:20

assumption that maybe biological neurons

play07:22

aren't that different from artificial

play07:25

neurons and so if you have a large

play07:27

neural network like this that guesses

play07:29

the next word really well maybe it will

play07:31

be not that different from what people

play07:33

do when they speak and that's what you

play07:35

get so now when you talk to

play07:37

a neural network like this it's because

play07:40

it has such a great such an excellent

play07:43

sense of what comes next what word comes

play07:46

next it can narrow down it can't see the

play07:48

future but it can narrow down the

play07:50

possibilities correctly from its

play07:52

understanding

play07:53

being able to guess what comes next very

play07:55

very accurately requires prediction

play07:57

which is the way you operationalize

play08:00

understanding

play08:01

what does it mean for a neural network

play08:03

to understand it's hard to come up with

play08:05

a clean answer but it is very easy to

play08:09

measure and optimize the Network's

play08:12

prediction error of the next word

play08:15

so we say we want understanding but we

play08:18

can optimize prediction and that's what

play08:20

we do

play08:21

and that's how you get this current

play08:23

large language models these are neural

play08:25

networks which are large they are

play08:27

trained with the back propagation

play08:29

algorithm which is very capable and if

play08:32

you allow yourself to imagine that an

play08:34

artificial neuron is not that different

play08:35

from a biological neuron then yeah like

play08:38

our brains are doing are capable of

play08:40

doing a pretty good job at guessing the

play08:42

next word if you pay if you pay very

play08:44

close attention

play08:46

so so if I let I love that and I just

play08:49

want to make this more concrete so just

play08:51

to push that analogy further between the

play08:53

biological brain and these neural um uh

play08:56

analog physical networks digital

play08:59

networks

play09:00

um if the human if if we consider you

play09:04

know before it was considered untenable

play09:05

for these machines to learn now it's a

play09:07

given that they can learn or do this

play09:10

um uh do predictive outcomes of what's

play09:12

going to come next if a human is at 1X

play09:15

learning and you have the visibility

play09:18

into the most recent chat gbt models

play09:21

what would you put the most recent chat

play09:22

gbt model as a ratio of where the humans

play09:26

are at so if humans are at 1X what's

play09:28

chat gpdn

play09:30

you know it's a bit hard to make direct

play09:33

comparisons between our artificial

play09:36

neural networks and people because at

play09:39

present

play09:41

people

play09:43

are able

play09:45

to learn more from a lot less data

play09:50

this is why these neural networks like

play09:52

Chad GPT are trained on so much data to

play09:55

compensate for their initial slow

play09:58

learning ability

play09:59

you know as we train these neural

play10:01

networks and we make them better

play10:03

faster learning abilities start to

play10:06

emerge

play10:07

but overall

play10:15

overall it is the case that we are we

play10:18

are quite different the way people learn

play10:20

is quite different from the way these

play10:21

neural networks learn like one example

play10:24

might be you know these neural networks

play10:26

they are

play10:27

you know solidly good at math or

play10:28

programming

play10:30

but like the amount of math books they

play10:32

needed to get let's say good at

play10:34

something like

play10:36

calculus is very high or as a person

play10:39

would need a fairly you know two

play10:41

textbooks and maybe 200 exercises and

play10:44

you're pretty pretty much good to go

play10:46

so there is just to get an order of

play10:48

magnitude sense if you relax the data

play10:50

constraint so if you let the machine

play10:52

consume as much data as it needs do you

play10:54

think it's operating at like one-tenth

play10:56

of a human right now or

play10:58

you know it's quite hard to answer that

play11:01

question still and let me tell you why I

play11:03

hesitate to like I think that any figure

play11:05

like this will be misleading and I want

play11:07

to explain why like because right now

play11:09

any such neural network is obviously

play11:14

very superhuman when it comes to the

play11:17

breadth of its knowledge and to the very

play11:19

large number of skills that these neural

play11:21

networks have for example they're very

play11:23

good at poetry and they're very good you

play11:25

know like they can talk eloquently about

play11:27

any topic pretty much

play11:29

and they can talk about historical

play11:31

events and lots of things like this

play11:34

on the other hand on the other hand

play11:37

people

play11:38

can go deep and they do go deep so you

play11:41

may have an expert like someone who

play11:44

understands something very deeply

play11:45

despite having read only a small amount

play11:48

of documents let's say on the topic so

play11:52

because of this difference I really

play11:55

hesitated to answer the question in

play11:58

terms of oh yeah it's like some some

play12:00

number between zero do you think there

play12:02

is a singularity point where the

play12:04

machines will surpass the humans in

play12:07

terms of the pace of learning and

play12:09

adaption yeah and when do you think that

play12:12

point will occur

play12:13

I don't know I don't know when it will

play12:15

occur I think some additional advances

play12:17

will need to do will happen but

play12:20

you know I absolutely would not bet

play12:22

against this point occurring at some at

play12:24

some point can you give me a range is it

play12:26

at some point next month is it next year

play12:30

you know

play12:32

I think it's like the the uncertainty on

play12:35

this thing is quite High because these

play12:37

advances I can imagine it can take in

play12:40

quite a while I can imagine it can take

play12:42

any disappoint in a long time I can also

play12:44

imagine it's taking

play12:48

you know

play12:54

some

play12:55

number of years but it's just very it's

play12:59

very hard to give a Cali braided answer

play13:02

and I I know there's lots of push

play13:04

forward so I'm going to ask one more

play13:05

question then move on to some of the

play13:07

other issues but um I know I read that

play13:09

when you were a child you were disturbed

play13:11

by the notion of Consciousness and I

play13:13

wasn't sure what that what that word

play13:14

meant Disturbed but I'm curious do you

play13:16

view Consciousness or sentience or

play13:19

self-awareness as an extenuation of

play13:22

learning do you think that that is

play13:23

something that also is an inevitability

play13:25

that will happen or not

play13:28

yeah I mean on the Consciousness

play13:30

questions

play13:32

like yeah I was as a child that would

play13:35

like you know look into my in my hand

play13:36

and I would be like how can it be that

play13:38

this is my hand that I get to see like I

play13:42

something of this nature I don't know

play13:43

how to explain it much better so that's

play13:46

been something I was curious about

play13:48

you know it's

play13:52

It's Tricky with Consciousness because

play13:54

how do you define it it's something that

play13:56

the looted definition for a long time

play13:58

and how can you test it in a system

play14:00

maybe there is a system which acts

play14:02

perfectly right but um perfectly the way

play14:06

you'd expect

play14:08

um

play14:09

a conscious system would act yet maybe

play14:11

it won't be conscious for some reason I

play14:14

do think there is a very simple way to

play14:17

there's there is an experiment which we

play14:19

could run on an AI system which we can't

play14:21

run on which we can't run just yet but

play14:24

maybe

play14:25

in like the Future Point when the AI

play14:28

learns very very quickly from less from

play14:30

less data we could do the following

play14:31

experiment

play14:33

very carefully with very carefully

play14:37

curate the data

play14:38

such that we never ever mention anything

play14:41

about consciousness it would only say

play14:44

you know here is

play14:45

here's a ball and here's a castle and

play14:47

here is like a little toy like you would

play14:50

imagine imagine you'd have data of this

play14:52

sort it would be very controlled maybe

play14:54

we'd have some number of years worth of

play14:57

this kind of training data

play14:58

maybe it would be maybe such an AI

play15:01

system would be interacting with a lot

play15:02

of different teachers learning from them

play15:05

but all very carefully you never ever

play15:07

mentioned Consciousness you don't talk

play15:09

about

play15:10

people don't talk about

play15:12

anything except for the most surface

play15:15

level Notions of their experience and

play15:18

then at some point you sit down this Ai

play15:20

and you say Okay I want to tell you

play15:22

about Consciousness it's the stain

play15:23

that's a little bit not well understood

play15:25

people disagree about it but that's how

play15:27

they describe it and imagine if the AI

play15:29

then goes and says oh my god I've been

play15:32

feeling the same thing but I didn't know

play15:34

how to articulate it that would be okay

play15:36

that would be definitely something to

play15:39

think about it's like if the AI was just

play15:42

trained on very mundane data around

play15:44

objects and going from place to place or

play15:47

maybe you know

play15:51

something like this from a very narrow

play15:53

set of Concepts we would never ever

play15:55

mention that and if it could somehow

play15:57

eloquently correctly talk about it in a

play16:00

way that we would recognize that would

play16:02

be convincing

play16:03

and do you think of it as a some as

play16:05

Consciousness as something of degree or

play16:08

is it something more binary uh

play16:12

I think it's something that's more a

play16:15

matter of degree

play16:18

I think that

play16:20

I think that like you know

play16:24

let's say if a person is very tired

play16:27

extremely tired and maybe drunk then

play16:30

perhaps if that's when when someone is

play16:32

in that state and maybe their

play16:33

Consciousness is already reduced to some

play16:35

degree

play16:37

I can imagine that animals have a more

play16:40

reduced form of Consciousness if you

play16:41

imagine going from

play16:43

you know large primates maybe dogs cats

play16:47

and then eventually you get mice you

play16:49

might get an insect like

play16:51

feels like I would say it's pretty

play16:53

continuous yeah

play16:54

okay I want to move on even though I

play16:57

could I would love to keep asking more

play16:58

questions along the lines of the

play16:59

technology but I want to move on to

play17:01

talking about the mission of openai and

play17:03

how you perceive or any issues around

play17:05

ethics and your role as Chief science

play17:08

officer how ethics informs if at all how

play17:11

you think about your role and so let me

play17:14

just lay a couple Foundation points out

play17:15

and then have you speak

play17:17

um as you know open ai's mission is to

play17:19

ensure the art of that artificial

play17:21

general intelligence benefits all of

play17:23

humanity and it started off as a

play17:26

non-profit and open source and it is now

play17:28

a for-profit and closed-sourced and with

play17:31

a close relationship with Microsoft and

play17:33

Elon Musk who I believe recruited you to

play17:35

originally join open Ai and gave 100

play17:38

million dollars when it was a non-profit

play17:40

has says that the original Vision was to

play17:42

create a counterweight to Google and the

play17:44

corporate world and he didn't want to

play17:45

have a world in which AI which is has

play17:48

which he perceives and others can have

play17:50

an existential threat to humanity to be

play17:53

solely in the holds of of corporate of a

play17:57

for-profit

play17:58

um and now

play18:00

open AI is neither open nor exclusively

play18:04

a non-profit it's also a for-profit with

play18:06

close ties to Microsoft and it looks

play18:08

like the world may be headed towards

play18:10

um a private duopoly between Microsoft

play18:12

and Google

play18:14

can you shed light on the calculus to

play18:17

shift from a for-profit to a non-profit

play18:19

and did you weigh in the ethics of that

play18:22

decision and do ethics play a role in

play18:26

how you conceive of your role as the

play18:27

chief science officer or do you view it

play18:29

more as something that somebody else

play18:31

should handle and you are mainly just

play18:33

tasked with pushing the technology

play18:34

forward

play18:40

yeah so this question is many parts let

play18:42

me yeah let me think about the best way

play18:45

to to approach it

play18:48

so there are several parts there is the

play18:52

there is the question around open source

play18:55

versus closed source

play18:57

there is a question around non-profit

play19:00

versus for-profit

play19:02

and the connection with Microsoft

play19:05

and how to see that in the context of

play19:08

Elon musk's recent comments

play19:10

and then the question about how I see my

play19:12

role in this maybe I'll start with that

play19:14

because I think that's easier Okay so

play19:24

I feel

play19:26

yeah

play19:28

the way I see my role

play19:31

I feel a lot I I feel

play19:33

direct responsibility for whatever open

play19:36

AI does even though

play19:39

I my role is primarily around advancing

play19:43

the science

play19:44

it is still the case I'm one of the

play19:46

founders of the company

play19:48

and

play19:49

ultimately I care a lot

play19:53

about open ai's overall impact

play19:57

now I want to go so with this context I

play20:00

want to go and talk about the open

play20:02

source versus closed source

play20:04

and the non-profit versus for-profit

play20:07

and I want to start with open source

play20:09

which is closed source

play20:10

because I think that

play20:13

you know the challenge with AI

play20:16

is that AI is so all encompassing

play20:19

encompassing

play20:21

and it comes with many different

play20:25

challenges it comes with many many

play20:27

different dentures

play20:31

which come into conflict with each other

play20:34

and I think the open source versus

play20:35

closed source is a great example of that

play20:38

why is it desirable well let me put it

play20:42

this way

play20:43

what are some reasons

play20:46

for which it is desirable to open source

play20:48

AI

play20:50

the answer there would be

play20:53

to

play20:55

prevent

play20:56

concentration of power

play20:59

in the hands of those who are building

play21:01

the AI so if you are in a world where

play21:04

let's say there is only a small number

play21:07

of companies you might that control this

play21:09

very powerful technology

play21:10

you might say this is an undesirable

play21:13

world and that AI should be open and

play21:16

that anyone could use the AI this is the

play21:18

argument for open source

play21:20

but this argument

play21:23

you know of course you know to State the

play21:25

obvious there are near-term commercial

play21:28

incentives against open source but there

play21:31

is another longer term argument against

play21:33

open sourcing as well which is

play21:36

if we believe

play21:38

if one believes that eventually AI is

play21:41

going to be unbelievably powerful

play21:45

if we get to a point where your AI is so

play21:48

powerful where you can just tell it hey

play21:50

can you autonomously

play21:54

create a biological I don't know a

play21:58

biological research lab

play22:00

autonomously

play22:01

do all the paperwork render space hire

play22:05

the technicians aggregate experiments do

play22:07

all this autonomously

play22:08

like that starts to get

play22:11

incredible that starts to get like

play22:13

mind-bandingly powerful should this be

play22:15

open sourced also

play22:17

so my position on the open source

play22:20

question is that I think that

play22:23

I think that there is a maybe

play22:26

a level of capability you can think

play22:27

about these neural networks in terms of

play22:29

capability

play22:31

how capable they are how smart they are

play22:34

how much how many how much how much can

play22:36

they do

play22:38

when the capability is on the lower end

play22:40

I think open sourcing is a great thing

play22:43

but at some point and you know there can

play22:46

be debate about where the pointer is

play22:48

but I would say that at some point the

play22:50

capability will become so vast that it

play22:53

will be obviously irresponsible to open

play22:55

source models and was that the driver

play22:58

Behind Closed sourcing it or was it

play23:00

driven by a a devil's compact or

play23:03

business necessity to get cash in uh

play23:06

from Microsoft or others to support the

play23:08

viability of the business was the

play23:10

decision making to close it down

play23:12

actually driven by that line of

play23:13

reasoning or was it driven by more so

play23:17

it's so so the way I'd articulate it you

play23:21

know my view is that the current level

play23:23

of capability is still not that high

play23:27

where it will be the safety

play23:29

consideration it will drive the close

play23:31

closed Source in the model this kind of

play23:34

this kind of research

play23:35

so in other words a claim that it goes

play23:38

in phases right now it is indeed the

play23:41

competitive phase but I claim that as

play23:44

the capabilities of these models keep

play23:45

increasing

play23:46

there will come a day where it will be

play23:49

the safety consideration that will be

play23:51

the obvious and immediate driver to not

play23:54

open source these models

play23:56

so this is the open source versus closed

play23:59

Source but your question had enough but

play24:01

your question in another part which is

play24:03

non-profit versus for-profit

play24:06

and we can talk about that also

play24:09

you know indeed it would be preferable

play24:13

in a certain meaningful sense if open AI

play24:17

could just be a for a non-profit from

play24:19

now until the mission of open AI is

play24:22

complete

play24:24

however one of the things that's worth

play24:26

pointing out is the very significant

play24:29

cost of these data centers I'm sure

play24:32

you're reading about various AI startups

play24:34

and the amount of money they are raising

play24:35

the great majority of which goes to the

play24:38

cloud providers

play24:39

why is that

play24:42

well the reason so much money is needed

play24:44

is because

play24:46

this is

play24:48

the nature of these large neural

play24:50

networks they need the compute end of

play24:53

story you can see something like this

play24:56

that's all you can see a divide that's

play24:58

now happening between Academia

play25:01

and the AI companies

play25:03

so for a long time for many decades

play25:06

Cutting Edge research in AI took place

play25:08

in academic departments in universities

play25:12

that cap being the case up until the

play25:15

mid-2010s

play25:18

but at some point when the complexity

play25:21

and the cost of this project started to

play25:23

get very large it no longer remained

play25:25

possible for universities to be

play25:27

competitive and now universities need a

play25:30

University Research in AI needs to find

play25:33

some other way in which to contribute

play25:35

those ways exist they're just different

play25:38

from the way they're used to and

play25:40

different from the way the companies are

play25:42

contributing right now

play25:44

now

play25:47

with this context you're saying okay

play25:50

the thing about non-profit a non-profit

play25:52

is the people who give money to a

play25:53

non-profit never get to see any any of

play25:56

it back it is a real donation

play25:59

and believe it or not it is quite a bit

play26:02

harder to convince people to give money

play26:04

to a non-profit and so we so so we think

play26:07

what's what's the solution there or what

play26:09

is a good course of action

play26:11

so we came up with an idea

play26:15

that to my knowledge is unique

play26:18

in all corporate structures in the world

play26:21

the open air corporate structure is

play26:23

absolutely unique

play26:25

open AI is not a for-profit company it

play26:28

is a capped profit company

play26:30

and I'd like to explain what that means

play26:33

what that means

play26:36

is that

play26:39

equity in open AI can be better seen as

play26:43

Bond rather than equity in a normal

play26:46

company the main feature of a bond is

play26:49

that once it's paid out it's gone

play26:52

so in other words

play26:53

open AI has a finite obligation to its

play26:57

investors as opposed to an infinite

play26:59

obligation to that normal companies have

play27:02

and does that include the founders do

play27:03

the founders have equity in open AI

play27:06

so Sam Altman does not have equity

play27:10

but the other Founders do and is it

play27:12

capped or is it unlimited it is capped

play27:15

and how does that cap is that capped at

play27:18

because the the founders I presume

play27:19

didn't buy in unless it's capped at the

play27:21

nominal

play27:23

Share value

play27:25

um

play27:27

I'm not sure I understand the question

play27:29

precisely but what I can say like what

play27:31

what I can answer the part which I do

play27:33

understand which is like

play27:36

there is certainly like it isn't there

play27:39

are it is a different it is different

play27:41

from normal startup Equity but there are

play27:43

some similarities as well where the

play27:45

earlier you join the company

play27:47

the higher the cap is because then

play27:51

the larger cap is needed to attract the

play27:54

initial investors as the company

play27:56

continues to succeed the cap decreases

play27:58

and why is that important it's important

play28:01

because it means that the company

play28:04

one once when once all the obligation to

play28:07

investors and employees are paid out

play28:09

open AI becomes a non-profit again

play28:12

and you can say this is totally crazy

play28:14

what are you talking about like it's not

play28:16

going to change anything

play28:18

but it's worth considering what we

play28:22

expect like it's worth

play28:25

looking at what we think AI will be

play28:28

I mean we can look at what AI is today

play28:31

and I think it is not at all

play28:33

inconceivable

play28:34

for open AI tool achieve its

play28:39

to pay out its obligation to the

play28:42

investors and employees become a

play28:44

non-profit at around the time when

play28:46

perhaps the computers will become so

play28:48

capable where the economic destruction

play28:50

will be very big where this transition

play28:53

will be very beneficial

play28:54

so this is the answer on the cap profit

play28:57

versus

play28:59

non-profit

play29:00

there was a last part to your question I

play29:02

know I'm speaking for a while but the

play29:04

question had many parts the last part of

play29:06

your question is the Microsoft

play29:08

relationship

play29:09

and

play29:12

so here

play29:14

the thing that's very fortunate

play29:17

is that Microsoft is a

play29:21

there thinking about these questions the

play29:24

right way they understand

play29:27

the potential and the gravity of AGI and

play29:31

so for example on the on all the

play29:34

investor documents that any investor in

play29:36

open AI has signed and by the way

play29:38

Microsoft is an investor into open AI

play29:40

which is a very different relationship

play29:42

from the

play29:43

deepmind

play29:46

any anyone who signed any document any

play29:48

investment document there is a a purple

play29:50

rectangle at the top of the investment

play29:54

document which says that the fiduciary

play29:57

duty of open AI is to the open AI

play29:59

mission

play30:00

which means that you run the risk of

play30:03

potentially losing all your money if the

play30:06

mission comes in conflict

play30:10

so this is something that all the

play30:12

investors have signed

play30:14

and let me just make this clear for

play30:15

everybody because Google Google acquired

play30:17

deepmind so deepmind was just an asset

play30:19

inside of Google but beholden to Google

play30:20

you're making the distinction that with

play30:22

openai Microsoft is an investor and so

play30:25

beholden to this fiduciary duty for the

play30:27

mission of openai which is held by the

play30:29

non-profit which is a is is a a GP or an

play30:34

LP in the

play30:36

um in in the for-profit

play30:38

um okay understood yeah so it's not

play30:41

something like this you know I am

play30:43

you know there are people

play30:46

I can't tell you the precise details

play30:50

yeah but so but this is the general

play30:54

picture

play30:55

and you know some have claimed though

play30:57

now especially it uh um Steve Wozniak

play30:59

the co-founder of apple and Elon Muska

play31:01

famously signed this very public

play31:02

petition saying that the point of no

play31:04

return is already passed or we're

play31:06

approaching it where it's going to be

play31:07

impossible to reign in Ai and it's and

play31:11

it's it's repercussions if we don't halt

play31:13

it now and they've called for halting AI

play31:17

um

play31:18

I'm curious

play31:19

on you are a world citizen Ilia you were

play31:22

born in Russia you were raised in Israel

play31:25

you're Canadian

play31:27

um and I'm and it's open ai's response

play31:31

to that public petition was

play31:33

um I know Sam basically said that you

play31:35

know this wasn't the right way to go

play31:36

about doing that but also in parallel

play31:38

Sam is on a world tour with many

play31:41

countries that also can be antagonistic

play31:44

towards the West

play31:46

are there any citizen obligations

play31:48

ethical obligations that you think also

play31:51

overweigh your your technological

play31:55

obligations when it comes to spreading

play31:57

the technology around the world right

play31:59

now through open AI do you think that

play32:00

should be beholden to a regulation or

play32:04

some oversight

play32:10

let me think

play32:16

once again the question had a number of

play32:17

Parts did I apologize I'm trying to give

play32:20

you the so you can respond however you

play32:22

want to on that I know we're going to

play32:24

come out of off of time so I just want

play32:26

to give you the mic and just share

play32:28

everything that's on my mind and you can

play32:29

decide how you want to handle it yeah

play32:30

thank you

play32:32

I mean

play32:34

you know

play32:38

it is true

play32:40

that AI is going to become truly

play32:43

extremely powerful and truly extremely

play32:45

transformative

play32:47

and I do think

play32:50

that we will want to move to a world

play32:52

with sensible government regulations and

play32:55

there you know there are several

play32:56

Dimensions to it

play32:59

we want to be in a world where there are

play33:01

clear rules about for example training

play33:05

more powerful neural networks

play33:09

we want there to be some kind of careful

play33:13

evaluation careful prediction of these

play33:17

of what we expect these neural networks

play33:20

of what they can do today and on what we

play33:22

expect them to be able to do let's say

play33:24

in a year from now or by the time they

play33:27

finish training I think all these things

play33:29

will be very necessary in order to

play33:35

like rational like rationally

play33:40

I wouldn't use the word slow down the

play33:42

progress I would use the term you want

play33:45

to make it so that the progress is

play33:47

sensible so that with each step we've

play33:50

done the homework and indeed we can make

play33:54

a credible story that okay

play33:57

the neural network the system that we've

play34:00

trained it has we are doing this and

play34:02

here all the steps and it's been

play34:05

verified or certified I think that is

play34:09

the world that we are headed to which I

play34:11

think is correct

play34:13

and as for the citizen obligation I feel

play34:16

like

play34:23

I mean

play34:25

15 what

play34:27

I'll answer it like this

play34:29

like I think I think like there are

play34:32

there are two answers to it so obviously

play34:33

you know I live I live in the United

play34:35

States and I really like it here and I

play34:37

want and I want this place to flourish

play34:39

as much as possible

play34:40

I care about that

play34:43

I think that of course there will be

play34:45

lots of but the world is much more than

play34:47

just the US

play34:48

and I think that these are the kind of

play34:50

questions which I feel a little bit

play34:54

let's say outside of my expertise how

play34:57

these

play34:58

between country relationships work out

play35:01

but I'm sure there will be lots of

play35:02

discussions there as well

play35:05

yeah

play35:06

um Julia can I turn a little bit towards

play35:08

strategy

play35:09

um I'm curious for you guys internally

play35:12

what metrics do you track as your North

play35:14

Star what are the most sacred kpis that

play35:18

you use to measure open ai's success

play35:20

right now

play35:24

the most sacred kpis

play35:27

you know I think this is also the kind

play35:29

of question where maybe different people

play35:31

will give you different answers

play35:33

but I would say I would say that there

play35:35

are

play35:36

if I were to really narrow it down

play35:39

I would say that there are

play35:43

there is a couple of really important

play35:45

kpi of a really important dimensions of

play35:48

progress

play35:49

one is undeniably the technical progress

play35:53

are we doing good research

play35:57

do we understand our systems better are

play35:59

we able to train them better can we

play36:01

control them better I is our

play36:05

is ours is our research plan being

play36:08

executed well is our safety plan being

play36:10

executed well how happy are we with it I

play36:13

would say this would be my description

play36:15

of the primary kpi which is do a good

play36:18

job of the technology then there is of

play36:21

course stuff around the product but

play36:25

which I think is cool but I would say

play36:27

that it is really the core technology

play36:30

which is the heart of openai the

play36:32

technology its development

play36:36

and on end

play36:38

its control

play36:40

it's steering

play36:43

and and do you view um right now chat

play36:45

gbt is a destination

play36:47

do you view open AI in the future being

play36:49

a destination that people go to like

play36:52

Google or will it be powering other

play36:55

applications and be the back end or be

play36:58

be you know used as part of the back end

play37:00

infrastructure

play37:02

um is it a destination or is it going to

play37:03

be more behind the scenes

play37:05

um in in five to ten years

play37:08

yeah well I mean things change so fast I

play37:11

I cannot make any claims about

play37:15

five to ten years in terms of the

play37:17

correct shape of the product I imagine a

play37:20

little bit of both perhaps but this kind

play37:23

of question

play37:27

I mean I think it remains to be seen I

play37:29

think there are I think this stuff is

play37:31

still so new

play37:32

okay

play37:33

I'm gonna ask one more question I'm

play37:34

gonna jump to the student questions if

play37:36

you're a student at Stanford today

play37:37

interested in AI if you were you know

play37:40

somebody who wants to be Ilia um what

play37:42

would you focus your time and another

play37:44

second question on this if you're also

play37:46

interested in entrepreneurship

play37:48

um where would you what would you what

play37:49

advice would you give for a Stanford

play37:50

undergrad engineer that's interested in

play37:52

Ai and Entrepreneurship

play37:56

so

play37:57

I think on the first one

play38:03

it's always hard to give generic advice

play38:06

like this

play38:08

but

play38:10

I can still provide some generic advice

play38:12

nonetheless

play38:14

and I think it's something like

play38:20

it it is generally a good idea to lean

play38:24

into

play38:25

one's unique predispositions

play38:28

you know every you know why if you think

play38:31

if you look if you think about the set

play38:34

of let's say inclinations or skills or

play38:36

talents that the person might have the

play38:39

combination is pretty rare so leaning

play38:41

into that is a very good idea no matter

play38:44

which direction you choose to go look to

play38:46

going and then on the AI research

play38:53

like I would say

play38:55

I would say that there

play38:58

you know

play39:00

I could say something but even but there

play39:02

especially you want to lean into your

play39:04

own ideas and really ask yourself what

play39:07

can you is is there something

play39:10

that's totally obvious to you that makes

play39:13

you go why is everyone else not getting

play39:15

it if you feel like this that's a good

play39:18

sign it means that you might be able

play39:21

that that you you want to lean into that

play39:24

and explore it and see if your instinct

play39:27

is true or not it may not be true

play39:29

but you know my my advisor Jeff Hinton

play39:31

says this thing which I really like he

play39:33

says you should trust your intuition

play39:35

because if your intuition is good you go

play39:38

really far and if it's not good then

play39:40

there's nothing you can do

play39:42

hmm

play39:43

and as far as entrepreneurship is

play39:45

concerned

play39:47

I feel like

play39:49

this is a place where the unique

play39:51

perspective is even more valuable or

play39:54

maybe equally it's because it's maybe

play39:55

maybe I'll I'll explain why I think it's

play39:57

more valuable than in research well in

play40:00

research it's very valuable too but in

play40:01

entrepreneurship like you need to like

play40:04

almost pull from your unique life

play40:05

experience where you say okay I see this

play40:08

thing I see this technology I see

play40:10

something like take a very very Broad

play40:12

View and see if you can hone in on

play40:15

something and then actually just go for

play40:17

it

play40:19

so that would that would be the

play40:21

conclusion of my generic advice okay

play40:24

which is great that's also great I'm

play40:26

going to move on to the student question

play40:27

so one of the most upvoted question is

play40:29

how do you see the field of deep

play40:30

learning evolving in the next five to

play40:33

ten years

play40:37

let's see

play40:39

you know I expect deep learning to

play40:41

continue to make progress I

play40:44

I expect that

play40:46

you know there was a period of time

play40:48

where

play40:50

a lot of progress came from scaling

play40:54

and you you saw that most in the most

play40:57

pronounced way in going from GPT 1 to

play41:00

gpd3

play41:02

but

play41:04

things will change a little bit the

play41:06

reason the reason that the reason that

play41:08

progress in scaling was so rapid is

play41:10

because people had all these data

play41:12

centers which they weren't using for a

play41:14

single training run

play41:15

so by simply reallocating existing

play41:19

resources you could make a lot of

play41:20

progress

play41:21

and it doesn't take that long

play41:23

necessarily to reallocate existing

play41:25

resources you just need to you know

play41:26

someone just needs to decide to do so

play41:30

it is different now because the training

play41:32

runs are very big and the scaling is not

play41:35

going to be progressing as fast as it

play41:36

used to be because building data center

play41:39

takes time

play41:41

but at the same time I expect deep

play41:43

learning to continue to make progress in

play41:45

uh from other places the Deep learning

play41:48

stack is quite deep and I expect that

play41:52

there will be improvements in many

play41:53

layers of the stack and together they

play41:56

will still lead to progress being very

play41:58

robust

play41:59

and so

play42:01

if I had to guess I'd imagine that there

play42:04

would be maybe

play42:07

I'm certain we will discover new

play42:09

properties which are currently unknown

play42:12

of deep learning and those properties

play42:14

will be utilized and I fully expect that

play42:17

the systems of five to ten years from

play42:19

now will be much much better than once

play42:21

they are we have right now

play42:22

but exactly how it's going to look like

play42:24

I think

play42:26

I think it's a bit harder to answer it's

play42:29

a bit like

play42:31

it's because the improvements that there

play42:33

is there will be maybe a small number of

play42:36

big improvements and also a large number

play42:38

of small improvements all integrated

play42:40

into a large complex engineering

play42:42

artifact

play42:43

and can I ask your you know your

play42:45

co-founder Sam Altman has said that

play42:46

we've reached the limits of what we can

play42:48

achieve by scaling to larger language

play42:51

models is do you agree

play42:53

um and if so you know what then what is

play42:56

the next Innovation Frontier that you're

play42:57

focusing on if that's the case yeah so

play43:02

I think maybe

play43:05

I don't remember I don't know exactly

play43:07

what he said but maybe he meant

play43:09

something like that the age of easy

play43:10

scaling has ended or something like this

play43:14

like of course of course the larger

play43:16

neural Nets will be better but it will

play43:17

be a lot of effort and cost to do them

play43:20

but I think there will be lots of

play43:22

different Frontiers and actually into

play43:24

the question of how can one contribute

play43:26

in deep learning identifying such a

play43:29

frontier perhaps one that's been missed

play43:31

by others is very fruitful

play43:34

and is it can I go even just deeper on

play43:36

that because I think there is this

play43:37

debate about vertical Focus versus

play43:39

General

play43:41

um uh General's training you know is it

play43:43

better do you think there's better

play43:45

performance that can be achieved in

play43:47

particular domains such as law or

play43:49

Medicine by training with special data

play43:51

sets or is it likely that generalist

play43:53

training with all available data will be

play43:55

more beneficial

play43:57

so

play43:59

like at some point we should absolutely

play44:01

expect Specialists training to make a

play44:03

huge impact

play44:04

but the reason we do the generalist

play44:06

training

play44:07

is just so that we can even reach the

play44:10

point where

play44:13

just so that we can reach the point

play44:18

where the neural network can even

play44:20

understand the questions that we are

play44:21

asking and only when it has a very

play44:24

robust understanding only then we can go

play44:27

into specialist training and really

play44:28

benefit from it so yeah I mean I think

play44:30

all these I think these are all fruitful

play44:32

directions

play44:33

but you don't think when do you think

play44:35

we'll be at that point when specialist

play44:37

training

play44:38

is the thing to focus on I mean

play44:44

you know like if you look at people who

play44:47

do open source work people who work with

play44:49

open source models they do a fair bit of

play44:52

this kind of specialist training because

play44:54

they have a fairly underpowered model

play44:56

and they try to get any ounce of

play44:59

performance they can out of it

play45:02

so I would say that this is an example

play45:07

I'd say that this is an example of it

play45:09

happening like it's already happening to

play45:11

some degree it's not a binary it's you

play45:13

might want to think of it as of like a

play45:15

continual Spectrum but do you think that

play45:17

the competitor do you think that the

play45:19

winning Advantage is going to be having

play45:20

these proprietary data sets or is it

play45:23

going to be having a much higher

play45:25

performance large language model when it

play45:28

comes to these applications of AI into

play45:30

verticals

play45:31

so I think it may be productive to think

play45:34

about about an AI like this as a

play45:36

combination of multiple factors where

play45:39

each factor makes a contribution

play45:42

and

play45:44

is it better to have a special data

play45:47

which helps you make

play45:49

your AI better in a particular set of

play45:51

tasks of course is it better to have a

play45:54

more capable base model of course from

play45:56

the perspective of the task so maybe

play45:57

this is the the answer it's not an

play45:59

either or

play46:00

I'm going to move down to the other

play46:02

questions

play46:03

um there's a question on what was the

play46:04

cost of training and developing GPT T3

play46:06

slash four

play46:08

yeah so

play46:11

you know for for obvious reasons I can't

play46:13

comment on that

play46:16

um but there I think there is a you know

play46:19

I think even from our research Community

play46:21

there's a strong desire to be able to

play46:24

get access to

play46:27

um uh to different aspects of open ai's

play46:31

technology and are there any plans for

play46:34

releasing it to researchers or to other

play46:37

startups to encourage more competition

play46:39

and Innovation some of the requests that

play46:41

I've heard are unfettered interactions

play46:43

without safeguards to understand the

play46:46

model's performance model specifications

play46:48

including details on how it was trained

play46:50

and access to the model itself I.E the

play46:53

trained parameters do you want to

play46:54

comment on any of that

play46:56

yeah I mean I think

play47:03

like it's related to our earlier

play47:04

question about open versus closed

play47:07

I think that there are some

play47:11

intermediate approaches which can be

play47:14

very fruitful

play47:15

for example

play47:17

model access and various combinations of

play47:20

that

play47:20

can be very very productive because

play47:23

these mineral networks already have such

play47:25

a large and complicated surface area of

play47:27

behavior

play47:28

and

play47:33

and studying that alone can be extremely

play47:36

interesting look if you have an academic

play47:38

access problem we provide various forms

play47:40

of access to the models and in fact

play47:43

plenty of academic research Labs do

play47:45

study them in this way

play47:46

so

play47:48

I think this kind of approach

play47:52

is viable

play47:54

and it's something that we could that we

play47:56

are doing

play47:57

I know we're coming up on time I want to

play48:00

end with just one final question which

play48:02

is can you just share any unintuitive

play48:04

but compelling use cases for how you

play48:06

love to use chat gbt that others may not

play48:09

know about

play48:10

um

play48:15

so I mean I don't I wouldn't say that

play48:19

it's unknown but I I really enjoy its

play48:22

poem writing ability

play48:24

it can write poems it can rap it can it

play48:28

can be it can be it can be pretty

play48:29

amusing

play48:32

and do you guys use it is it is it an

play48:33

integrated part of the

play48:36

um of teamwork at open I assume it is

play48:39

but I'm curious do you have any insights

play48:41

on how it changes Dynamics with teams

play48:43

when you have ai deeply integrated into

play48:47

you know a human team and how they're

play48:49

working and any insights into to what we

play48:51

may not know but that will come

play48:53

I would say I would say to today the

play48:57

best way to describe the impact is that

play49:00

everyone is a little bit more productive

play49:03

people are a little bit more on top of

play49:04

things I wouldn't say that right now

play49:06

there is a dramatic impact on Dynamics

play49:08

which I can say oh yeah the Dynamics

play49:10

have shifted in this pronounced way

play49:12

okay I'm curious if it depersonalizes

play49:15

conversations because it's the AI bot or

play49:17

maybe it but maybe we're not at that

play49:19

point yet where it's specifically

play49:21

that I definitely I I don't think that's

play49:25

the case and

play49:27

I predict that will not be the case but

play49:30

we'll see

play49:32

well thank you Ilya for a fascinating

play49:34

discussion time is always too short

play49:36

you're always invited back to the farm

play49:39

um we'd love to have you either

play49:40

virtually or in person

play49:42

um so thank you thank you thank you um

play49:43

to our audience thank you for tuning in

play49:45

for this session of the entrepreneurial

play49:47

thought leader series next week where

play49:49

we're going to be joined by the

play49:50

executive chairman and co-founder of

play49:52

OCTA Frederick karist and you can find

play49:54

that event and other future events in

play49:56

this ETL series on our Stanford e-corner

play49:58

YouTube channel and you'll find even

play50:00

more of the videos podcasts and articles

play50:01

about entrepreneurship and Innovation at

play50:03

Stanford e-corner that's

play50:06

ecorner.stanford.edu and as always thank

play50:09

you for tuning in to ETL

play50:20

thank you

Rate This

5.0 / 5 (0 votes)

Related Tags
斯坦福创新创业OpenAI人工智能深度学习Ilya Sutskever技术进步行业影响未来展望学术研究
Do you need a summary in English?