OpenAI Insider Talks About the Future of AGI + Scaling Laws of Neural Nets

Wes Roth
5 Mar 202418:30

Summary

TLDRこの動画は、人工知能(AI)の発展と、将来的に人工汎用知能(AGI)に到達する可能性について議論しています。特に、ニューラルネットワークのパラメーター数が人間の脳と同等になった時に、AGIが実現する可能性があることを示唆しています。また、OpenAIの研究員であるScott Aronsonが、AGIの安全性と制御の問題に取り組んでいることにも言及しています。さらに、AGIの実現が人類に与える影響や倫理的課題についても触れられています。

Takeaways

  • 🧠 人間の脳には約1000兆のパラメーター(シナプス)があり、AIモデルの能力は一般にパラメーター数に比例する。
  • 📈 GPT-3は1750億のパラメーターを持ち、猫の脳に匹敵する規模。完全なAGI(人工般人工知能)を実現するには、人間の脳と同等の1000兆パラメーターが必要とされている。
  • ⚡ AIの進化は主にデータ量とコンピューティングパワーの増加によって促進されてきた。AGIの基本的な理論は1950年代から存在していた。
  • 🔮 一部の専門家は、AGIを実現するための技術的詳細はすでに解決済みで、あとはハードウェアの進化を待つだけだと主張している。
  • 🤖 遠隔労働者が行える作業はほとんどすべてAGIが代替可能になると予測されている。
  • 🕵️‍♂️ スコット・アロンソン氏は量子コンピューター研究者からOpenAIのAI安全性・整合性の研究に携わるようになった。
  • 🤔 GPT-4やOpenAIの最新モデルが1000兆パラメーターに達したという噂は、作者自身があまり信用していないようだ。
  • 📝 スコット・アロンソンの研究「線形光学の計算複雑性」は、量子コンピューターが古典コンピューターで効率的にシミュレートできないことを示唆している。
  • 🔑 AGIを実現する上で重要なポイントは10個程度で、それらはこれまでの様々な研究テキストの中に隠されている可能性がある。
  • ✍️ 作者は今回の内容をやや散漫で分かりづらいと認めているが、AGIの話題はますます興味深くなってきている。

Q & A

  • 人工知能はいつ人間レベルの能力に到達すると考えられていますか?

    -この記事によると、人工知能のパラメーター数が人間の脳のシナプス数に匹敵する約2*10^14個になると、人間レベルの能力に達すると予測されています。ただし、この予測には幅があり、最低でGPT-3レベル、最高で現在の1万倍のパラメーター数が必要となる可能性があると述べられています。

  • GPT-3とはどのようなモデルですか?

    -GPT-3は2020年にリリースされた大規模な言語モデルで、1750億個のパラメーターを持ち、ある程度の推論や文章生成が可能でした。この能力には多くの人が驚きを示しましたが、人間レベルの知能とはまだ程遠い段階でした。

  • ScottAronsonはどのような人物ですか?

    -ScottAronsonは量子コンピューターの研究者で、2022年半ばにOpenAIの人工知能の安全性と整合性に関する部門に入社しました。彼は人工知能の危険性について警鐘を鳴らす一方で、その進歩を阻止すべきではないと主張しています。

  • パラメーター数が多ければ多いほど、人工知能の性能は向上するのでしょうか?

    -はい、おおむねその傾向があると考えられています。パラメーター数が多ければ、ニューラルネットワークに組み込めるデータの量が増え、より複雑なパターンを認識できるようになります。しかし、単にパラメーター数を増やすだけでは不十分で、適切な学習データとアルゴリズムも重要になります。

  • 人工知能がAGI(汎用人工知能)に到達するための技術的詳細は、すでに解決されていると考えられていますか?

    -はい、John Carmackら一部の専門家は、AGIを実現するための技術的詳細は過去数十年の研究で概ね解決されていると主張しています。彫琢が必要なのは、計算能力とデータの確保だけだと述べられています。ただし、この見解には異論もあります。

  • ニューラルネットワークにおけるパラメーターとは具体的に何を指しますか?

    -ニューラルネットワークにおけるパラメーターとは、ノード間の結合の重みや、各ノードへの入力に対する閾値を表すバイアス値のことを指します。これらのパラメーターを調整することで、ネットワークの出力を最適化できます。生物の脳のシナプス結合に相当します。

  • AA (Anthropic AI)のTIIモデルとは何ですか?

    -TIIは、Anthropic AIによって開発されているAGIモデルの名称です。このモデルは、リモートワーカーが行える作業のほとんど全てを人間レベルで実行できることを目標としています。パラメーター数の予測値も示されています。

  • AIの性能予測において、パラメーター数以外に重要な要素はありますか?

    -パラメーター数だけでなく、学習データの質や量、使用されるアルゴリズム、モデルアーキテクチャなども、AIの性能に大きな影響を与えます。単にパラメーターを増やすだけでは、必ずしも性能が向上するわけではありません。適切な設計と学習が重要です。

  • AGI実現の鍵は、新しい技術ブレークスルーにあるのでしょうか?

    -この記事では、AGIを実現するための重要な技術的詳細はすでに過去の研究で明らかになっている可能性があると示唆されています。新しいブレークスルーよりも、既存の知識を組み合わせて適用することが鍵となるかもしれません。

  • この動画の主な主張は何ですか?

    -この動画の主な主張は、人工知能の進化はパラメーター数の増加と適切なデータの組み合わせによって加速する可能性があり、AGIに到達するための技術的詳細はすでに存在している可能性があるということです。ただし、具体的な時期予測には不確定要素が多いことも示唆されています。

Outlines

00:00

👾 AGIに対する懐疑的な見方とGPTの能力の進化

この段落では、GPTやChatGPTのような最新のAIシステムの能力は、単なる確率モデルや関数近似にすぎず、人間の脳や意識とは本質的に異なるものだという懐疑的な見方を紹介している。しかし同時に、GPT-3やChatGPTの登場により、AIが文章を理解し、学習し、道徳的判断ができるかのように見えるようになったことで、文明に大きな変化をもたらす可能性にも言及している。

05:00

🧠 ニューラルネットワークにおけるパラメータの役割

この段落では、ニューラルネットワークにおけるパラメータ(重み・バイアス)の役割について説明している。パラメータは、ニューロン同士の結合の強さを表し、より多くのパラメータを持つモデルほど、複雑なパターンを学習できる。また、パブロフの犬の実験を例に、条件付けによる学習プロセスとニューラルネットワークのパラメータ更新の類似性を示している。

10:02

🔢 ヒト脳とAIモデルのパラメータ数の比較

この段落では、ヒト脳のシナプス数(パラメータに相当)が1000億個程度であることを示し、AIモデルのパラメータ数とヒト脳のパラメータ数を比較している。GPT-3のパラメータ数は175億個であり、ネコの脳と同程度だが、ヒト脳に匹敵するには1000兆個のパラメータが必要であると指摘している。また、パラメータ数とAIの性能には相関関係があり、ヒト並みの能力を持つAIを実現するためには、ヒト脳と同等のパラメータ数が重要であると述べている。

15:02

📈 AGIを実現するためのパラメータ数の予測

この段落では、AGI(人工般化知能)を実現するために必要なパラメータ数の予測について述べている。あるレポートでは、AGIに必要なパラメータ数の中央値が1000万億個程度と推定されている。また、Quantum Computingの研究者でOpenAIのAI安全性の研究に従事しているScott Aronsonの見解に触れ、今後のAGIの進展に対する期待と懸念が示されている。

Mindmap

Keywords

💡ニューラルネットワーク

ニューラルネットワークとは、人間の脳の神経細胞のように情報を処理するコンピューターシステムのことです。この動画では、ニューラルネットワークの「パラメータ」と呼ばれる数値の重みが、人工知能の能力を決定する重要な要素として説明されています。ニューラルネットワークのパラメータ数が多ければ多いほど、より複雑な情報処理が可能になるといわれています。

💡AGI(人工般化知能)

AGI(人工般化知能)とは、人間と同等の汎用的な知能を持つ人工知能のことを指します。この動画では、AGIを実現するために必要なニューラルネットワークのパラメータ数について議論されています。AGIが実現されれば、リモートワーカーが行うほとんどすべての作業を人工知能が行えるようになると考えられています。

💡パラメータ

ニューラルネットワークにおけるパラメータとは、ニューロン同士の結合の強さを表す数値のことです。人間の脳の中の神経細胞同士の結合が「シナプス」に相当します。動画の中で、人間の脳のシナプスの数が約1,000兆個あると推定されており、AGIを実現するためにはそれに匹敵する規模のパラメータ数が必要だと説明されています。

💡スケーリング則

スケーリング則とは、ニューラルネットワークの規模(パラメータ数)が大きくなれば、その性能も比例して向上するという法則のことです。この動画では、スケーリング則に基づいて、AGIを実現するためのパラメータ数を予測する研究が紹介されています。パラメータ数が人間の脳と同等の規模に達すれば、人間レベルの知能が実現できるという仮説が示されています。

💡GPT-3

GPT-3はOpenAIが開発した、膨大な量の文章データを学習した大規模な言語モデルです。この動画では、GPT-3がパラメータ数約1,750億個と、当時として最大規模のニューラルネットワークであったことが紹介されています。GPT-3の出現により、人工知能が人間並みの文章を生成できるようになり、大きな驚きをもたらしました。

💡量子コンピューター

量子コンピューターとは、量子力学の原理に基づいて動作する次世代のコンピューターです。この動画では、オープンAIのAI安全性研究者であるスコット・アロンソンが、もともと量子コンピューターの研究者であったことが紹介されています。彼の経歴から、量子コンピューターの高速計算能力がAGI実現に役立つ可能性があると示唆されています。

💡データ

ニューラルネットワークを学習させるために必要なのがデータです。この動画では、インターネットを通じて莫大な量のデータが入手可能になったことが、近年のAI能力の飛躍的な向上に寄与していると指摘されています。質の高いデータを多く与えることで、ニューラルネットワークの性能が大幅に向上するためです。

💡バックプロパゲーション

バックプロパゲーションとは、ニューラルネットワークの学習において、出力結果と正解の差を逆算して、ネットワーク内の結合の重み(パラメータ)を修正する手法のことです。動画の中で、「温かいかヒントに基づいて修正」というゲームの比喩が使われていました。バックプロパゲーションによってパラメータを徐々に調整し、望ましい出力が得られるようにチューニングしていくのです。

💡トランスフォーマーモデル

トランスフォーマーモデルは、注意機構(Attention Mechanism)を活用したニューラルネットワークの一種で、言語処理などの自然言語タスクに優れた性能を発揮します。有名な例としてGPT(Generative Pre-trained Transformer)がありますが、この動画でもパラメータ数の増加に伴うトランスフォーマーモデル性能の向上が示唆されています。

💡パブロフの実験

パブロフの実験とは、条件付け学習の代表例として有名な心理学実験です。この動画では、犬に鐘を鳴らすと同時に餌を与えると、次第に鐘の音だけで唾液が出るようになったという実験が紹介されています。これは、ニューラルネットワークのパラメータが、刺激と結果の関係を学習するメカニズムに例えられています。

Highlights

The author discusses the concept of 'deflationary claims' about AI systems, where each person believes they are the first to make such claims, but the author sees these claims as reductionistic and lacking a principle that separates AI from human intelligence.

The author mentions Scott Aaronson, who worked for OpenAI on AI alignment and safety, and his blog post titled 'Letter to His 11-Year-Old Self' where he discusses the creation of an AI that can converse like humans and the ethical implications surrounding it.

The author expresses skepticism about a leaked paper claiming that GPT-4 or another model built in 2022 has 100 trillion parameters, stating that they don't buy it.

The author discusses the concept of scaling laws and when digital neural networks might exceed the complexity of the human brain, defining AGI as the ability to perform any intellectual task that a smart human can.

The author explains the basics of neural networks, including neurons, connections (parameters/synapses), and how they are trained through forward and backward propagation to adjust the weights and biases to produce desired outputs.

The author highlights John Carmack's belief that we've had the technical details of AGI solved for many decades, but lacked the computing power, data, and internet infrastructure to achieve it.

The author cites a figure that the human brain has around 100 trillion to 200 trillion synaptic connections (parameters), which is used as a benchmark for comparing the parameter count of AI models.

The author discusses a paper that predicts AI performance by parameter count, showing that as models approach the parameter count of the human brain, they are expected to reach human-level abilities.

The author mentions a transformative model called 'AA' (possibly a codename for a specific AI model) and its ability to perform tasks that remote human workers can do, as a potential indicator of AGI.

The author presents estimates for the number of parameters required for a 'transformative model' (potentially AGI) to achieve human-level abilities, ranging from the size of GPT-3 to one quintillion (10^18) parameters, which is 10,000 times more than the human brain.

The author expresses uncertainty about how much of the information presented is believable but finds Scott Aaronson's work on the computational complexity of linear optics and his role at OpenAI's AI safety and alignment team particularly interesting.

The author acknowledges that the video might have been a bit disjointed but expresses that the topic of AI and AGI is becoming increasingly interesting, with more to come soon.

Transcripts

play00:00

I'm going to call it the religion of

play00:02

jism okay so so there's like the you

play00:04

know there's this whole sequence of

play00:07

deflationary claims right like each

play00:09

person who makes them thinks that

play00:11

they're like the first one right and

play00:12

they you know there there's like I've

play00:14

seen like like 500 different variants of

play00:17

this now right chat gbt you know it

play00:19

doesn't matter how impressive it looks

play00:21

because it is just a stochastic paret it

play00:24

is just a next token predictor it is

play00:27

just a function approximator it is just

play00:30

a gargantuan autocomplete right and what

play00:32

these people never do what it never

play00:34

occurs to them to do is to ask the next

play00:37

question what are you Justa

play00:39

right right aren't you just the bundle

play00:42

of neurons and synapses right I mean

play00:45

like we could take that deflationary

play00:47

reductionistic stance about you also

play00:50

right or or if not then we have to give

play00:52

some principle that separates the one

play00:54

from the other right you know it is our

play00:56

burden to give that principle so the way

play00:59

that someone was putting it on my blog

play01:01

was okay you know they they gave this

play01:02

giant litany you know look GPT does not

play01:05

interpret sentences it seems to

play01:07

interpret them it does not learn it

play01:10

seems to learn it does not judge moral

play01:13

questions it seems to judge moral

play01:15

questions and so I just responded to

play01:17

this I said you know that's great and it

play01:18

won't change civilization it will seem

play01:20

to change

play01:23

it so the person that was talking his

play01:25

name is Scott arenson and he recently

play01:28

went to work for openi of the AI

play01:30

alignment and safety his previous work

play01:32

in research was into Quantum Computing

play01:35

and so he started working for open AI in

play01:37

2022 probably around the middle of the

play01:39

year and by the end of the year he put

play01:41

out a blog post titled letter to his

play01:44

11-year-old self in it he says this

play01:47

there's a company building an AI that

play01:49

fills giant rooms eats a Town's worth of

play01:52

electricity and has recently gained an

play01:54

astounding ability to converse like

play01:56

people we can write essays or poetry on

play01:59

any topic it can Ace college level exams

play02:02

it's daily gaining new capabilities that

play02:04

the engineers who tend to the AI can't

play02:06

even talk about in public yet those

play02:08

Engineers do however sit in the company

play02:10

cafeteria and debate the meaning of what

play02:13

they're creating what will it learn to

play02:15

do next week which jobs might it render

play02:18

obsolete should they slow down or stop

play02:20

so as not to tickle the Tail of the

play02:23

Dragon but wouldn't that mean someone

play02:25

else probably someone with less Scruples

play02:28

would wake the Dragon first is there an

play02:31

ethical obligation to tell the world

play02:34

more about this is there an obligation

play02:36

to tell it less and he's saying that his

play02:38

job at the company is to develop a

play02:40

mathematical theory of how to prevent

play02:42

the AI and its successors from wreaking

play02:44

havoc so that's Aron right there in 2011

play02:48

this is from this paper that was leaked

play02:50

uh and my take is I I think this is BS

play02:53

after reading it and trying to verify

play02:55

some of it I mean it's I just don't buy

play02:57

it here's the thing it starts out really

play02:59

good it it had me going but at some

play03:01

point it kind of rapidly falls apart and

play03:04

it's trying to push this idea that GPT 4

play03:07

or some other model that they built in

play03:09

2022 has 100 trillion parameters now

play03:13

again I I don't buy it I'll post it down

play03:15

below if you guys want to take a look at

play03:16

it but anyways my take is is a lot of

play03:18

this is is nonsense but in this PDF

play03:21

there are three interesting links to

play03:24

papers or things that other very

play03:26

credible researchers have wrote and

play03:29

specific spefically this guy talking

play03:31

about AI is also really interesting so

play03:34

in this video Let's briefly look at

play03:36

scaling laws and when we can expect

play03:39

digital neural Nets to exceed kind of

play03:41

the complexity of the human brain and

play03:44

basically the definition of a gii that

play03:47

the author kind of states is it can do

play03:49

any intellectual task that a smart human

play03:51

can 2020 was the first time I was

play03:54

shocked by an AI system that was gpt3 so

play03:58

the world was catching up just something

play04:00

that these people were interacting with

play04:02

years before so people were surprised by

play04:04

its ability to reason even as early as

play04:06

gpt3 which gpt3 I feel like most people

play04:09

haven't even interacted with this

play04:11

because Chad GPT the big thing that most

play04:13

people got their hands on that was GPT

play04:15

3.5 kind of an updated version and he's

play04:18

saying that somewhere in there there was

play04:19

this massive leap because before that

play04:21

Chad Bots had no ability to respond

play04:24

coherently at all why was gpt3 such a

play04:27

massive leap and so here we're getting

play04:29

into parameter count so really fast

play04:31

exactly what is a parameter so in neural

play04:34

networks we're kind of replicating the

play04:36

human brain so here's kind of a diagram

play04:38

of a neural network these little round

play04:40

things they're called neurons which is

play04:43

the digital version of the neuron that's

play04:45

in our brain basically these neurons

play04:47

connect to each other and pass

play04:48

information back and forth so let's say

play04:50

there's a neuron in your brain that's

play04:52

responsible for food I'm simplifying

play04:55

obviously but let's say there's another

play04:57

one that's it's the smell of cooking so

play05:00

when you when you smell something

play05:01

delicious cooking on the stove that

play05:03

triggers this neuron now obviously the

play05:06

actual brain is much more complicated

play05:07

there's a whole it's not like one neuron

play05:09

does this or that but just for

play05:11

illustrative purposes like let's say

play05:13

this neuron is food and this neuron is

play05:15

smell of cooking each time you smell

play05:17

something cooking and then you get food

play05:19

the connection between these two neurons

play05:21

gets a little bit stronger over time it

play05:23

gets stronger and stronger and stronger

play05:25

until the smell of cooking gets to be

play05:27

kind of a predictive thing for you

play05:30

getting food whenever you smell cooking

play05:32

you know that there's food around you're

play05:34

going to get food this is kind of how

play05:36

your brain is able to predict the future

play05:38

if you will and this is how brains work

play05:40

in humans now so in dogs if a dog smells

play05:42

something cooking or the smell of food

play05:46

whatever smells trigger food for them

play05:48

you know it might start salivating

play05:49

because it knows food is coming so one

play05:51

day this handsome fella decided to

play05:54

decided to see if he can trick these

play05:56

dogs into creating other neur neural

play05:59

connections that aren't triggered by

play06:03

smells but instead by something kind of

play06:05

random like ringing a bell so this is

play06:08

Ivan Pavlov uh if you've ever hear that

play06:10

term pavlovian response that's kind of

play06:13

his doing he would ring a bell every

play06:16

time before he served dogs food so he'd

play06:19

ring a bell and give him some food and

play06:21

this would go over a course of however

play06:23

along ring a bell give him food ring a

play06:25

bell give the dog food it was a whole

play06:28

thing they really went all in on this

play06:30

now obviously beforehand if you just

play06:31

rang a bell the dog didn't really have

play06:33

any response to it it didn't mean

play06:36

anything to the dog but after doing this

play06:38

for a Time the dogs started salivating

play06:40

after hearing a bell the dogs were

play06:42

conditioned to salivate and expect food

play06:45

whenever that bell rang by the way this

play06:47

is why the office was such a great show

play06:49

cuz that whole prank that Jim plays on

play06:50

Dwight with the breath mints was

play06:53

literally him conditioning that

play06:55

pavlovian response by giving him a

play06:58

breath mint every time there was a the

play07:01

Microsoft Office ding or whatever but

play07:03

the point here is with the dogs and

play07:05

Dwight I guess as this this thing kept

play07:08

happening where a bell would ring and

play07:09

then he and the dog would get a treat

play07:12

the actual like physical wiring in the

play07:13

brain these neural connections would get

play07:16

stronger and stronger so the Bell became

play07:18

a stronger and stronger signal for you

play07:20

know there's food coming until the dog

play07:22

was like okay anytime I hear a bell that

play07:24

means I get food like I was convinced of

play07:26

that so in the neural Nets in the AI

play07:29

weights and biases they determine kind

play07:32

of the strength of that connection how

play07:35

often those connections get called how

play07:36

strong they are so for example before

play07:38

the pavlovian conditioning of the dog

play07:40

you know a bell ringing might have a

play07:42

very low connection to you know getting

play07:45

food the dog doesn't connect a bell to

play07:49

getting food but as he keeps hearing you

play07:52

know being food being food this

play07:54

connection gets stronger to where

play07:56

there's a stronger there's a stronger

play07:58

predictive ability between the being and

play08:01

the food between the Bell and the food

play08:04

and all these various connections are

play08:05

referred to as parameters and so the

play08:09

more parameters the more connections the

play08:12

more possible I guess predictive

play08:14

abilities and so when we refer to the

play08:16

the size of the AI model the size of the

play08:19

llm we refer to it as the number of

play08:21

parameters the number of total

play08:23

connections and then when we train the

play08:25

model when we give it data you can think

play08:27

of all these as little knobs and Dot FES

play08:29

that we kind of twist and turn to try to

play08:32

create these connections that make sense

play08:34

then we then we have our input and the

play08:36

output and we try to understand like how

play08:38

good is this brain this series of

play08:41

connections weights and biases how good

play08:44

is it at producing the response if it's

play08:47

way off then we have a process called

play08:49

back propagation where we go back and

play08:51

kind of like flip these dials into

play08:53

different positions and we try again and

play08:55

these back and forward passes over time

play08:58

set all the little dials and knobs into

play09:01

the correct position to get the outputs

play09:04

that we're looking for so I kind of

play09:06

think of this as that game where you say

play09:07

if you're getting hotter or colder right

play09:09

so you move in a certain direction

play09:10

that's the forward pass and then the

play09:12

person you're playing with goes you're

play09:13

getting warmer and so you do the back

play09:15

propagation so that's where you know

play09:17

maybe you turn in a slightly different

play09:19

direction and you head in that direction

play09:21

right and then the person goes oh you're

play09:22

almost there you're getting hot right so

play09:24

basically the hotter you get the less

play09:27

changes you make to what you're doing if

play09:29

they're saying oh your ice cold then you

play09:31

make a lot of changes and you flip all

play09:32

these dials in in different directions I

play09:35

mean slightly more complicated than that

play09:36

but I feel like what I've described is a

play09:39

pretty good analogy and so the paper

play09:41

continues deep learning is a concept

play09:42

that essentially goes back to the

play09:44

beginning of AI research in the 1950s

play09:46

the first neural network was created in

play09:48

the ' 50s and modern neural networks are

play09:50

just deeper meaning they contain more

play09:52

layers these are the layers so there's

play09:54

just more more layers across the network

play09:57

and most of the major techniques in AI

play09:59

today I rooted in the basic 1950s

play10:01

research combined with a few minor

play10:03

engineering Solutions like back

play10:05

propagation and Transformer models yeah

play10:08

it's not it's just a few minor tweaks I

play10:11

think most people would say that these

play10:12

are kind of I mean big deals but his

play10:14

point is a lot of this the idea of

play10:17

neural networks isn't exactly new so

play10:20

he's saying there's only two reasons for

play10:22

the recent explosion of AI capability

play10:24

size and data uh so maybe a different

play10:27

way of saying that is just we can have

play10:29

massive progress massive improvements we

play10:31

we've just improving the size and the

play10:34

data like this alone will create massive

play10:36

progress without necessarily other

play10:39

breakthroughs I think that's fair to say

play10:41

and a growing number of people in the

play10:42

field are beginning to believe we've had

play10:44

the technical details of AGI solved for

play10:46

many decades we just didn't have the

play10:48

computing power we didn't have the data

play10:50

and we didn't have the internet for all

play10:52

the data so this is John CarMax so he

play10:54

was the guy that created the original

play10:56

Doom him and John Romero and so he's

play10:59

he's kind of a big deal like he's

play11:01

wellknown highly respected here's him

play11:03

with Elon Musk here's him with Notch

play11:06

that's the creator of Minecraft who sold

play11:08

it for billions I think to Microsoft I

play11:11

think he like single-handedly coded

play11:13

Minecraft way back in the day here's him

play11:15

on stage with Steve Jobs I think he

play11:17

worked for meta on the whole virtual

play11:19

reality for for quite some time and on

play11:21

the Lex freedom in podcast he talked

play11:24

about kind of this very idea and he

play11:26

actually founded recently announced that

play11:29

he started his own AGI lab and he's

play11:31

saying like for the first time in kind

play11:32

of humor history just one or a handful

play11:35

of people can have like an incredible

play11:38

result on the world this leverage by

play11:41

creating AGI and he kind of said

play11:44

something similar that we probably have

play11:47

the technical details of AGI like we

play11:50

we've had it solved and I believe he

play11:52

also said that if you had to write write

play11:54

out like all the things that you needed

play11:56

to know to solve AGI it would probably

play11:58

fit on a napkin like there might be 10

play12:00

things that we kind of needed to solve

play12:02

that would allow for egi to happen and a

play12:05

lot of them are probably hidden away in

play12:07

various texts and textbooks over the

play12:10

past you know power many decades so this

play12:13

idea that it's probably not going to

play12:14

come from some brand new thing that no

play12:16

one has expected but rather from

play12:19

something that has been already talked

play12:21

about right like just like neural Nets

play12:23

you know the first one was created in

play12:24

the 50s right so it's been around for a

play12:27

while and so they're saying what is this

play12:28

parameter well it's kind of like a

play12:29

sinapse synapse sinapse however you

play12:32

pronounce that so it's like a syapse in

play12:34

a biological brain connection between

play12:36

neurons and each neuron in the

play12:38

biological brain has roughly a thousand

play12:40

connections to other neurons and of

play12:42

course digital neural networks are

play12:44

analogous to biological brains so this

play12:46

is interesting how many parameters right

play12:49

synapses or parameters are in a human

play12:52

brain so the figure that is commonly

play12:55

cited is 100 trillion so keep that

play12:58

number in M 100 trillion 100 trillion

play13:02

parameters in the human brain so with

play13:04

AGI we're trying to achieve something

play13:06

similar to a human brain or the human

play13:09

brain's capabilities the general

play13:11

intelligence so in nature that's 100

play13:13

quote unquote parameters so Yale

play13:16

Neuroscience 100 trillion synaptic

play13:18

connections human brain there are more

play13:20

neurons in a single brain than there are

play13:22

stars in the Milky Way and a cat has 250

play13:25

billion synapses a dog has 530 billion

play13:28

sinaps says sinapse count generally

play13:30

seems to predict higher intelligence

play13:32

this guy is now just talking crap about

play13:33

cats I'm not sure how I feel about that

play13:36

and he know so yeah there's there's some

play13:37

exceptions for example elephants have

play13:40

higher count than humans yet display

play13:42

lower intelligence and he kind of

play13:43

explains that that the quality of data

play13:46

might answer for those uh exceptions so

play13:49

human brains evolved from higher quality

play13:51

socialization and communication data

play13:53

than elephants but the point is syapse

play13:55

count is definitely important and so

play13:57

gpt2 the syapse count is less than a

play14:01

mouse's brain gpt3 is approaching a

play14:05

cat's brain so it's intuitively obvious

play14:08

that an AI system the size of a cat's

play14:11

brain would be superior to an AI system

play14:13

the size of a mouse's brain so all other

play14:15

things equal certainly that's the case

play14:17

predicting AI performance in 2020 after

play14:20

the release of the 175 billion parameter

play14:24

gpt3 many speculate about the potential

play14:26

performance of a model that is 600 times

play14:28

larger so that's the 100 trillion

play14:30

parameters kind of where it's equivalent

play14:33

to the human brain so gpt3 is like.

play14:39

175% so it's like a tenth of 1% of what

play14:43

the human brain is in terms of

play14:45

parameters so is it possible to predict

play14:47

AI performance by parameter count and as

play14:49

it turns out the answer is yes and so

play14:52

here's the paper so saying there are

play14:53

roughly 2 2 * 10 14th power synapses in

play14:57

the human brain

play14:59

which that's that's 200 trillion so it's

play15:02

double than that Yale quoted earlier

play15:05

Yale NE science so here they're saying

play15:07

200 you know double what double that

play15:09

amount right and this line here looks

play15:12

like it's this line on the chart that's

play15:16

where the parameters equal synapses in

play15:19

the brain so this is kind of where that

play15:21

line when we cross it over that's when

play15:22

neural networks match the parameters in

play15:25

of the human brain according to this

play15:27

article and the dark green line that's

play15:29

the next line here so again this is

play15:31

where it matches the human brain this is

play15:34

the tii the transformative model so this

play15:38

is AA cotra and so in this little speech

play15:41

she gave the introducer said AA so I'm

play15:43

just going to go with that AA so this is

play15:45

a ja and this is from less wrong.com

play15:47

I'll post this in the show notes we

play15:49

probably it's a little bit older at this

play15:51

point August 2022 but we might look into

play15:55

it but she talks about tii which is I

play15:58

mean you can think of it as AGI so

play16:00

basically kind of a similar idea so like

play16:02

at what point is it going to get to the

play16:04

point of maybe replacing human workers

play16:07

uh or or at least being as capable as

play16:08

human workers but but this kind of

play16:10

jumped out of me so she was saying when

play16:12

writing my report I was imagining that a

play16:14

transformative model would likely need

play16:17

to be able to do almost all the tasks

play16:19

that remote human workers can do and

play16:21

again we might do a deep dive into this

play16:23

article but I got to say in my mind I

play16:25

think this is a much better sort of

play16:27

conceptual way thinking about AGI like

play16:30

at what point can it do anything that a

play16:33

remote human worker can do so basically

play16:36

if you have a job where a person doesn't

play16:37

need to come into the office and you

play16:39

communicate through emails and they do

play16:42

whatever whatever it is that they do

play16:43

whether that's Excel or writing or

play16:46

coding design whatever like when will

play16:49

tii or AGI when can AI kind of just do

play16:52

all of that or in other words if you're

play16:54

a remote worker and currently you spend

play16:57

half of the time that you're supposed to

play16:58

supposed to be working playing hell

play16:59

divers 2 at what point can you spend all

play17:03

of your time playing hell divers 2 but

play17:06

the point is this green line that's the

play17:09

median that's the average estimate for

play17:10

the number of parameters in that

play17:13

transformative model that tii and the

play17:16

80% confidence interval so kind of is

play17:18

between these two number of parameters

play17:21

so meaning at what point are we fairly

play17:23

certain that we've achieved that at how

play17:25

many parameters have we achieved AGI

play17:28

well it's as low as gpt3 and uh and as

play17:31

high as so 10 to the 18th that's one

play17:35

quintilian so if I'm doing my math right

play17:38

so that's 10,000 times more than what

play17:41

the human brain would be so this

play17:43

extrapolation shows that AI performance

play17:46

will reach human level abilities as it

play17:49

reaches human level size parameter count

play17:52

so what we just went over I don't know

play17:54

how much of it I believe I went through

play17:57

it I ended up cutting most of it but

play17:59

this Scott Aronson guy really jumped out

play18:02

to me here's a paper that he did the

play18:04

computational complexity of linear

play18:06

Optics talking about giving new evidence

play18:09

that quantum computers cannot be

play18:12

efficiently simulated by classical

play18:13

computers it's interesting that a

play18:15

quantum computer guy is working at open

play18:17

ai's you know AI safety and Alignment

play18:20

there's some more interesting stuff

play18:22

ahead I apologize if today's video was a

play18:24

little bit disjointed but this whole

play18:25

thing is getting uh a lot more

play18:27

interesting more to come very soon

Rate This

5.0 / 5 (0 votes)

Do you need a summary in English?