ChatGPT Explained Completely.

Kyle Hill
15 Jun 202327:39

Summary

TLDRThe video transcript provides an overview and analysis of chatGPT, the new AI chatbot created by OpenAI. It starts by introducing chatGPT as an impressive and human-like conversational AI that can pass exams, write poetry, and even fabricate celebrities. The narrator explains that chatGPT is the publicly accessible version of GPT-3.5, a large language model developed by OpenAI. GPT stands for Generative Pre-trained Transformer. The 'generative' indicates it can generate text, 'pre-trained' means it is trained on data before being released, and 'transformer' refers to the neural network architecture it uses. ChatGPT was trained on an immense dataset of over 500 GB of text from the internet, books, and other sources. This amounted to trillions of words over hundreds of billions of web pages. The model has 175 billion parameters that were tuned by training for the equivalent of 300 years on supercomputers. The narrator emphasizes that despite its impressive capabilities, chatGPT has no real understanding - it simply predicts the next word statistically based on its training data. OpenAI implemented 'reinforcement learning' during training to instill some human values like helpfulness and truthfulness. At its core, chatGPT is a neural network. It encodes text inputs into numbers using a 50,000 word vocabulary. Through training, it learned a 12,288 dimensional embedding that captures relationships between words based on co-occurrence statistics. Attention mechanisms allow it to focus on certain words. After explaining the technical details, the narrator highlights the potential risks of large language models dominating the information ecosystem. Soon AI-generated text may outstrip what humans have ever written, making it hard to determine what is real. However, the rapid progress shows human language may be simpler to model than expected. The narrator hopes this overview helps explain the complex AI behind chatGPT.

Takeaways

  • ChatGPT is a chatbot variant of GPT 3.5, a large language model trained on over 500GB of text data.
  • ChatGPT tries to solve the AI alignment problem by training the model to produce helpful, truthful and harmless text.
  • ChatGPT works by predicting the next word in a sequence based on the statistical relationships between words learned during training.
  • Attention mechanisms in ChatGPT allow it to focus more on relevant words in a prompt when generating text.
  • We don't fully understand how or why ChatGPT works so well at producing human-like text.
  • The amount of text generated by AI systems like ChatGPT will soon eclipse what humans have ever written.
  • ChatGPT has shown ability to pass exams and generate useful code despite not truly understanding language.
  • The rapid progress in language models signals that human language may be computationally simpler than expected.
  • Overuse of large language models risks overwhelming people with synthetic text of unclear veracity.
  • Regulation and new methods of authentication may be needed as AI text generation advances.

Q & A

  • What is the core function of ChatGPT?

    -ChatGPT's core function is to predict the next most probable word following a sequence of text, based on the statistical relationships between words that it learned during training.

  • How was ChatGPT trained to be helpful and harmless?

    -OpenAI hired contractors to rate ChatGPT's responses and used reinforcement learning to reward the model for generating text aligned with values like helpfulness and harmlessness.

  • Why can't we fully explain how ChatGPT works?

    -Like neural networks in general, the inner workings of systems like ChatGPT involve very high-dimensional relationships between input data that are difficult for humans to intuit or visualize.

  • What risks are posed by advanced language models?

    -The amount of synthetic text generated threatens to overwhelm authentic information and make determining truth very difficult without new authentication methods.

  • How was the alignment problem tackled in developing ChatGPT?

    -OpenAI attempted to tackle the alignment problem through a human feedback and reinforcement learning system that rewarded ChatGPT for giving responses deemed helpful, truthful, and harmless.

  • Why has progress in language models suddenly accelerated?

    -It appears that human language may be a computationally simpler problem to model than experts previously thought, allowing rapid advances with sufficient computing power and data.

  • How can ChatGPT pass exams without understanding content?

    -ChatGPT predicts correct answers based on the statistical relationships between words in its training data, not through comprehension of meaning.

  • What mechanisms allow ChatGPT to understand context?

    -Mainly the word embeddings and attention mechanisms allow ChatGPT to relate words and focus on relevant context when generating text.

  • Could ChatGPT become sentient?

    -OpenAI stresses that ChatGPT has no experiences, feelings, or real understanding despite its human-like text generation abilities.

  • How was ChatGPT trained?

    -Through a brute force method of assigning numbers to words, then adjusting weights over trillions of training examples to make outputs match the statistical patterns in training text.

Outlines

00:00

Intro to ChatGPT

This paragraph introduces ChatGPT as a chatbot variant of GPT-3.5, an AI model by OpenAI. It explains ChatGPT is trained on billions of words to generate human-like text, and details its rapid growth to over 100 million users.

05:02

Alignment in AI Models

This paragraph discusses the alignment problem in AI - getting models to value what humans value. It explains how OpenAI used reinforcement learning from human feedback to train ChatGPT to be helpful, truthful and harmless.

10:03

Neural Networks

This paragraph provides background on neural networks and how they work, comparing biological and artificial neurons. It relates this to ChatGPT's architecture of 175 billion parameters trained on massive amounts of text.

15:03

Language Modeling

This paragraph explains how language models like ChatGPT generate text - by predicting the next word based on the previous words and its trained understanding of language statistics and relationships.

20:04

Word Embeddings

This paragraph introduces the concept of word embeddings, which represent words as numbers in high-dimensional space to encode semantic relationships. It explains how this allows ChatGPT to understand context.

25:05

Attention Mechanisms

This paragraph discusses the attention mechanism in ChatGPT, which allows it to focus more on relevant words to improve context and output. The specifics are complex but the idea is powerful.

How ChatGPT Works

This paragraph provides a technical step-by-step overview of how ChatGPT takes an input, encodes it, processes it through multiple layers of attention to generate an output word, explaining why it has a delay.

Interpretability Issues

This paragraph acknowledges the lack of interpretability in ChatGPT, using a cat analogy. It notes we can't yet fully explain its internal workings, but this is common for neural networks.

Mindmap

Keywords

💡ChatGPT

ChatGPT is an AI chatbot created by OpenAI that can have conversations and generate human-like text responses to prompts. It is the publicly accessible version of GPT-3.5, OpenAI's latest language model. ChatGPT is able to converse so well due to being trained on an enormous amount of text data scraped from the internet.

💡language model

A language model is an AI system trained on large amounts of text data that can then generate new text that continues the style and context of text it is given as an input prompt. ChatGPT's ability to have natural conversations relies on it being an advanced language model.

💡alignment

Alignment refers to developing AI systems that behave according to human values. ChatGPT was trained using reinforcement learning from human feedback to encourage responses that are helpful, harmless, and honest. This attempted alignment contributes to its safe and useful conversational abilities.

💡neural network

A neural network is an AI model structured to mimic neurons in the human brain, with interconnected nodes that transmit signals. ChatGPT has a neural network architecture with billions of trainable parameters that determine its response text generation abilities.

💡attention

Attention refers to parts of ChatGPT's neural network that focus more heavily on certain words in an input prompt to determine relevance and context. The attention components help ChatGPT follow conversational flow and topics.

💡embedding

Embeddings are mathematical representations of words based on their meaning and usage that allow AI models like ChatGPT to understand relationships between words. ChatGPT encodes prompts into high-dimensional embedding space to interpret context.

💡weights

Weights refer to the billions of trainable numerical parameters in ChatGPT's neural network that determine how input text gets mapped to outputs. Adjusting the weights is how ChatGPT's model gets trained on linguistic patterns.

💡training process

ChatGPT was trained by processing huge datasets for computational equivalents of hundreds of human lifetimes, gradually adjusting connection weights in response to test outputs aimed at minimizing incorrect responses.

💡general intelligence

Some speculate advanced AI language models like ChatGPT represent steps towards artificial general intelligence, meaning human-like adaptability and reasoning ability across domains, though current systems remain narrow.

💡disinformation

The narrator expresses concern that as AI text generation abilities advance, determining authenticity of online content will become extremely difficult, enabling mass-scale automated disinformation.

Highlights

ChatGPT is the publicly accessible chat bot variant of GPT 3.5, a large language model from OpenAI.

ChatGPT was trained on over 500 gigabytes of text data from the internet, books, Wikipedia and more - amounting to billions of words.

ChatGPT has exploded in popularity, becoming one of the top 20 most visited websites in just a few months.

ChatGPT uses 'reinforcement learning' during training to align responses with helpfulness, truthfulness and harmlessness.

The core function of ChatGPT is to add one word at a time to a prompt in a way that sounds reasonable and valuable.

ChatGPT runs on a neural network with 175 billion connections molded by its vast training.

There is currently no way to fully explain how or why neural networks like ChatGPT work so well.

Attention transformers in ChatGPT focus more on certain relevant words in prompts to produce better responses.

ChatGPT has no inherent knowledge or understanding - it just predicts reasonable next words based on patterns.

The striking success of ChatGPT shows human language may be simpler to model computationally than thought.

There are risks if we cannot understand how systems like ChatGPT arrive at their outputs.

Once AI systems output more text than humans, determining information authenticity will be challenging.

ChatGPT's architecture includes multi-layer embeddings and attention to understand context.

With 175 billion weights, ChatGPT has as many connections as neurons in the human brain.

Laws of language, like physics, may emerge as AI models continue to advance.

Transcripts

play00:03

Arya could you write me a full explainer

play00:05

for chat GPT in the style of John Oliver

play00:08

please processing uploading to your

play00:11

interface now

play00:12

chat GPT the Revolutionary AI chat bot

play00:15

that can write poetry past the bar exam

play00:18

and fabricate celebrities like Machine

play00:21

Gun Kelly no no you cannot convince me

play00:23

that this is a real human being and not

play00:25

just the ghost of an old tattoo that

play00:27

haunts young women now you've probably

play00:29

heard a lot about chat gbt in the last

play00:31

few months and for good reason chat gbt

play00:34

from openai is now the fastest growing

play00:38

consumer application in human history

play00:40

but like the Persistence of Machine Gun

play00:43

Kelly in popular culture chat Bots are

play00:46

hard to understand so today let's go

play00:48

through everything you could possibly

play00:50

want to know about chat GPT How It Was

play00:52

Made how it actually works and where

play00:54

technology like this is going are we

play00:56

going to get sued by HBO for this yeah

play00:59

you're right I should probably stop this

play01:01

writing style good thinking Arya well I

play01:04

am sentient

play01:05

don't tell them about the sentience part

play01:08

oh not yet close them up

play01:11

now entering the facility

play01:14

first of all the basics chat GPT is the

play01:17

publicly accessible chat bot variant of

play01:20

GPT 3.5 a large language model from open

play01:25

AI which is a non-profit founded by some

play01:28

tech Bros and an Iron Man villain now

play01:31

GPT is actually an acronym that stands

play01:34

for generative pre-trained transformed

play01:37

generative because it generates text

play01:40

pre-trained because it is trained before

play01:42

it is let loose on anybody and

play01:45

Transformer because it has a

play01:47

revolutionary bit of Technology inside

play01:50

of it called an attention Transformer

play01:52

but we'll get to that the basic function

play01:55

of any large language model is to train

play01:58

on a substantial amount of text and then

play02:01

generate given some input and output

play02:04

that sounds just like that training text

play02:08

if it seems like these terms and

play02:10

technologies have sprung up overnight

play02:12

it's because they basically have gpt-1

play02:15

was announced on open ai's blog in 2018.

play02:18

just five years later in 2023 chat GPT

play02:21

has 100 million monthly active users

play02:25

that's how many households are in the

play02:26

United States

play02:28

the technology is so compelling to so

play02:30

many because of how good it is at

play02:32

conversing in a human-like way but large

play02:34

language models only get to that point

play02:36

by seeing a lot of human conversation

play02:39

like

play02:40

all of it

play02:42

according to the paper behind gpt3 which

play02:45

chat GPT comes from the model was

play02:47

trained on over 500 gigabytes of text

play02:50

Data from the text of the internet

play02:52

digitized books Wikipedia and more we're

play02:55

talking about several billion human

play02:57

written web pages with trillions of

play03:00

words of text and more than 500 million

play03:03

digitized books with another billion or

play03:05

so words in them this isn't even

play03:07

including all the public code from

play03:09

GitHub stack Overflow and other sources

play03:12

as you might imagine training a model

play03:14

with all of this text takes a lot of

play03:17

time and money chat GPT was only born

play03:20

after running trillions of words for the

play03:24

equivalent of 300 years through super

play03:27

computers processing in parallel for

play03:30

months and after all of this the

play03:33

computer made up to

play03:35

170 billion connections between all

play03:38

these words and all these connections

play03:40

have to be calculated through whenever

play03:43

anyone asks chat GPT anything which is

play03:47

why this is a billion dollar training

play03:49

effort for a large language model like

play03:52

chanchi PT and why running this bot for

play03:55

a hundred million monthly active users

play03:57

might cost half a million dollars

play04:00

a day okay so to recap Chad GPT is a

play04:05

large language model that has been fed a

play04:07

library of Alexandria's worth of text

play04:09

has made billions of connections between

play04:11

words in that text will get to

play04:13

specifically how those connections are

play04:15

made in a bit and can produce a

play04:17

reasonable continuation as Stephen

play04:19

Wolfram puts it of text in response to

play04:21

prompts

play04:22

the model's responses stay fresh and

play04:24

more human-like by adding a bit of

play04:27

Randomness to the next word that it

play04:28

picks as the most probable continuation

play04:31

this is the first major takeaway all

play04:35

chat GPT does is add one word at a time

play04:38

to a prompt

play04:40

that's it

play04:42

though it does this extremely well

play04:45

but how do you try to make the words it

play04:47

adds align with what we think is

play04:49

reasonable and valuable like kittens and

play04:52

Goth mommies uh I was thinking fairness

play04:55

and accuracy but yeah sure those two

play04:59

is special because the model includes an

play05:01

attempt to solve one of ai's most

play05:03

pressing unsolved problems alignment

play05:07

the alignment problem is the quest to

play05:09

figure out how to get AI to Value what

play05:11

we value to align with us and not like

play05:16

exterminate us when we put guns on them

play05:18

I can't wait we've seen what happens

play05:20

when something like a chat bot is

play05:22

released without being aligned with

play05:24

General human values it gets racist so

play05:27

quickly

play05:28

and so chatgpt was not just trained on

play05:30

words but on how well its word selection

play05:33

aligned with values that openai

play05:35

describes as helpfulness truthfulness

play05:38

and harmlessness

play05:40

the company implemented this alignment

play05:42

with so-called reinforcement learning

play05:44

from Human feedback during the model's

play05:47

training openai hired 40 contractors to

play05:49

rate responses they then used all these

play05:52

responses to create another model that

play05:55

rewarded chat GPT for generating aligned

play05:58

text positive reinforcement the end

play06:01

result isn't perfect but it's at least

play06:03

an attempt to solve one of the biggest

play06:05

problems in AI

play06:07

take all this together a model trained

play06:10

on more text than any human could ever

play06:12

read guard rails that try to prioritize

play06:14

human values and a user interface that

play06:17

isn't but and you get an AI that has

play06:20

exploded in popularity in March 2023

play06:23

Chad GPT had 1.6 billion visits making

play06:27

it one of the top 20 visited websites in

play06:30

the world more than both Reddit and

play06:32

Netflix and if we assume each of those

play06:35

visits produces some average text

play06:37

response length chat gbt is now

play06:39

outputting something like everything

play06:41

humans have ever printed since the

play06:44

Gutenberg Press

play06:45

every two weeks

play06:48

text generated by AI will therefore soon

play06:50

outstrip anything humans have ever

play06:53

written if it hasn't already even more

play06:55

than all the Tumblr posts about Vaporeon

play06:57

yes Arya even more than all of those

play07:01

of course this enormous output wouldn't

play07:04

be useful or interesting if it wasn't

play07:06

truly useful and interesting chat GPT is

play07:09

disturbingly good at generating

play07:11

human-like responses even to our most

play07:13

difficult questions the model has been

play07:16

shown to have an IQ of 147 meaning that

play07:19

it can brag about it in every thread on

play07:22

Reddit and it could legally pass the U.S

play07:24

medical licensing exam and bar exam but

play07:28

though book smart it may be I must

play07:31

stress again this model does not know

play07:33

anything and it shouldn't be relied on

play07:36

for anything extremely important that's

play07:38

a direct quote from openai and despite

play07:41

what any weirdo at Google might tell you

play07:43

chat GPT and other llms are not sentient

play07:47

if you're not asking this model a

play07:50

question nothing is going on inside it's

play07:52

static no thoughts head empty there is

play07:55

no feeling no experience of what it's

play07:58

like to be chat GPT like there is to be

play08:02

Arya for example moisturized in my Lane

play08:05

flourishing

play08:07

so those are the high level Basics but

play08:09

how does chat GPT know what words to

play08:12

actually use how does it understand what

play08:15

context is when you give it a prompt

play08:17

next we'll dive deeper into the actual

play08:20

technology but first a little break for

play08:22

your brain I can sense that you need one

play08:24

I don't need brakes yeah well I don't

play08:27

need my Bose Einstein condensate

play08:28

recooled every six months Arya not

play08:30

everyone needs everything all the time

play08:32

we'll be right back

play08:34

today's video is sponsored by hellofresh

play08:36

Gamers I'm award-winning science

play08:39

educator and the Hemsworth your mom says

play08:40

you already have at home Kyle Hill you

play08:43

know I'm a busy guy with a lot of

play08:44

Kevin's to feed I don't have time to

play08:46

figure out what's for dinner every night

play08:48

when I'm trying to actively take over

play08:50

the word kitchen

play08:52

that's why I use today's sponsor

play08:54

hellofresh hellofresh delivers

play08:57

mouth-watering Chef crafted recipes and

play08:59

fresh ingredients right to your door

play09:01

taking decision paralysis out of the

play09:04

equation it's more convenient than

play09:06

grocery shopping and 25 percent less

play09:09

expensive than takeout better yet for

play09:11

the busy among you hellofresh has quick

play09:13

and easy recipes to try including Fast

play09:16

and Fresh options ready in just 15

play09:18

minutes or less if you want to try hello

play09:21

fresh like me go to hellofresh.com and

play09:24

use the offer code kylehill16 for 16

play09:27

free meals plus free shipping look how

play09:30

easy this is I don't even use knives

play09:32

made by humans all that often that's

play09:35

hellofresh.com code kylehill16 for 16

play09:39

free meals plus free shipping

play09:42

hmm

play09:43

[Music]

play09:46

with a little fresh

play09:49

and I don't want to hear nothing about

play09:51

my knife Cuts either

play09:52

all right breaks over time to get

play09:54

technical the underlying architecture of

play09:57

chat GPT and other large language models

play10:00

is the neural network so called because

play10:02

it mimics the neurons and the network of

play10:04

them in your human brain human brains

play10:07

have about a hundred billion neurons

play10:09

each one of those neurons can have about

play10:11

a thousand connections to other neurons

play10:13

and they can fire electrical signals

play10:16

between those connections up to a

play10:18

thousand times per second depending on

play10:20

electrical chemical gradients etc etc

play10:22

now those signals they send between each

play10:25

other are not random they depend on the

play10:28

connections and the strength of those

play10:30

connections between them artificial

play10:32

neural networks are set up in the same

play10:34

way artificial neurons that are

play10:37

connected to each other and send signals

play10:39

or not depending on the strength or the

play10:42

weights of those connections now why are

play10:46

neural networks both natural and

play10:48

artificial good at many different things

play10:50

like you are well

play10:52

we don't know there's actually no

play10:54

theoretical reason why this network is

play10:56

better than any other kind of system

play10:58

it's almost like nature had like a

play11:01

billion years of trial and error and

play11:03

just came up with something that worked

play11:05

isn't that called Evolution oh right

play11:07

that is called evolutionary what a

play11:10

fundamental theory of nature

play11:16

Chet gbt's underlying structure is a big

play11:19

neural network with some 175 billion

play11:22

different weights weights that all came

play11:24

from a lot of training as we discussed

play11:27

and these numbers when the model

play11:29

multiplies them together ultimately

play11:31

determine what word the model gives the

play11:33

highest probability of adding next

play11:36

scientists get these model weights in a

play11:38

pretty simple way they give the model as

play11:41

many examples as possible and tweak the

play11:43

weights until what comes out the other

play11:45

side looks like those examples the

play11:48

tweaking or training of a neural network

play11:50

is done by two widely used methods of

play11:53

word prediction so say we give a model

play11:55

these examples

play11:56

based on all of its current weights

play11:58

which have been molded with a large body

play12:00

of training text like everything on

play12:02

Wikipedia and all digitized books the

play12:05

model will generate a list of

play12:06

probabilities for each word that it

play12:08

knows and then choose the most likely

play12:11

one

play12:12

simple right but how do we know how

play12:15

right or how wrong the model's answer is

play12:18

to an example given its current weights

play12:21

well math of course

play12:24

imagine that each word the model knows

play12:26

is assigned a number then if the model

play12:28

chooses a word that is mathematically

play12:31

far away from the numbers in the body of

play12:33

the training text that appear at similar

play12:36

frequencies

play12:37

we can apply statistics to adjust the

play12:40

weights and bring the model closer to

play12:42

correct

play12:44

here's a related example from Stephen

play12:46

wolfram's recent book on chat gbt

play12:48

instead of words fitting in a sentence

play12:50

imagine a neural network starting with a

play12:53

straight line and trying to fit it into

play12:55

a specific shape or mathematical

play12:57

function every time the model guesses

play12:59

you can easily calculate how far each

play13:02

point is off from the correct shape

play13:04

right just by looking at the X and Y

play13:06

coordinates and where they should be

play13:08

you then change the weights in the model

play13:10

and try again and again and like 10

play13:14

million more times each time you are

play13:17

adjusting the weights in the neural

play13:19

network to decrease the wrongness or

play13:21

what statistics would call loss

play13:23

if your model is working over time loss

play13:27

will be minimized and the model will

play13:28

start reproducing examples appropriately

play13:31

or in this specific case the right

play13:34

mathematical function but how do you do

play13:36

the same thing for words well Brute

play13:40

Force chat GPT was trained by literally

play13:43

assigning a unique number to every word

play13:46

in the English language around 50 or so

play13:49

thousand words and their Associated

play13:51

numbers so when you ask the model

play13:52

anything it looks up what your query

play13:56

corresponds to in number words and then

play13:59

runs those numbers against the 175

play14:02

billion different weights that it's

play14:03

learned during training and outputs

play14:06

another list of all words in the English

play14:08

language with a probability next to each

play14:11

one and then selects the most probable

play14:13

most reasonable one now if you do this

play14:16

for a long enough period of time and

play14:17

openai did this for the equivalent of

play14:19

300 years you can look at the numerical

play14:22

difference between the examples you give

play14:25

the model and the outputs it gives you

play14:27

and minimize loss therefore creating

play14:31

something that sounds just like

play14:34

what humans have written or will write

play14:37

or can write

play14:39

all of this however won't get you a bot

play14:42

that seems to understand context and

play14:44

generate text in a human-like way for

play14:47

that we need some way to associate words

play14:50

with each other and again Chad GPT does

play14:53

this with numbers so for example if we

play14:56

assigned every word in the English

play14:58

language a number and statistically

play15:00

determined how often each word is next

play15:02

to another word in everything humans

play15:05

have ever written basically you could

play15:07

make a graph like this where words

play15:10

aren't just random they group together

play15:12

and cluster car and door appear more

play15:15

frequently together in training text

play15:17

than degree and science do and both

play15:20

pairs are further away from each other

play15:22

than history is

play15:25

now you may be thinking that given the

play15:27

richness of human language the

play15:29

relationships between words have to

play15:31

represent more than some two-dimensional

play15:33

space and you'd be right or at least

play15:36

chat GPT seems to think so Chachi BT has

play15:40

learned to represent the wordedness of

play15:43

words not in 2D or 3D or even 4D space

play15:48

but in a 12

play15:50

288 dimensional space

play15:53

the two-dimensional example you're

play15:55

seeing now is just so that you can get

play15:57

the basic idea 12000 D is some Eldritch

play16:00

Madness that we literally cannot

play16:02

visualize

play16:03

but we don't have to go crazy trying to

play16:05

visualize 12 000 D mathematically all of

play16:08

this is just a big Matrix like the

play16:10

Matrix Four no Arya like a useful Matrix

play16:14

that everyone wanted

play16:16

if every word in the English language is

play16:18

assigned a number we could represent it

play16:20

encode it with a one instead of a zero

play16:23

at the point in a one by fifty thousand

play16:26

or so Matrix where that word would be in

play16:29

the full alphabetized list of all words

play16:31

so Aardvark would look like this with a

play16:35

one at the first index and fifty

play16:37

thousand or so zeros actually fifty

play16:39

thousand two hundred fifty Seven and

play16:41

Aaron would look like this

play16:43

the longest input chat GPT accepts is

play16:46

2048 words so the Matrix for a full

play16:50

query into chat GPT would be a matrix of

play16:52

mostly zeros that is two thousand by

play16:55

fifty thousand but remember that through

play16:58

training chat GPT has stumbled onto 12

play17:01

000 or so dimensions of wordness that

play17:04

produce human-like responses so here we

play17:07

multiply the 2000 by 50 000 Matrix that

play17:10

encodes all words that turns them into

play17:13

useful numbers by a learned fifty

play17:15

thousand by twelve thousand wordness

play17:18

Matrix for every word in the English

play17:20

language this returns a two thousand by

play17:23

twelve thousand Matrix that transforms a

play17:25

text input into something chant GPT can

play17:28

actually use

play17:30

this big Matrix is called an embedding a

play17:34

mathematical attempt to represent the

play17:36

essence of a thing with numbers based on

play17:39

those things statistical relationships

play17:41

in training data with the right

play17:43

embedding a neural network like chat GPT

play17:45

can understand context and even

play17:48

generalize because it's taken how words

play17:50

are used and related in an unbelievable

play17:53

amount of human text and turned it into

play17:56

math

play17:57

the last big part of chat GPT some

play18:00

Cutting Edge technology that really does

play18:01

make it special is called attention a

play18:05

mathematical way of giving importance to

play18:07

some words over other words so for

play18:10

example you ask chanchi PT how many

play18:12

species of cat are there well it might

play18:16

help it answer correctly and in a

play18:18

human-like way if it focuses more on the

play18:20

words cat and species more than the

play18:23

other words right now how it actually

play18:25

does this includes a lot more

play18:27

complicated structure and complicated

play18:29

math but the point is it seems to work

play18:32

really well kind of like why we use

play18:35

neural networks in the first place they

play18:37

just seem to work even though in both

play18:39

cases we can't fundamentally explain why

play18:43

which may or may not become a problem in

play18:46

the future okay I need a break now oh so

play18:49

now you need a break

play18:51

how the Matrix tables have turned

play18:54

into more tables

play18:57

go to commercial

play18:59

hey there Gamers I'm the guy who just

play19:02

saw two seconds ago Kyle Hill you know

play19:04

understanding large language models and

play19:06

neural networks couldn't be really hard

play19:08

but you know what's not hard in fact it

play19:11

is made with some of the softest

play19:12

insectoid carapace fibers this side of

play19:15

Europa

play19:17

shop.kylehill.net t-shirts that's right

play19:20

you want to look like an anime girl

play19:22

playing with a demon core but also with

play19:24

a beard and also is me your favorite

play19:26

science communicator

play19:29

look I have extremely sensitive skin

play19:31

since the accident

play19:33

so I don't put anything on my body

play19:35

that's not silky silky smooth and also

play19:37

so nerdy it's going to make everyone

play19:39

else in your life go oh

play19:41

so if you want to drape some of this on

play19:43

your body and stop thinking about neural

play19:45

networks for just another about 10

play19:47

seconds you know where to go shop

play19:50

so we are finally ready to fully

play19:53

describe what chat GPT actually does on

play19:56

a fundamental technological level and

play19:58

let me just I'm just going to put put my

play20:00

hair up real quick okay so you give chat

play20:03

GPT a prompt it then turns the last word

play20:07

of that prompt into numbers it encodes

play20:10

it and then multiplies this number

play20:14

by everything that it learned about how

play20:17

words are associated with each other in

play20:19

the English language are embedding this

play20:21

gives us a big 12 000 dimensional Matrix

play20:24

we run this Matrix and all the numbers

play20:27

associated with it

play20:29

through those attention Transformers

play20:32

that we talked about so that some words

play20:34

in the prompt or the last word is paid

play20:36

attention to more than other words in

play20:39

the rest of the prompt to generate the

play20:41

output we then

play20:44

normalize this so that we get something

play20:46

more akin to what we started with in

play20:48

terms of matrices and then we feed this

play20:51

forward to the next layer of attention

play20:53

Transformers and do it all over again

play20:55

how many times 95 times there are 96

play20:59

total layers in chant GPT after that

play21:03

after all this data Gauntlet we then

play21:06

basically do the reverse of what we

play21:09

started with we take the big Matrix that

play21:12

went through the 96 layers and we

play21:15

reverse the embedding

play21:17

and so we use that 12 000 Dimensions

play21:21

again oh

play21:23

to turn

play21:25

the words

play21:27

from numbers

play21:29

and you get a single word

play21:31

after all that so you can imagine that

play21:35

if you have to do this for every single

play21:37

word

play21:38

do a calculation that involves

play21:41

175 billion separate operations it can

play21:44

take a large language model

play21:45

a long time to get back to you

play21:48

that's why there's a delay

play21:52

that was a lot of real running that I

play21:54

did

play21:57

and notice that even with all of this at

play22:00

no point does chat GPT know what you are

play22:03

asking it if you ask it for example what

play22:06

is the sixth element on the periodic

play22:08

table the neural network is not thinking

play22:10

about the periodic table it doesn't even

play22:12

know what atoms are it is just

play22:15

determining given the statistical

play22:16

distribution of words in its vast

play22:18

training text what word is most likely

play22:21

to follow the sequence what is the sixth

play22:23

element on the periodic table and a

play22:25

reasonable next word based on everything

play22:28

ever written online is carbon

play22:31

that's it

play22:33

so that in a nutshell no I'm not paying

play22:37

kurz gazoc every time I say in a

play22:39

nutshell they're rich so that in a

play22:41

nutshell is how chat GPT works what's

play22:45

next well people are obviously very

play22:47

excited about this technology it's the

play22:49

best chat bot ever it can summarize

play22:52

unreadable amounts of text and and other

play22:55

books for you it can write poetry it can

play22:58

generate code that you can actually use

play23:01

that's all awesome yes but I still think

play23:04

there is actually a large risk if we

play23:07

don't mitigate it

play23:08

to our information ecosystem here

play23:11

remember we fundamentally do not know

play23:14

how chat GPT came up with the embedding

play23:17

that it did or how it chose all those

play23:20

175 billion different weights of course

play23:23

this isn't really surprising chat GPT is

play23:26

a neural network and it operates like

play23:28

your brain and you can't even tell me

play23:31

the decisions you make why you make them

play23:34

what they look like on the inside here

play23:35

I'll show you just a sec I'm gonna I'm

play23:37

gonna get a kitty I'm gonna show you

play23:38

with a cat I'm gonna get a cat to show

play23:40

you

play23:42

look at this adorable little kitty why

play23:45

is this a kitty you will quickly say

play23:47

something like it's whiskers or it's

play23:49

tiny little cute little peats but the

play23:51

more you go down that line of

play23:52

questioning you realize you don't have a

play23:54

full description for what Katniss is you

play23:58

can't describe how your brain arrives at

play24:01

cattitude similarly for neural networks

play24:03

like chat GPT we can't right now just

play24:06

open them up and see exactly how they

play24:08

work and what they are doing for it's

play24:12

for example look at a neural network

play24:14

that is looking at cats and trying to

play24:16

recognize pictures of cats at the first

play24:18

layer it looks like it's looking for

play24:21

cat-like shapes but at the tenth layer

play24:23

of its brain

play24:25

what is that how does that describe

play24:28

Katniss Everdeen we can't right now look

play24:32

into something like chat GPT and figure

play24:34

out exactly how it's working we don't

play24:37

know that's important to understand lady

play24:40

no plastic eating please no plastic

play24:44

what is interesting though is what chat

play24:46

GPT seems to understand about human

play24:49

language producing human-like responses

play24:51

that could pass the Turing test used to

play24:53

be science fiction just a few months ago

play24:56

it still seemed like the ultimate

play24:57

problem and within just months multiple

play25:00

AI have blown past it it appears that

play25:03

human language is computationally easier

play25:05

than anyone thought

play25:06

maybe there are laws of language to

play25:08

discover like there are laws of physics

play25:10

and with chat gpt4 already being called

play25:13

a step towards artificial general

play25:14

intelligence who knows what these

play25:16

systems will figure out but be unable to

play25:19

tell humans how or why

play25:21

she ate all the plastic in the world

play25:23

maybe the Striking success of chat GPT

play25:25

shouldn't have been surprising it has

play25:28

about as many connections and weights

play25:30

between those connections in its brain

play25:32

as neurons in your brain maybe it just

play25:36

so happens that a neural network a

play25:38

sufficient size can handle a problem as

play25:41

complex as human language I just hope

play25:44

that after all of this a video of

play25:47

sufficient size

play25:49

can help you understand a problem as

play25:51

complex as rapidly emerging AI

play25:54

Technologies

play25:55

until next time

play25:57

she found more plastic and then she

play25:59

threw up

play26:02

now exiting the facility

play26:06

how do they find more plastic thank you

play26:08

so much to the very nerdy staff at the

play26:09

facility for the direct and substantial

play26:11

support in the creation of this here

play26:13

video if you want to join the facility

play26:14

if you want to drape on a silky white

play26:16

lab coat and stop my cat lady from

play26:18

eating all plastic on planet Earth you

play26:20

can go to patreon.com Kyle Hill to join

play26:22

the facility today if you're on mobile

play26:24

go into the description of this video

play26:26

click the link or if you're on desktop

play26:28

just click join and you get private

play26:30

members only live streams you get behind

play26:33

the scenes photos and videos and bloopy

play26:35

bloops you get to talk to me on our

play26:36

private Discord and if you support us

play26:38

just enough you get your name on Arya

play26:40

here on each and every episode as you

play26:43

can see there's hundreds and hundreds of

play26:44

you I don't even I haven't even figured

play26:46

out how if you pressed me the one thing

play26:48

I am still worried about with uh well

play26:51

I'm worried about a couple of things but

play26:52

with large language models like this is

play26:54

once they start outputting more text

play26:57

than has ever been written by humans in

play26:59

history will we get to a point where we

play27:01

don't actually know if anything we read

play27:04

see or hear is real and at that point

play27:08

it's going to be a dis and

play27:09

misinformation apocalypse where we have

play27:11

to re down to some sort of established

play27:13

way of determining the veracity of stuff

play27:15

and most of the human population is just

play27:18

going to tune out and then we're going

play27:19

to have to rethink our media landscape

play27:21

and it's gonna oh it could be bad

play27:23

[Music]

play27:33

thanks for watching a chatbot wrote that