What children can teach us about building AI

Dr Waku
5 Mar 202420:46

Summary

TLDRDieses Video untersucht die Geheimnisse des menschlichen Lernprozesses, insbesondere die Fähigkeit von Säuglingen, komplexe Sprachen und Motorik schnell zu erlernen, was AI-Algorithmen weit übertrifft. Professorin Elizabeth Spels Forschung in der Kindpsychologie zeigt, dass Säuglinge bereits im Schlaf lernen und sechs grundlegende Wissensarten teilen, die auch bei Tieren zu finden sind, sowie eine siebte, die Sprache, die einzigartig für den Menschen ist und kognitive Fähigkeiten katalysiert. Die Diskussion um AI-Forschung und das Potential, diese Wissensarten mit künstlicher Intelligenz zu verbinden, um KI-Systeme zu verbessern, wird ebenso behandelt.

Takeaways

  • 🧠 Kinder lernen komplexe Fähigkeiten wie Sprache und Motorik schneller als kürzlich erkannt, was für KI-Forschung relevant ist.
  • 👶 Elizabeth Spelke, eine Harvard-Professorin, untersucht, wie Kinder Wissen aufbauen und welche kognitiven Fähigkeiten sie von Geburt an mitbringen.
  • 🔍 Spelke verwendet das Infant Gaze Tracking, um zu erforschen, was Säuglinge gelernt haben, indem sie die Dauer ihrer Aufmerksamkeit auf Objekte misst.
  • 🌟 Säuglinge lernen bereits im Schlaf und können z.B. die Anzahl von Objekten oder die Worttrennung erkennen, bevor sie die Bedeutung der Wörter kennen.
  • 🗣️ Die Fähigkeit zur Worttrennung ist ein Kernaspekt der Sprachentwicklung und wird von Säuglingen intuitiv erlernt, bevor sie die Grammatik oder die Bedeutung der Wörter kennen.
  • 🧩 Professor Spelke identifizierte sechs Kernwissensbereiche, die sowohl bei Säuglingen als auch bei Erwachsenen und Tieren zu finden sind: Orte, Objekte, lebende Wesen, soziale Wesen, Zahlen und Geometrie.
  • 🌐 Die siebte kognitive Fähigkeit, die nur beim Menschen vorkommt, ist die Sprache, die als kognitives Katalysator für das menschliche Denken gilt.
  • 🤖 KI-Forschung konzentriert sich oft auf die Sprachfähigkeit, aber um KI-Systeme intelligenter zu machen, könnte es notwendig sein, auch die anderen sechs Kernkompetenzen zu integrieren.
  • 🔢 Ein Beispiel für KI-Forschung, die auf Spelkes Forschung aufbaut, ist die SLM, die einen kleinen Sprachmodell mit einem symbolischen Löser kombiniert, um arithmetische Probleme zu lösen.
  • 📈 KI-Systeme wie große Sprachmodelle sind derzeit nicht gut darin, hierarchisches Planen durchzuführen, was für KI-Systeme, die kompetent kodieren können, notwendig sein könnte.

Q & A

  • Was könnte das Geheimnis sein, um KI auf menschliches Niveau zu bringen?

    -Das Geheimnis könnte in den Gehirnen von Kindern liegen, da ihre Fähigkeit, komplexe motorische Fertigkeiten und Sprachen zu erlernen, weit über die Fähigkeiten der fortschrittlichsten maschinellen Lernalgorithmen hinausgeht.

  • Wie unterscheidet sich die kognitive Entwicklung von Säuglingen von der von KI-Systemen?

    -Säuglinge können in ihren ersten Lebensjahren ohne Anweisungen natürliche Sprachen erlernen, sich und Objekte bewegen, soziale Beziehungen aufbauen und Konzepte wie Zahl, Geometrie und Kausalität anwenden, während KI-Systeme dies oft nicht können oder es sehr langsam lernen.

  • Was ist der Unterschied zwischen dem Lernen von Säuglingen und dem von KI-Systemen?

    -Säuglinge können im Schlaf lernen und Informationen aufnehmen, während KI-Systeme normalerweise aktiv trainiert werden müssen. Säuglinge lernen auch, während sie schlafen, was KI-Systemen nicht möglich ist.

  • Wie verwendet Professor Spel die Methode der Säuglingsbeobachtung in ihren Studien?

    -Professor Spel verwendet die Säuglingsbeobachtung, um zu bestimmen, ob ein Säugling etwas gelernt hat, indem sie beobachtet, wie lange ein Säugling auf ein Objekt schaut, nachdem es ihm präsentiert wurde.

  • Was zeigte eine Studie über Säuglinge, die während des Schlafes Geräusche hören ließ?

    -Es zeigte, dass Säuglinge visuelle Interesse an Punkten zeigen, die einer Anzahl von Schallimpulsen entsprechen, die ihnen während des Schlafes gespielt wurden, was auf das Lernen während des Schlafes hindeutet.

  • Wie sind die sechs Kerntypen von Wissen, die Professor Spel in Säuglingen identifizierte, definiert?

    -Die sechs Kerntypen von Wissen sind Orte, Objekte, bewegte Wesen, soziale Wesen, Zahl und Geometrie. Diese Typen von Wissen interagieren nicht miteinander und scheinen unabhängig voneinander entwickelt worden zu sein.

  • Warum ist Sprache nach Ansicht von Professor Spel ein kognitives Katalysator für menschliches Denken?

    -Sprache ist ein kognitiver Katalysator, weil sie es ermöglicht, die sechs anderen Kernsysteme des Gehirns zusammenzuarbeiten, indem sie sie in Harmonie bringen und Ideen austauschen kann, was eine höhere kognitive Fähigkeit ermöglicht.

  • Was ist die Bedeutung der Kombinatorik und Rekursion in der Sprache für das menschliche Denken?

    -Die Kombinatorik und Rekursion in der Sprache ermöglichen es uns, einfache Konzepte zu kombinieren, um komplexe, hierarchisch strukturierte Gedanken zu bilden, was das abstrakte Denken ermöglicht.

  • Wie unterscheidet sich die Sprachfähigkeit von Tieren von der menschlichen Sprachfähigkeit?

    -Obwohl Tiere Kommunikationssysteme haben, die ein begrenztes Vokabular umfassen, können sie nur einen kleinen Teil der Konzepte ausdrücken, die sie in ihrem Kopf haben, im Gegensatz zur menschlichen Sprache, die nahezu jedes Konzept ausdrücken kann.

  • Was ist das Hauptziel der sLM-Forschung, die im Skript erwähnt wird?

    -Das Hauptziel der sLM-Forschung ist es, eine kleine Sprachmodell, das nicht in der Lage ist, numerische Probleme zu lösen, mit einem symbolischen Solver zu kombinieren, um die Fähigkeit des Systems zu verbessern, arithmetische Probleme in natürlicher Sprache zu lösen.

  • Warum sind hierarchische Planungsfähigkeiten für KI-Systeme wichtig?

    -Hierarchische Planungsfähigkeiten sind wichtig, weil sie es ermöglichen, über verschiedene Ebenen zu planen und zu denken, was natürlichen menschlichen Denkprozessen entspricht und für KI-Systeme zur Entwicklung von KI auf menschlichem Niveau notwendig ist.

Outlines

00:00

🧠 Kindliche Psychologie und KI-Forschung

Dieser Absatz stellt die Verbindung zwischen der kindlichen Psychologie und der KI-Forschung her. Es wird besprochen, dass die Fähigkeit von Kindern, komplexe Motorfertigkeiten und Sprachen zu erlernen, schneller ist als die von KI-Algorithmen. Die Rede ist von der Arbeit von Professorin Elizabeth Spelke, die sich mit der kindlichen Psychologie auseinandersetzt und versucht, herauszufinden, was Menschen von anderen Tieren unterscheidet. Ihre Forschung konzentriert sich auf das frühkindliche Erlernen von Sprachen und die Entwicklung von Wissen in verschiedenen Kategorien. Sie verwendet die Infant Gaze Tracking-Methode, um zu erforschen, wie Kinder lernen, und fand heraus, dass Kinder bereits im Schlaf lernen können. Diese Forschung könnte KI-Forschung inspirieren, indem sie zeigt, wie das menschliche Gehirn Wissen verarbeitet und wie es möglicherweise KI verbessern kann.

05:00

🔢 Sechse Kernwissensbereiche und die Sprache

In diesem Absatz werden sechs grundlegende Wissensbereiche beschrieben, die sowohl Menschen als auch Tiere teilen: Orte, Objekte, lebende Wesen, soziale Wesen, Zahlen und Geometrie. Diese Wissensbereiche interagieren nicht miteinander und scheinen unabhängig voneinander entwickelt worden zu sein. Menschen haben jedoch ein siebtes Wissen, das Sprache ist, das einzigartig für den Menschen ist und das Fundament für alle Sprachentwicklung im ersten Lebensjahr bildet. Sprache verbessert das Objekterkennen und die Zahlendarstellung und ist ein kognitives Katalysator, der es ermöglicht, dass verschiedene Teile des Gehirns zusammenarbeiten. Sprache hat rekursive Eigenschaften, die es ermöglichen, komplexe, hierarchisch strukturierte Gedanken zu bilden. Tiere haben zwar auch Kommunikationssysteme, diese sind jedoch eingeschränkt und können nicht alle Konzepte, die sie im Kopf haben, ausdrücken.

10:02

🤖 KI-Forschung und Kombination von Wissensbereichen

Dieser Absatz behandelt die KI-Forschung und wie man kleine Sprachmodelle mit symbolischen Lösern kombiniert, um Probleme zu lösen, die einen arithmetischen Anteil in natürlicher Sprache haben. Die Idee ist, Probleme in eine formale Sprache umzuwandeln, die dann an einen symbolischen Löser übergeben werden kann. Die Forschung konzentriert sich auf kleine Sprachmodelle mit weniger als 50 Milliarden Parametern, da numerische Fähigkeiten erst ab diesem Punkt erscheinen. Die Kombination aus Sprachmodell und symbolischem Löser ermöglicht es, arithmetische Probleme zu lösen, was direkt analog zu einer Kombination aus einem der sechs Kernsysteme und einem Sprachverarbeitungsmodul ist. Die Forschung zeigt, dass durch die Feinabstimmung des Modells anstatt eines erneuten Trainings eine tiefere Integration zwischen dem Sprachmodell und dem symbolischen Ausführungsmodul erreicht werden kann.

15:03

🏡 Hierarchische Planung und Weltmodelle

Hierarchische Planung ist ein menschliches Denkvermögen, das KI-Systeme derzeit nicht sehr bewältigen können. Um hierarchische Planung zu ermöglichen, ist ein hierarchisches Weltmodell notwendig, das verschiedene Ebenen der Realität repräsentiert. Das menschliche Gehirn kann aufgrund seiner verschiedenen kognitiven Fähigkeiten, wie räumliche Verarbeitung, Langzeitgedächtnis und numerische Repräsentation, automatisch hierarchische Pläne bilden. KI-Systeme, die nur über Sprachfähigkeiten verfügen, haben Schwierigkeiten, solche hierarchischen Pläne zu entwickeln. Die Integration von anderen kognitiven Fähigkeiten in KI-Systeme könnte zu einer verbesserten Fähigkeit zur hierarchischen Planung führen und zu KI-Systemen, die ähnlich kompetent sind wie der menschliche Verstand.

20:04

👋 Schlussfolgerung und Ausblick

Die Schlussfolgerung des Videos betont die Bedeutung der Forschung von Professor Spelke, die zeigt, wie Kinder lernen und welche sechs Kernwissensbereiche Menschen und Tiere teilen. Sie identifizierte auch das siebte System, die Sprache, als kognitives Katalysator, der es ermöglicht, dass verschiedene Teile des Gehirns zusammenarbeiten. In Bezug auf KI-Forschung wurde die SLM-Forschung vorgestellt, die kleine Sprachmodelle mit symbolischen Lösern kombiniert, um numerische Probleme zu lösen. Schließlich wurde die Notwendigkeit von hierarchischer Planung und Weltmodellen diskutiert, um KI-Systeme zu verbessern. Der Sprecher regt dazu an, über die verschiedenen Definitionen von Intelligenz nachzudenken und wie KI-Systeme intelligent sein könnten, unabhängig von menschlichen Perspektiven.

Mindmap

Keywords

💡Kinderpsychologie

Kinderpsychologie ist das Studium des Denkens, Fühlens und Verhaltens von Kindern. Im Video wird diese Disziplin genutzt, um zu untersuchen, wie Kinder schnell komplexe Fähigkeiten erlernen und wie dies für die Entwicklung von KI von Vorteil sein könnte. Die Forscherin Elizabeth Spelke untersucht, welche grundlegenden Kategorien des menschlichen Wissens existieren und wie das Wissen in jeder Kategorie von Anfang an entwickelt wird, was zentral für das Verständnis der kognitiven Fähigkeiten von Kindern und möglicherweise auch für KI ist.

💡Sprachentwicklung

Sprachentwicklung bezieht sich auf den Prozess, wie Kinder natürliche Sprachen erlernen. Im Video wird erwähnt, dass Kinder ohne Anleitung eine oder mehrere Sprachen erlernen und dies als Teil ihrer kognitiven Entwicklung. Die Fähigkeit, Sprache zu erlernen, wird als ein Schlüsselmerkmal der menschlichen kognitiven Fähigkeiten betrachtet und könnte für KI-Systeme, die nach menschlicher Intelligenz streben, von großer Bedeutung sein.

💡Kognitive Fähigkeiten

Kognitive Fähigkeiten umfassen die verschiedenen Aspekte des Denkprozesses, wie das Lernen, Behalten von Informationen und das Problemlösen. Im Video werden sechs grundlegende kognitive Fähigkeiten identifiziert, die sowohl von Menschen als auch von Tieren geteilt werden, und eine siebte, die spezifisch für Menschen ist: die Sprache. Diese Fähigkeiten sind für die Interaktion mit der Welt und die Entwicklung von KI-Systemen entscheidend.

💡Rekursive Kompositionalität

Rekursive Kompositionalität bezieht sich auf die Fähigkeit, einfache Konzepte zu kombinieren, um komplexere, hierarchisch strukturierte Gedanken zu bilden. Im Video wird dies als ein Schlüsselmerkmal der menschlichen Sprache und kognitiven Fähigkeiten dargestellt, das es Menschen ermöglicht, abstrakte Gedanken zu denken und sich von Tieren abzuheben. Diese Eigenschaft könnte für KI von entscheidender Bedeutung sein, die nach menschlicher Intelligenz strebt.

💡Kombinatorische Macht

Die kombinatorische Macht ist die Fähigkeit, Elemente zu verschiedenen Kombinationen zu bringen, um neue Ideen oder Lösungen zu schaffen. Im Video wird dies als eine Eigenschaft der Sprache beschrieben, die es ermöglicht, Konzepte flexibel zu verbinden und komplexe Gedanken zu bilden. Diese Fähigkeit ist für die kognitive Flexibilität und die Fähigkeit zur abstrakten Denkweise von zentraler Bedeutung.

💡Infant Gaze Tracking

Infant Gaze Tracking ist eine Methode, die verwendet wird, um zu bestimmen, ob ein Säugling etwas gelernt hat, indem man beobachtet, auf welche Objekte sie länger schauen. Im Video wird dies als eine Methode beschrieben, die von Elizabeth Spelke verwendet wird, um zu untersuchen, wie Säuglinge lernen und welche Wissenskategorien sie entwickeln. Dies ist ein wichtiger Aspekt der Forschung zur kognitiven Entwicklung von Säuglingen.

💡Symbolische Lösung

Symbolische Lösung bezieht sich auf die Verwendung formeller Algorithmen, um mathematische Probleme zu lösen. Im Video wird dies als eine Methode beschrieben, die in der KI-Forschung verwendet wird, um KI-Systemen die Fähigkeit zu geben, arithmetische Probleme zu lösen, indem sie kleine Sprachmodelle mit symbolischen Lösern kombinieren. Dies ist ein Beispiel für die Integration verschiedener kognitiver Fähigkeiten in KI-Systeme.

💡Hierarchical Planning

Hierarchical Planning ist der Prozess, bei dem Ziele und Aufgaben in einer Hierarchie organisiert werden, um komplexe Aufgaben effizienter zu planen und auszuführen. Im Video wird dies als eine Fähigkeit betrachtet, die KI-Systemen fehlt, und die für die Entwicklung von KI von hoher Intelligenz und Autonomie notwendig ist. Die Fähigkeit zur Hierarchischen Planung erfordert eine Hierarchie von Weltmodellen, die verschiedene kognitive Fähigkeiten integrieren.

💡Kognitive Katalysator

Ein kognitiver Katalysator ist ein Element, das die Interaktion und Integration verschiedener kognitiver Fähigkeiten fördert. Im Video wird die Sprache als einen kognitiven Katalysator beschrieben, der es Menschen ermöglicht, verschiedene kognitive Module wie numerische, räumliche und soziale Fähigkeiten zu verbinden und zusammenzuarbeiten, um komplexere Denkprozesse zu unterstützen.

💡Künstliche Intelligenz (KI)

Künstliche Intelligenz, oder KI, bezieht sich auf Systeme, die nach menschlicher Intelligenz gestaltet sind und in der Lage sind, Aufgaben zu lösen, die normalerweise intelligente Entscheidungen erfordern. Im Video wird KI als ein Ziel beschrieben, das durch das Verständnis und die Integration der kognitiven Fähigkeiten von Säuglingen und Kindern erreicht werden kann, was zu KI von noch höherer Komplexität und Autonomie führen könnte.

Highlights

The secret to unlocking human-level AI might lie in understanding how children learn.

Children's rapid mastery of complex skills outpaces current machine learning algorithms.

Elizabeth Spelke's research focuses on cognition in babies to understand human uniqueness.

Infant gaze tracking is used to study what infants have learned without their ability to speak.

Infants show interest in numbers and can learn concepts even while sleeping.

Babies are familiar with word segmentation before birth, a skill that takes adults much longer to learn.

Infants prefer listening to their native language and regional accents.

Professor Spelke identified six core types of knowledge shared by humans and animals.

Humans possess a seventh knowledge type, language, which is unique and a cognitive catalyst.

Language learning in humans enhances object recognition and potentially other cognitive functions.

Animal communication systems lack the complexity and combinatorial power of human language.

The paper 'sLM' discusses integrating a symbolic solver with a small language model for arithmetic reasoning.

Symbolic reasoning is incorporated into the fine-tuning of models, not just as a tool.

Hierarchical planning is a capability that current AI systems lack.

Integrating multiple core knowledge systems may lead to more advanced AI capabilities.

The video concludes by emphasizing the importance of hierarchical world models for AI planning.

Transcripts

play00:00

hi everyone the secret to unlocking

play00:02

human level AI might very well lie

play00:05

within the minds of children their rapid

play00:07

Mastery of complex motor skills and

play00:09

languages far outpaces the capabilities

play00:12

of our most advanced machine learning

play00:14

algorithms what could be the key to this

play00:16

remarkable learning prowess can the core

play00:19

knowledge and brain architecture

play00:21

observed in infants Inspire

play00:22

unprecedented advancements in AI keep

play00:25

watching to learn more this video has

play00:27

three parts child psychology the special

play00:29

soft

play00:30

and relation to AI research part one

play00:33

child psychology I recently attended an

play00:36

academic AI conference in Vancouver

play00:38

called triple AI it was a massive

play00:40

conference with I think 2,200 accepted

play00:44

papers I can't believe the scale of AI

play00:46

conferences and it had some very

play00:47

interesting invited talks one of those

play00:50

talks was by someone from a slightly

play00:51

different field Elizabeth S spel is a

play00:54

professor at Harvard and she studies

play00:56

child psychology essentially she's the

play00:58

author of a few books including what

play01:00

babies know and her research focuses on

play01:03

studying cognition in babies to figure

play01:05

out what makes humans different from

play01:07

other animals and in particular what are

play01:09

the Bedrock categories of human

play01:11

knowledge and how does knowledge in each

play01:13

category develop from its very

play01:15

Beginnings here's a quote from her talk

play01:17

abstract in their first years and with

play01:19

no instruction children learn one or

play01:21

more natural languages they develop a

play01:23

host of skills for moving themselves and

play01:26

manipulating objects they learn the

play01:27

paths connecting places in their homes

play01:29

and the bonds connecting people in their

play01:31

social networks and they deploy concepts

play01:33

of number geometry and causality that

play01:36

continue to guide the reasoning of

play01:38

adults how children do this is a great

play01:40

unanswered question so how does

play01:42

Professor spel study this problem and

play01:44

apologies if I'm mispronouncing her name

play01:46

there I don't remember the pronunciation

play01:48

the experimental method that she uses in

play01:50

her studies is to do infant gaze

play01:52

tracking basically she figured out that

play01:54

if an infant has learned something then

play01:57

they're going to look at that thing for

play01:58

a longer time period they're going to be

play02:00

more interested in it because we're

play02:01

talking about infants that are much too

play02:03

young to talk at this phase so you have

play02:05

to use an indirect method to figure out

play02:06

if they've learned something one

play02:08

experiment that the professor performed

play02:10

was to try to get infants to learn the

play02:11

difference between four objects and 12

play02:14

objects she had them look at four dots

play02:16

for example and then later was able to

play02:18

show that they had learned that because

play02:19

they would look more at four dots than

play02:20

they would at 12 and it's even

play02:22

multimodal an infant can listen to a

play02:24

sound that occurs four times and then

play02:26

later they'll be interested more in four

play02:28

dots basically their brain is learning

play02:30

the concept of number but what's really

play02:32

fascinating is that the researchers

play02:34

actually played The Sounds the four

play02:36

sounds for example or the 12 sounds in a

play02:38

row to infants while the infants were

play02:40

sleeping and even in that case they

play02:42

actually were more interested in four or

play02:45

12 dots visually at a later date so in

play02:48

other words even while they were

play02:49

sleeping the infants were picking up

play02:51

information and learning it there was

play02:52

another study that showed that babies

play02:54

are familiar with word segmentation

play02:57

before their birth word segmentation is

play02:59

basically the the problem of you listen

play03:00

to a spoken language and you can figure

play03:02

out where the word breaks are supposed

play03:04

to be adults tend to learn word

play03:05

boundaries very slowly if you've learned

play03:07

a second language as I have been doing

play03:09

with Japanese you'll know how difficult

play03:11

it can sometimes be to go was that one

play03:13

word or was that three but babies and

play03:15

native speakers know word boundaries

play03:17

intuitively this is even before they

play03:19

know the meanings of any of the words

play03:21

and before they know the grammar of the

play03:23

words as well but there's enough

play03:24

patterns in the speech that they're able

play03:26

to figure it out I actually have

play03:27

secondhand confirmation of babies being

play03:29

able to to learn word segmentation I

play03:31

have a friend who has English as her

play03:33

first language but her mother spoke

play03:35

Spanish while she was a really young kid

play03:37

I don't know if this was before or after

play03:39

her birth but regardless my friend was

play03:41

never taught any Spanish at all however

play03:44

when she tried to learn Spanish as an

play03:46

adult her teacher said you have the ear

play03:48

of a native speaker she was very good at

play03:50

reproducing exactly the right sounds and

play03:52

I would guess had very little trouble

play03:54

with word segmentation well again this

play03:56

is probably because both Spanish and

play03:58

English were being spoken spoken around

play04:00

her before she was born and as she was a

play04:02

very young child but I suppose the

play04:04

Spanish speaking must have stopped at

play04:05

some point and then she just picked up

play04:07

English as her only native language

play04:09

Anyway by using this infant gaze

play04:11

tracking metric researchers were able to

play04:13

show that babies pay more attention to

play04:15

someone speaking their native language

play04:17

and not only that but they pay even more

play04:19

attention to someone who's speaking with

play04:21

their own Regional accent slightly older

play04:23

kids would rather make friends with

play04:25

someone who sounds like them than looks

play04:27

like them if someone is of the same race

play04:29

but has a foreign accent for example

play04:31

then they would not really prefer that

play04:33

person as a friend compared to someone

play04:35

of A different race but that sounds like

play04:36

they came from the same neighborhood so

play04:38

word segmentation and being able to

play04:40

recognize your native language are

play04:41

really core parts of how infants and

play04:44

children interact with the world and all

play04:46

of that learning starts even before the

play04:48

baby is born part two the Special Sauce

play04:51

in her research and in her book what

play04:53

babies know Professor spel identified

play04:55

six core types of knowledge these six

play04:58

core types of knowledge that are

play05:00

represented in babies brains and adult

play05:02

brains are also shared with animals they

play05:05

are places objects animate beings social

play05:09

beings number and geometry these six

play05:11

types of core knowledge don't really

play05:13

interact with each other and therefore

play05:15

potentially evolved independently when

play05:17

an animal or human is engaging with one

play05:20

type of knowledge then they don't use

play05:21

the others and as soon as they switch to

play05:23

something else their brain completely

play05:24

switches focus and yes plenty of animals

play05:27

exhibit all the signs of these six

play05:29

knowledge types as well presumably they

play05:31

are what evolution deemed necessary to

play05:33

survive in the physical world however

play05:35

humans have a seventh knowledge type

play05:38

this seventh type is language and it's

play05:40

Unique to humans and it underlies all

play05:43

the language learning that happens

play05:44

within a baby's first year of Life we'll

play05:46

talk more in a minute about why animal

play05:48

systems of communication don't really

play05:50

count as language in this sense when

play05:52

humans are learning language doing so

play05:54

actually improves their object

play05:56

recognition if you learn the name for

play05:58

something it's easier for you to recogn

play05:59

ize it it might improve number

play06:01

representation as well this is my

play06:03

supposition but I think the reverse is

play06:05

true as well if you learn a different

play06:06

representation for a word or a concept

play06:09

that exists in your language then that

play06:10

will result in a more solid

play06:12

understanding it only seems natural for

play06:14

example recently I was looking at a map

play06:15

of Southeast Asia which is the first

play06:17

time I've looked at it in great detail

play06:19

in a while and I spotted where Myanmar

play06:21

was which I didn't really know and then

play06:23

when I started reading news stories

play06:24

about Myanmar it was easier for me to

play06:26

remember because I had a visualization

play06:28

or representation of what that country

play06:31

was in fact Professor spel talks about

play06:33

this seventh knowledge type or language

play06:36

learning as an actual cognitive Catalyst

play06:39

she thinks it's the secret ingredient

play06:41

for human cognition and what allows us

play06:43

to be so different from animals

play06:44

specifically she thinks that human

play06:46

language is the Catalyst that allows our

play06:49

numeric architectonic and social modules

play06:52

of our brain to all act in harmony it

play06:55

allows them to join forces swap ideas

play06:57

and essentially engage more multiple

play06:59

parts of your brain at once instead of

play07:01

just switching from one completely to

play07:02

the other here's a quote what's special

play07:04

about language is its combinatorial

play07:06

power we can use it to combine anything

play07:09

with anything apparently for example

play07:11

children start integrating what they

play07:13

know about their physical environment

play07:15

and their navigational sense at around

play07:17

the same time as the age when they begin

play07:19

to master spatial language and think

play07:21

about what's left and what's right

play07:23

another paper I read points out that

play07:25

language has recursive machinery and

play07:27

this recursiveness allows us to flexibly

play07:30

combine any concepts any concepts we

play07:32

like into complex hierarchically

play07:35

structured thoughts yes language as in

play07:38

speech itself is also hierarchical and

play07:41

contains very formalized structure but

play07:44

the idea here is that that structure

play07:46

that recursive property can actually be

play07:48

applied to thoughts as well as the paper

play07:50

puts it compositionality is a key

play07:53

component of linguistically structured

play07:55

thought so basically the ability to

play07:57

think about something very small and

play08:00

very simple and something else very

play08:01

small and simple and combine them

play08:03

together into a larger composite thought

play08:05

is really what makes abstract thought

play08:08

possible the legendary Chomsky even

play08:10

called this the basic property and

play08:12

argues that it evolved in humans because

play08:14

of the way that it allows us to

play08:15

structure thoughts rather than the way

play08:17

that it allows us to structure speech

play08:19

although that was definitely a desirable

play08:22

side effect so although Dr spel is

play08:24

basically calling the seventh system the

play08:26

language system I guess you could also

play08:28

call it the recursive ly structured

play08:30

system or the recursion and composition

play08:32

system or something like that so what

play08:34

about language in animals after all some

play08:37

animals seem pretty smart like you can

play08:39

train a dog to push a button when it

play08:41

wants to get the humans to do something

play08:43

and you have monkeys and chimpanzees and

play08:45

birds that all communicate with

play08:47

languages that seem to have reasonably

play08:49

large vocabularies but this is the

play08:51

difference pretty much any concept that

play08:53

a human might want to think about can be

play08:55

expressed in language but for all known

play08:58

animal communication systems the animals

play09:00

can only express a small subset of the

play09:03

concepts that they can actually

play09:04

represent in their head for example

play09:06

honeybees have color vision they can

play09:08

remember the colors of flowers that they

play09:10

visit but the honeybee language where

play09:12

they dance to communicate to other

play09:14

honeybees where to find food or whatever

play09:16

allows them to communicate only the

play09:18

spatial location the spatial direction

play09:20

and distance of that food rather than

play09:22

the color of the flower that they're

play09:24

sending them to and that's all the

play09:25

information that's actually needed for

play09:27

the honey bees to survive which is good

play09:29

but it means that the bees will never be

play09:31

able to scaffold thoughts upon thoughts

play09:33

and communicate to one another and build

play09:36

up a culture over time every honeybee

play09:38

has to start from scratch so the Special

play09:41

Sauce is language or recursive

play09:43

compositionality or whatever you want to

play09:45

call it therefore if you're trying to

play09:47

create really smart like human level AI

play09:49

it makes sense to concentrate on those

play09:51

six core types of knowledge that other

play09:54

animals have and combine it with the

play09:56

existing language expertise because so

play09:58

far most models skip over all of those

play10:01

six core competencies and they just head

play10:03

straight for the language which is great

play10:05

because that's the most complicated one

play10:07

and the one that we interact with on a

play10:08

day-to-day basis but it also means that

play10:10

if you ask a language model to do some

play10:13

reasoning it has only the language

play10:15

capabilities there it can't fall back on

play10:17

some spatial reasoning module to solve a

play10:20

geometric problem or whatever perhaps

play10:22

easier said than done though part three

play10:24

relation to AI research one of the

play10:26

papers I saw at triple AI was called s

play10:30

LM which I think stands for symbolic

play10:32

reasoning language model the idea of

play10:34

this work is to take a small language

play10:36

model they call it a Frugal language

play10:38

model because it doesn't have many

play10:39

parameters and get that model to offload

play10:42

some computations to a symbolic solver

play10:44

this would effectively be taking one of

play10:46

those six core competencies that I

play10:48

talked about and fusing it with the

play10:50

language model in this work they focused

play10:52

on small language models with less than

play10:54

50 billion parameters because apparently

play10:57

numeric reasoning capability only only

play10:59

starts to emerge once you exceed that

play11:00

amount the idea is to take in problems

play11:03

that have an arithmetic component in

play11:05

natural language and convert them into a

play11:07

formal language which you can then pass

play11:09

to a symbolic solver the symbolic solver

play11:11

will use mathematically precise

play11:13

algorithms to figure out the solution

play11:15

and then pass it back up to the language

play11:17

model which can answer the original

play11:18

queries for example you could imagine

play11:20

asking a language model to generate

play11:22

python code which then calls a symbolic

play11:25

execution engine they did experiments

play11:27

with two main models GPT DJ which has 6

play11:30

billion parameters and vuna which has 13

play11:33

billion as the larger model vuna can

play11:35

generate python code but gptj could only

play11:38

generate what they called pseudo code

play11:40

actually though what they call pseudo

play11:41

code is essentially an Assembly Language

play11:43

they can do additions subtractions and

play11:45

other operations on variables that the

play11:48

LM defined so the Assembly Language can

play11:50

look up variables from the original

play11:52

prompt basically or if you want to be

play11:54

technical it looks a lot like an SSA

play11:56

form inside a compiler but anyway how

play11:58

does this paper work so first they take

play12:00

the base model and they don't retrain it

play12:02

at all everything done in this work is

play12:04

based around fine-tuning there's two

play12:05

main stages first they use Laura low

play12:08

rank adaptation to fine-tune the model

play12:11

to the formal language generation task

play12:13

remember the formal language could be

play12:15

just python or it could be their pseudo

play12:17

code language second they fine-tuned the

play12:19

model to actually use the solver and

play12:21

they Incorporated a feedback loop to try

play12:23

to judge the quality of the formal

play12:26

language representation including for

play12:27

example whether the python code actually

play12:29

ran or not the results were pretty

play12:31

impressive for example they showed a 30%

play12:34

in terms of absolute percentage

play12:35

Improvement on one of the tasks compared

play12:38

to the base model that hadn't been

play12:39

fine-tuned along with the symbolic

play12:41

reasoning engine and that's a key Point

play12:43

as well this isn't just a system where

play12:45

they exposed a tool to an existing large

play12:48

language model which is sort of how

play12:50

internet search gets exposed for example

play12:52

instead the symbolic reasoning is

play12:54

actually included in the fine-tuning of

play12:56

the model not quite during the training

play12:58

phase days because they just used a

play13:00

pre-trained model but it's absolutely

play13:02

integral to the fine-tuning stage in

play13:04

other words during fine tuning the model

play13:05

would actually be generating some python

play13:07

running it checking the output this is a

play13:09

much deeper integration between the

play13:11

language model and the symbolic

play13:13

execution engine than just using it as a

play13:15

tool would be and because all their work

play13:17

was done on small models that did not

play13:19

need to be retrained but just fine-tuned

play13:21

this work was also very computationally

play13:23

efficient and easy for researchers to

play13:25

replicate so I'm a big fan of this work

play13:27

and it's interesting because it does

play13:29

represent combining one of those six

play13:31

systems along with the language

play13:33

reasoning capabilities and I suspect

play13:34

there will have to be a lot more work

play13:36

like this in the future as we start to

play13:38

put together heterogeneous combinations

play13:40

of different types of reasoning modules

play13:42

in fact that was very specifically one

play13:44

of the ways to build AGI that was called

play13:47

out and I talked about that in a

play13:48

previous video there was another talk at

play13:50

the triple AI conference by the famous

play13:52

deep learning Pioneer Yan Lun he called

play13:55

out one of the main things that current

play13:57

AI systems are lacking and that is

play13:59

hierarchical planning based on

play14:01

hierarchical world models so what does

play14:03

that mean well I've touched on that in

play14:05

the past as well by saying that large

play14:07

language models are essentially

play14:08

feedforward neural networks they can't

play14:11

think on something deeply for a long

play14:12

time because they don't really have

play14:15

Cycles they're not recurrent neural

play14:16

networks that's one of the reasons that

play14:18

if you want an llm to solve a

play14:20

complicated problem you should use chain

play14:22

of thoughts and ask it to split its

play14:24

reasoning into small steps because it's

play14:26

a lot more likely that the difference

play14:27

between one step and the next is going

play14:29

to be feasible for a feed forward neural

play14:32

network than trying to get it to do it

play14:33

all in one shot another way of putting

play14:35

this is that the llm doesn't have any

play14:38

planning capabilities it can't plan

play14:40

ahead and say oh in step five I'm going

play14:42

to need this so I should do this now it

play14:44

can figure that out if you ask it to

play14:46

explicitly break down its reasoning into

play14:48

steps but overall the planning abilities

play14:50

are not very good even harder is a

play14:53

hierarchical planning ability if you

play14:55

have a hierarchical plan it means you

play14:57

have a top level goal a slightly lower

play14:59

level goal and even lower level goals

play15:01

and furthermore the really lowlevel

play15:03

goals don't have to be worked out in

play15:05

detail in their entirety for example

play15:07

let's say my goal is to make a sandwich

play15:09

for myself but I don't have the

play15:11

ingredients for the sandwich so a lower

play15:13

level thing is I need to go to the

play15:14

grocery store to get those ingredients

play15:16

and lower level than that is I need to

play15:18

figure out which direction I should go

play15:20

in order to walk to the grocery store

play15:22

like which path would be best to take

play15:24

but in order to do that I need to figure

play15:25

out what the weather is to decide if I

play15:27

should wear a raincoat or just go out as

play15:29

I am but in order to do that I need to

play15:31

get up out of my chair and actually go

play15:32

to that room to check the weather so as

play15:34

soon as I think to myself wait I need to

play15:36

make a sandwich then that hierarchical

play15:38

plan can just form itself in my mind and

play15:41

suddenly I'm standing up to go check the

play15:43

weather and of course I didn't have to

play15:44

plan out all the Tiny Steps in advance

play15:46

once I step into that other room and

play15:48

figure out what the weather is by I

play15:49

don't know asking my Amazon device for

play15:51

example I didn't have to map out in my

play15:53

head all of the individual steps I'm

play15:55

going to take all the way to the grocery

play15:56

store or whatever because my brain can

play15:58

Define the plan as time goes on also

play16:01

important in this discussion is a

play16:03

hierarchical world model you can only do

play16:05

planning at different levels in a

play16:06

hierarchy if you have a model of the

play16:09

world that is actually accurate at

play16:10

different levels hierarchical planning

play16:12

seems very natural to humans but if you

play16:15

have certain conditions then it can be

play16:17

really hard for example if you have ADHD

play16:20

you might go out to see a friend and

play16:21

forget your jacket and get soaked by

play16:23

rain along the way it's easy to get

play16:25

fixated on a goal and Overlook the small

play16:28

things or to live in the current moment

play16:30

and not really plan ahead too much for

play16:32

ADHD it can depend on the energy level

play16:35

that you have whether or not you think

play16:36

about all the necessary steps and I

play16:38

think it's a skill we have to learn and

play16:40

consciously practice as well in fact

play16:42

when a professor first asked me a

play16:44

problem in other words at the equivalent

play16:46

of how do I get to the grocery store I

play16:48

didn't know how to answer him I was just

play16:49

like well you figure out the closest

play16:51

grocery store and you go there but he

play16:52

was like no no no in programming you

play16:54

first have to figure out get up from

play16:55

your chair take one step to the right

play16:57

Etc and of course course now if I'm

play16:59

doing a coding task or anything like

play17:01

that I'm engaged in many many levels of

play17:04

hierarchical planning so it will be

play17:06

interesting to see when AI systems can

play17:08

really engage in planning at similar

play17:10

levels or even more advanced levels

play17:12

because I think that's one of the key

play17:14

components that's missing when you think

play17:15

about can an AI system write code

play17:17

competently for example and the way this

play17:19

relates to everything else in this video

play17:22

is that you want hierarchical planning

play17:23

but that needs a hierarchical world

play17:25

model and I think it's difficult to have

play17:27

a hierarchical world model model if you

play17:29

just have one reasoning component in

play17:30

your system that only understands

play17:32

language but if you have Incorporated

play17:34

one or more of those other six core

play17:36

knowledge systems then I think it's much

play17:38

easier to have a hierarchical world

play17:40

model my spatial reasoning capabilities

play17:43

can tell me how to get up out of this

play17:44

chair and over to the other room my

play17:46

episodic or long-term memory can tell me

play17:49

which path which turns I have to take to

play17:50

get to the grocery store my memory of

play17:52

what food tastes like can help me

play17:54

remember which ingredients I need to

play17:56

make my sandwich and my number

play17:57

representation system can help me figure

play17:59

out if the grocery store is overcharging

play18:01

me for things or not so it will be very

play18:03

interesting to see what AI systems will

play18:05

be capable of once they start

play18:07

integrating hierarchical World models

play18:09

hierarchical planning and more of those

play18:12

six core competencies that humans and

play18:14

other animals have as a sort of

play18:16

cognitive Baseline finally in conclusion

play18:18

we talked about how Professor s spel

play18:22

work on studying child psychology has

play18:24

helped us figure out what babies know we

play18:27

discovered that babies can start to

play18:29

learn about the number of items they can

play18:31

start to engage their social muscles

play18:33

they start to engage language learning

play18:35

and word segmentation even before

play18:37

they're born and a lot of this learning

play18:39

even happens while they're asleep the

play18:40

professor also identified six core areas

play18:44

in which humans seem to share a lot of

play18:46

similarities with animals and those six

play18:48

core knowledge systems do not interact

play18:51

with each other if the brain is focusing

play18:52

on one then the other types of reasoning

play18:54

are shut off humans have a seventh

play18:56

reasoning system which is what deal

play18:58

deals with language and this seventh

play19:00

system is what enables humans to think

play19:02

abstractly to reason and to do far more

play19:05

than other animals can in fact this

play19:07

seventh system is a cognitive Catalyst

play19:10

because it enables those six other core

play19:12

systems to actually work together you

play19:14

can use some of this bit of reasoning

play19:16

along with some of that and combine it

play19:18

together with language we're using

play19:19

language somewhat Loosely here because

play19:22

structured thought recursive thought and

play19:24

compositional thought is actually the

play19:27

mechanism that enables these different

play19:29

types of reasoning to be used together

play19:31

and then once you have thoughts in a

play19:32

structured and whatever manner then by

play19:35

using language processing parts of your

play19:36

brain you can turn it into speech in

play19:38

terms of AI research we talked about s

play19:41

LM which takes a really small language

play19:43

model that is not capable of numeric

play19:45

reasoning and combines it with a

play19:47

symbolic solver this enables the

play19:49

combined system to solve arithmetic

play19:51

problems it's directly analogous to one

play19:53

of those six core systems the one that

play19:55

deals with numbers getting combined with

play19:57

a language processing module finally we

play20:00

also talked about hierarchical planning

play20:02

which is something that existing AI

play20:03

systems like large language models are

play20:05

not very proficient at hierarchical

play20:07

planning seems possible once you have

play20:10

hierarchical World representations that

play20:12

could come about for example because

play20:14

you're representing more of those six

play20:16

core knowledge systems so next time you

play20:18

walk out of the house without your

play20:20

wallet or your keys you know that you

play20:22

can blame your brain's hierarchical

play20:24

planning system if you liked this video

play20:26

please check out this previous one I

play20:28

made where I talked about different

play20:30

definitions of intelligence and how we

play20:32

could think about whether an AI system

play20:34

is intelligent or not outside of the

play20:36

human lens well that's all I have for

play20:39

today thank you very much for watching

play20:44

bye

Rate This

5.0 / 5 (0 votes)

Related Tags
KinderpsychologieKI-ForschungSprachentwicklungKognitive FähigkeitenInfant GazeLernprozesseSprachkompetenzSymbolisches DenkenKognitive KatalysatorenHierarchische Planung
Do you need a summary in English?