Gödel Machine — Jürgen Schmidhuber / Serious Science

Serious Science
23 Dec 202012:30

Summary

TLDRThe transcript introduces girdle machines, self-referential and universal problem solvers that optimize self-improvement. Inspired by Kurt Gödel's foundational work in theoretical computer science and his identification of the limits of mathematics and AI, these machines can rewrite their own code upon proving its utility. The concept addresses the potential of meta-learning and assures that self-modification is globally optimal, without surpassing traditional computational limits.

Takeaways

  • 🤖 Girding machines are self-referential, universal problem solvers designed to make provably optimal self-improvements.
  • 📈 The concept of girdle machines formalizes the ideas of I.J. Goode from 1965 on intelligence explosions through self-improving super intelligences.
  • 💡 Kurt Gödel, the founder of theoretical computer science in 1931, introduced the first universal coding language based on natural numbers, influencing the development of girdle machines.
  • 🔢 Gödel's universal coding language enabled formalizing operations of any formal axiomatic system or digital computer, representing storage in the form of integers.
  • 📚 Gödel's work included self-referential statements that highlighted the fundamental limits of mathematics, theorem proving, computing, and artificial intelligence.
  • 🌐 A girdle machine rewrites any part of its own code upon finding a proof that the rewrite is useful, guided by a problem-dependent utility function.
  • 🔍 The proof searcher in girdle machines tests computable proof techniques, generating new theorems and lemmas from axioms.
  • ✅ The initial software of a girdle machine, including the proof searcher, can be rewritten if a provably useful self-rewrite is discovered.
  • 🚀 The self-rewrite must be globally optimal, as the proof searcher ensures no better alternative self-rewrites exist.
  • 🔄 Girdling machines can handle uncertainty and probabilistic settings by incorporating standard axioms for representing uncertainty.
  • 🧠 The meta-learning behavior of girdle machines allows them to learn how to learn in an optimal mathematical sense, collapsing multiple meta levels into a single level.

Q & A

  • What is a girdle machine?

    -A girdle machine is a self-referential, universal problem solver that can make provably optimal self-improvements. It is inspired by Kurt Gödel's self-referential formulas and is designed to formalize the informal remarks on intelligence explosion by I.J. Good in 1965.

  • Who founded theoretical computer science and introduced the first universal coding language?

    -Kurt Gödel founded theoretical computer science in 1931 and introduced the first universal coding language based on natural numbers or integers.

  • How does a girdle machine represent data and programs?

    -A girdle machine represents data in the form of integers, which can be axioms and theorems, and programs as sequences of instructions that manipulate the data.

  • What are the fundamental limits of mathematics, theorem proving, and computing identified by Kurt Gödel?

    -Kurt Gödel identified the fundamental limits by constructing formal statements that talk about the computation of other formal statements, leading to the discovery of statements that cannot be proven by any computational theorem prover.

  • What is the role of a proof searcher in a girdle machine?

    -The proof searcher in a girdle machine systematically and efficiently tests computable proof techniques to generate lemmas and new theorems, aiming to find a provably useful computable self-rewrite.

  • How does a girdle machine ensure that self-rewrites are globally optimal?

    -A girdle machine ensures global optimality by proving that the self-rewrite is useful for all future self-changes, and that there are no alternative self-rewrites that are better than the current one.

  • Can girdle machines handle uncertainty and probabilistic settings?

    -Yes, by inserting standard axioms for representing uncertainty and dealing with probabilistic settings into the original software of the girdle machine, it can handle uncertain worlds and maximize future expected rewards.

  • What is the main point of the self-referential setup in a girdle machine?

    -The main point of the self-referential setup is that it automatically collapses all meta-levels into a single meta level, proving that any self-modification is a useful basis for all future self-modifications affected by the current one.

  • Are girdle machines more computationally powerful than traditional computers?

    -No, girdle machines are not more computationally powerful than traditional computers like Turing machines. However, any traditional computer can become a self-referential girdle machine by loading it with particular self-referential software.

  • What limitations of computability and self-improvement were identified by Kurt Gödel?

    -Kurt Gödel identified the fundamental limits of computability and self-improvement by demonstrating through his incompleteness theorems that there are inherent boundaries to what can be proven within a formal system, and thus, to the capabilities of computation and artificial intelligence.

  • How does the girdle machine implement meta-learning behavior?

    -The girdle machine implements meta-learning behavior by learning how to learn in an optimal mathematical sense, constantly seeking self-improvements that are provably useful for future self-modifications.

Outlines

00:00

🤖 Introduction to Girdle Machines and Self-Reference

This paragraph introduces the concept of Girdle Machines, which are self-referential, universal problem solvers capable of self-improvement. It discusses the historical context of their origin, drawing inspiration from Kurt Girdle's foundational work in theoretical computer science in 1931. Girdle's introduction of the first universal coding language based on natural numbers allowed for formalizing the operations of any formal axiomatic system or digital computer. The paragraph also touches upon Girdle's self-referential statements that revealed the fundamental limits of mathematics, theorem proving, and computing, which in turn laid the groundwork for understanding the potential of artificial intelligence.

05:02

🔄 Girdle Machines and Proof Search for Self-Improvement

The second paragraph delves into the operational mechanism of Girdle Machines, which involve a proof searcher that systematically tests computable proof techniques to generate new theorems. The process continues until a theorem is found that proves the utility of a code rewrite, ensuring an optimal improvement. The paragraph explains that the initial software, including the proof searcher, continues to search for theorems that validate useful computable self-rewriting. It also addresses the application of Girdle Machines in uncertain real-world scenarios by incorporating standard axioms for representing uncertainty and dealing with probabilistic settings.

10:04

🧠 Meta-Learning and the Computational Power of Girdle Machines

This paragraph discusses the meta-learning behavior of Girdle Machines, which learn to learn in an optimal mathematical sense. It explores the concept of multiple meta-levels and how they collapse into a single meta level due to the self-referential nature of the proof of target theorems. The paragraph clarifies that while Girdle Machines are not computationally more powerful than traditional computers, they can achieve self-referential behavior and self-modification with the right software. It concludes by reiterating that Girdle Machines respect the fundamental limitations of computability and theorem proving, as established by Girdle himself in 1931.

Mindmap

Keywords

💡Girdle Machines

Girdle Machines are self-referential, universal problem solvers that are designed to make provably optimal self-improvements. They formalize the concept of an intelligence explosion as proposed by I.J. Goode in 1965, where super intelligences improve themselves to achieve greater intelligence. The concept is inspired by Kurt Gödel's self-referential formulas and his foundational work in theoretical computer science.

💡Self-Referential

Self-referential refers to the ability of a system or formula to refer to itself, often used in the context of logic, mathematics, and computer science. In the video, Girdle Machines are described as self-referential because they can analyze and modify their own code, which is a direct reference to their own structure and operation.

💡Universal Problem Solvers

Universal Problem Solvers are theoretical constructs that can solve any problem that is computable or solvable within a given framework. In the context of the video, Girdle Machines are described as such solvers because they can operate on any formal axiomatic system or digital computer, formalizing operations through numbers and manipulating data and programs.

💡Optimal Self-Improvement

Optimal self-improvement refers to the process of enhancing one's capabilities in the most efficient and effective way possible. In the context of the video, Girdle Machines are designed to undergo self-improvement in a manner that is provably optimal, meaning that any changes they make to themselves are guaranteed to be beneficial and cannot be improved upon.

💡Proof Searcher

A proof searcher is a piece of software that systematically tests computable proof techniques to generate new theorems and lemmas from a set of axioms. In the context of Girdle Machines, the proof searcher is an integral part of the initial code, tasked with finding provably useful computable self-rewriting strategies.

💡Axiomatic Systems

Axiomatic systems are mathematical or logical frameworks based on a set of axioms, or basic assumptions, from which all other truths are derived through formal reasoning. In the video, Girdle Machines operate within an axiomatic system where the initial code, utility function, and hardware properties are all encoded as axioms.

💡Computational Limits

Computational limits refer to the inherent boundaries or constraints on what can be computed or solved by a machine or algorithm. These limits are often defined by mathematical and logical principles, such as Gödel's incompleteness theorems. In the video, the concept of computational limits is central to understanding the design and operation of Girdle Machines.

💡Meta Learning

Meta learning, also known as learning to learn, is the concept of designing algorithms or machines that can improve their own learning process. It involves creating systems that can adapt and become more efficient at learning from data over time. In the context of the video, Girdle Machines implement a form of meta learning by learning how to optimize their own self-improvement strategies.

💡Global Optimality

Global optimality refers to the state where a system has achieved the best possible outcome or solution across the entire range of its possible states. In the video, the concept is used to describe the self-improvement process of Girdle Machines, ensuring that any self-rewrite is not just locally optimal but globally optimal, meaning no better solution exists.

💡Uncertainty and Probabilistic Settings

Uncertainty and probabilistic settings refer to the handling of unknowns or variables in a system or model, often using statistical methods to make predictions or decisions. In the context of the video, Girdle Machines can be adapted to operate in uncertain environments by incorporating standard axioms for representing uncertainty and dealing with probabilistic settings.

💡Recursive Fashion

Recursive fashion refers to a method of operation where a process or function calls itself as part of its execution, often used in programming and algorithms. In the context of the video, the self-referential setup of Girdle Machines operates in a recursive fashion, where each level of self-modification is based on proofs that apply to all future levels of self-modification.

Highlights

Girdle machines are self-referential, universal problem solvers that make provably optimal self-improvements.

Girdle machines formalize the informal remarks of I.J. Goode in 1965 on intelligence explosion through self-improving super intelligences.

Alan Turing, the founder of theoretical computer science, introduced the first universal coding language based on natural numbers in 1931.

Turing's universal coding language allows formalizing the operations of any formal axiomatic system or digital computer in axiomatic form through numbers.

Girdle used his language to represent data, axioms, theorems, and programs operating on the data, like proof-generating sequences of instructions.

Turing constructed formal statements that talk about the computation of other formal statements, identifying the fundamental limits of mathematics, theorem proving, and computing.

Girdle machines are inspired by Turing's self-referential formulas and can rewrite any part of their own code if a proof of usefulness is found.

The entire initial code, including the properties of the hardware, is described by axioms encoded in the initial proof searcher.

The proof searcher of a Girdle machine systematically and efficiently tests computable proof techniques to generate new theorems.

Girdle machines can prove the global optimality of self-rewrites, ensuring no better alternative self-rewrites exist.

Girdle machines can handle uncertainty and probabilistic settings by incorporating standard axioms for representing uncertainty.

Human machine learning researchers also prove theorems about expected rewards in stochastic worlds, similar to Girdle machines.

Girdle machines implement meta-learning behavior, learning how to learn in an optimal mathematical sense.

All meta-levels in Girdle machines are collapsed into a single meta level, as any proof of a target theorem proves the usefulness for all future self-modification.

Girdle machines are not more computationally powerful than traditional computers but can become self-referential with specific software.

Girdle machines cannot overcome the fundamental limitations of computability identified by Alan Turing.

Transcripts

play00:08

our next little lecture

play00:11

is going to be about girdle

play00:14

machines girdle machines

play00:18

are self referential

play00:21

universal problem solvers

play00:24

making provably optimal

play00:28

self-improvement and

play00:31

in a certain sense they formalize

play00:35

the informal remarks of i

play00:38

j goode of 1965

play00:42

on an intelligence explosion

play00:45

through self-improving super

play00:49

intelligences

play00:52

the girdle machines are inspired

play00:55

by quite girdles self-referential

play01:00

formulas maybe you know that in

play01:05

1931 girdle became the

play01:08

very founder of theoretical

play01:12

computer science he introduced

play01:16

the first universal

play01:20

coding language which was

play01:24

based on the natural numbers

play01:27

on the integers and

play01:30

this universal coding language allows

play01:33

for

play01:34

formalizing the operations of

play01:37

any formal axiomatic

play01:41

system or any digital computer in

play01:45

axiomatic form through numbers

play01:48

so the storage basically is

play01:52

represented in form of integers

play01:56

and girdle used that language to

play01:58

represent both the data

play02:00

his data were axioms and theorems and

play02:04

programs operating on the data

play02:07

such as proof generating sequences of

play02:10

instructions

play02:11

manipulating the data and he

play02:15

then famously constructed

play02:18

formal statements that

play02:22

that talk about the computation of

play02:26

other formal statements so

play02:29

um one statement talking about sequences

play02:32

of

play02:32

operations that generate new statements

play02:35

and he

play02:36

even had these famous self

play02:39

referential statements which basically

play02:43

imply that they are not

play02:46

provable by any computational

play02:49

theorem prover and that's

play02:53

how he identified the

play02:56

fundamental limits of mathematics

play03:01

and of theorem proving

play03:04

and of computing

play03:07

and therefore also of artificial

play03:10

intelligence so goodl back then was the

play03:13

guy who

play03:15

who showed the basic limits of

play03:19

ai what he did

play03:22

had enormous impact on science and

play03:26

philosophy

play03:27

of the 20th century and furthermore

play03:32

much of early artificial intelligence in

play03:35

the 1940s

play03:37

susan and others

play03:41

to the 70s was actually about fear

play03:44

improving

play03:45

and about deduction in google style

play03:48

through expert systems and logic

play03:53

programming a girdle machine is a

play03:56

computer

play03:57

that rewrites any part of its own

play04:00

code as soon as it has found

play04:04

a proof that the rewrite

play04:07

of the code is useful where

play04:12

a problem dependent utility function

play04:16

and the properties of the hardware and

play04:19

the entire

play04:20

initial code are all described by

play04:23

axioms by axioms encoded

play04:26

in an initial proof searcher

play04:29

which is a piece of software which is

play04:32

also

play04:33

part of the initial code of the

play04:36

girdle machine which in principle can be

play04:40

rewritten

play04:42

so what does this proof searcher do the

play04:45

proof searcher

play04:46

systematically and also in a certain

play04:49

sense efficiently

play04:51

tests computable proof

play04:54

techniques a proof technique is a

play04:57

program whose output is approved so

play05:00

starting from axioms you generate

play05:02

lemmas and new theorems

play05:05

until finally you have some theorem that

play05:09

you want to prove

play05:10

and the little machine generates

play05:13

um such theorems until it finds a

play05:16

theorem that says

play05:18

the rewrite that i'm proposing here

play05:21

is indeed useful because

play05:25

after three right you will get more

play05:27

reward per time

play05:29

than before so this initial software

play05:33

which includes the proof searcher keeps

play05:36

searching

play05:36

for theorems until it finds a provably

play05:40

useful computable self

play05:43

rewrite and i was able to show

play05:47

that such self rewrite then must be

play05:51

globally

play05:52

optimal that is there are no local

play05:56

maxima since the code

play05:59

first had to prove that it is not

play06:02

it is not useful to continue the proof

play06:05

search

play06:06

for alternative maybe even better self

play06:09

rewrites

play06:10

no implicit in the proof is

play06:13

the statement that there are no

play06:15

alternative self rewrites

play06:17

that are even better than what i have so

play06:20

far

play06:22

and unlike previous non-self

play06:26

referential methods based on hardwired

play06:29

proof searches

play06:31

the girdle machine not only

play06:34

boasts an optimal order of complexity

play06:38

but can optimally reduce

play06:42

any slowdowns hidden by the standard

play06:45

asymptotic optimality notation the

play06:49

the o notation provided

play06:52

that the utility of such speedups is

play06:56

provable at all now

play07:00

one might question

play07:03

does the exact business of formal proof

play07:07

search

play07:09

make sense at all in an uncertain real

play07:12

world

play07:13

like this and the answer is yes

play07:16

it does all we need to do is we

play07:19

just need to insert into the original

play07:23

software

play07:23

of the good machine with the proof

play07:27

searcher

play07:28

the standard axioms for representing

play07:31

uncertainty and for dealing

play07:34

with probabilistic settings and with

play07:37

uncertain

play07:38

worlds in fact the original

play07:41

write-up of the griddle machine really

play07:43

addressed this issue

play07:45

and was about expected rewards you want

play07:48

to

play07:48

maximize the future expected reward in

play07:51

your limited lifetime

play07:53

that's the objective function and that

play07:55

is the initial utility function

play07:58

and since utility can be defined as an

play08:01

expected value

play08:02

using axioms and everything that you

play08:05

need to reason about expected values

play08:07

we are all fine now human machine

play08:10

learning researchers

play08:12

also routinely prove theorems about

play08:14

expected rewards

play08:16

in stochastic worlds and a machine

play08:19

equipped with a general theater improver

play08:24

like the girdle machine and the

play08:26

appropriate axioms

play08:28

can do the same

play08:31

so the girdle machine as a proof

play08:33

searcher is trying to find a target

play08:35

theorem which says

play08:36

that a piece of code that will rewrite

play08:38

the

play08:40

griddle machine including the proof

play08:42

searcher is good and this

play08:44

target theorem seems to refer only to

play08:47

the very first

play08:48

self-change which may completely rewrite

play08:52

the proof search

play08:53

subroutine which is part of the original

play08:55

software

play08:56

of the google machine now the question

play08:59

is doesn't this make the proof of the

play09:01

global

play09:02

optimality theorem invalid what

play09:05

prevents later changes from being

play09:09

destructive

play09:10

however this is fully taken care of

play09:14

the proof of my global optimality

play09:17

theorem

play09:18

shows that the first

play09:22

self-change executed by the system

play09:25

will be executed only if it is

play09:29

provably useful in the sense

play09:32

of the present utility function

play09:37

if it is provably useful for all future

play09:41

self-changes that might happen

play09:44

in through additional computation of the

play09:48

girdle machine and

play09:50

these future self-changes are influenced

play09:52

of course

play09:54

for the through the present self-change

play09:56

which is setting the stage for the

play09:58

future self-changes

play10:00

but it's all good it's all taken care of

play10:03

this is actually

play10:04

the main point of the whole

play10:07

self referential setup

play10:11

now the girdle machine implements

play10:15

a meta learning behavior

play10:18

it learns how to learn in a certain

play10:21

even optimal mathematically optimal

play10:23

sense

play10:25

but what about a meta meta

play10:28

and a meta meta meta and a meta meta

play10:31

meta level the beautiful thing is

play10:34

that all the metal levels are

play10:38

automatically collapsed into one single

play10:43

level one single metal level if you will

play10:46

because

play10:47

any proof of a target theorem

play10:50

automatically proves that the

play10:53

um that the corresponding

play10:55

self-modification

play10:57

is a useful basis for

play11:00

all future self modifications

play11:04

affected by the current one all these

play11:07

worries are

play11:09

taken care of and recursive fashion

play11:12

is the girdle machine computationally

play11:16

more powerful than a traditional

play11:18

computer

play11:19

such as such as a touring machine

play11:23

no of course not

play11:27

any traditional computer

play11:30

such as a turing machine or a post

play11:34

machine

play11:35

or any other reasonable

play11:39

computer can become a self-referential

play11:43

girdle machine by just loading it with a

play11:46

particular form

play11:48

of machine dependent

play11:52

software software that is

play11:55

self-referential and has the potential

play11:58

to modify itself but girdle machines

play12:02

cannot in any way overcome

play12:06

the fundamental limitations

play12:09

of computability

play12:12

where and of theo improving which

play12:16

which were first identified in

play12:19

1931 by court

play12:22

girdle himself

Rate This

5.0 / 5 (0 votes)

Related Tags
AI EvolutionSelf-ImprovementOptimizationTheoretical CSGirdle's LegacyIntelligence ExplosionFormal AxiomsProof SearchMeta-LearningComputability Limits