Max Tegmark | On superhuman AI, future architectures, and the meaning of human existence

Sana
20 May 202444:54

Summary

TLDRThe speaker reflects on their journey from pondering the cosmos to delving into artificial intelligence and neuroscience at MIT. They draw parallels between the unexpected advancements in AI and the history of flight, emphasizing the importance of humility in the face of rapid technological progress. The discussion explores the potential of AI to revolutionize education, the need for safety standards in AI development, and the future where AI could help solve humanity's most pressing issues. The speaker advocates for a future where humans remain in control, with AI as a tool for societal betterment, and stresses the importance of creating meaning in a universe that doesn't inherently provide it.

Takeaways

  • 🌌 The speaker's lifelong fascination with the mysteries of the universe and the human mind led them to a career in AI and neuroscience research at MIT.
  • 🤖 The rapid development of AI, particularly in language mastery, has surprised many experts, highlighting the importance of remaining humble and open to unexpected advancements.
  • 🔄 The speaker compares the evolution of AI to the development of flight, suggesting that sometimes simpler solutions emerge before fully understanding complex biological counterparts.
  • 💡 AI research can benefit from a broader perspective that includes both the vastness of cosmology and the intricacies of the human brain.
  • 🔗 The speaker believes that future AI advancements will likely incorporate elements from the brain's structure, such as loops and recurrence, which are missing in current transformer models.
  • 🧠 Human brains are efficient, using significantly less power than current AI data centers, indicating a need for more efficient AI models that can learn from fewer examples.
  • 🔑 The speaker suggests that the key to creating superhuman AI lies in combining neural networks with symbolic reasoning, similar to human cognitive processes.
  • 🌐 AI's ability to make analogies and generalize from one domain to another is a significant area of research, with implications for how models understand and translate knowledge.
  • 🚀 The potential for AI to revolutionize education by personalizing learning and providing deep insights is an exciting prospect for the future.
  • ⚠️ There is a call for safety standards and regulations in AI development to ensure that powerful technologies are used responsibly and ethically.
  • 🛠️ The speaker emphasizes the need for AI systems to be more interpretable and trustworthy, suggesting that AI could help us understand itself through automated reasoning and symbolic regression.

Q & A

  • What were the two biggest mysteries that inspired the speaker's career?

    -The speaker was inspired by the mysteries of the universe and the universe within the mind, which refers to intelligence and the workings of the human brain.

  • How has the speaker's research focus evolved over time?

    -The speaker started with researching the outer universe and then shifted their focus to artificial intelligence and neuroscience, conducting research at MIT for the past eight years.

  • What comparison does the speaker make between the development of AI and the history of flight?

    -The speaker compares the unexpected advancements in AI to the development of flight, noting that just as mechanical birds were not necessary to invent flying machines, understanding the human brain may not be necessary to create advanced AI.

  • What does the speaker think about the current state of AI and its potential future developments?

    -The speaker believes that current AI technologies, like transformers, will be seen as primitive in the future, and that we will develop much better AI architectures that require less data, power, and electricity.

  • How does the speaker view the relationship between the human brain and AI systems?

    -The speaker suggests that while AI systems like transformers are powerful, the human brain operates differently, using loops and less data. They believe future AI will incorporate elements of how the brain works.

  • What role do analogies play in the development of AI according to the speaker?

    -Analogies are crucial for both human reasoning and AI development. They allow AI models to derive higher abstractions and perform transfer learning between different domains.

  • How does the speaker perceive the future of AI in terms of its ability to supersede human knowledge?

    -The speaker is confident that AI will not only match but surpass human knowledge, especially with its ability to draw analogies across different disciplines and integrate vast amounts of data.

  • What potential changes in AI architecture does the speaker foresee?

    -The speaker anticipates a shift from the current focus on transformers to new AI architectures that may be more efficient and powerful, possibly incorporating symbolic reasoning and other elements of human cognition.

  • What is the speaker's perspective on the integration of AI with tools and its implications for the future?

    -The speaker sees the integration of AI with tools as a way to enhance AI capabilities, allowing AI to perform tasks and make decisions that combine the strengths of both AI and traditional software systems.

  • How does the speaker view the role of AI in education and knowledge dissemination?

    -The speaker believes AI can revolutionize education by providing personalized learning experiences, understanding students' knowledge gaps, and presenting information in engaging ways.

  • What is the speaker's stance on the importance of safety and ethical considerations in AI development?

    -The speaker emphasizes the need for safety standards and ethical considerations in AI development, advocating for regulations similar to those in other industries to ensure the responsible advancement of AI.

  • What historical event does the speaker draw a parallel with the current state of AI development?

    -The speaker compares the current state of AI to the moment when the first nuclear reactor was built, suggesting that significant advancements and potential risks are on the horizon.

  • What does the speaker believe the best-case scenario for the future of AI looks like?

    -In the best-case scenario, the speaker envisions a future where humans are still in control, major problems like famine and wars are solved, and AI is used responsibly for the betterment of society.

Outlines

00:00

🤖 Journey from Cosmology to AI and Neuroscience

The speaker reflects on their early fascination with the mysteries of the universe and the human mind, leading to a career that began with cosmology and transitioned into artificial intelligence and neuroscience research at MIT. They highlight the rapid advancements in AI, particularly in language and knowledge mastery, drawing parallels to the early days of flight. The speaker emphasizes the importance of humility in the face of technological progress and the potential for AI to evolve beyond current understanding, much like the transition from mechanical birds to flying machines.

05:01

🧠 The Future of AI Architecture and Learning from the Brain

The speaker discusses the limitations of current AI models, such as transformers, and suggests that future architectures will likely incorporate elements from the human brain, such as loops and recurrence, to improve consciousness and learning. They argue that AI can learn from fewer examples and with less energy than current systems, and that while understanding the brain is beneficial, it's not necessary to replicate its complexity. The speaker also touches on the potential 'missing ingredients' for AI to perform more advanced reasoning.

10:08

🔗 The Power of Analogies in AI and Human Reasoning

The speaker explores the concept of analogies in AI, drawing a comparison between the human brain's use of symbols and languages for reasoning and the way AI models can learn and transfer knowledge between different domains. They discuss the importance of pattern recognition and how AI systems can discover and apply these patterns to answer novel questions, even without explicit training. The speaker also delves into the idea of AI systems superseding human knowledge by drawing analogies across various disciplines.

15:09

🚀 The Potential of AI as Autonomous Agents and Tools

The speaker predicts that AI will evolve from language models to autonomous agents capable of performing tasks and making decisions to achieve goals. They discuss the potential for AI to use tools and databases effectively, similar to how humans use calculators and other aids. The speaker also considers the implications of AI in education, suggesting that AI could revolutionize teaching by deeply understanding students' knowledge and misconceptions, thus providing personalized and effective learning experiences.

20:14

🛠 The Path to Self-Improving AI and its Societal Impact

The speaker addresses the gradual transition of AI towards self-improvement, using the analogy of technological progress and productivity growth. They discuss the current state of AI development, where humans are still involved in the loop, but foresee a future where AI systems become increasingly independent. The speaker also emphasizes the importance of safety and transparency in AI, advocating for precautionary measures and responsible development to ensure AI is used for the greater good.

25:19

🌐 The Vision for a Safer and More Transparent AI Future

The speaker recounts the 2015 AI Safety Conference, which aimed to bring together leading figures in AI to discuss safety and ethical considerations. They highlight the importance of having a proactive conversation about AI's impact on society and the need for research into making AI safe and controllable. The speaker also reflects on the collective responsibility to shape the future of AI and the potential for a few dedicated individuals to make a significant difference.

30:23

💡 The Search for Novel AI Architectures and Innovations

The speaker advocates for thinking beyond the current transformer models and exploring entirely new AI architectures that could lead to more efficient and powerful systems. They draw parallels to historical shifts in computing architecture and suggest that AI could discover these innovative architectures, leading to significant advancements in AI capabilities with less hardware and energy consumption.

35:28

🔍 The Quest for Interpretability and Trust in AI Systems

The speaker discusses ongoing research aimed at making AI systems more interpretable and trustworthy. They describe their work in symbolic regression, which seeks to understand the complex formulas learned by neural networks. The speaker also talks about the potential for AI to help verify and prove the correctness of other AI systems, emphasizing the importance of safety and understanding in AI development.

40:34

🛑 The Call for AI Safety Standards and Ethical Considerations

The speaker compares the current state of AI to the early days of nuclear power, emphasizing the need for safety standards and ethical considerations. They argue for the establishment of regulations that would ensure AI technologies are safe before they are released to the public. The speaker also discusses the importance of balancing the benefits of AI with the potential risks and the need for a measured approach to AI development.

🌟 Envisioning a Future Where AI Contributes to Human Well-Being

The speaker paints a hopeful picture of a future where AI has been harnessed to solve some of humanity's most pressing issues, such as disease, climate change, and conflict. They emphasize the importance of creating a world where humans are still in control and where AI serves to enhance human life and well-being. The speaker also reflects on the meaning of life and the role of consciousness in creating meaning in the universe.

Mindmap

Keywords

💡Artificial Intelligence (AI)

Artificial Intelligence refers to the simulation of human intelligence in machines that are programmed to think like humans and mimic their actions. In the video, AI is the central theme, with discussions ranging from its development, safety, and potential future impacts. The script mentions AI's progression from simple tasks to complex problem-solving and the ethical considerations surrounding its advancement.

💡Cosmology

Cosmology is the scientific study of the universe's origin, evolution, and eventual fate. The speaker's interest in cosmology as a teenager is highlighted, indicating a lifelong fascination with big questions and mysteries, which later influenced their AI research by encouraging a broader perspective on problem-solving and pattern recognition.

💡Neural Networks

Neural networks are a set of algorithms designed to recognize patterns. They are inspired by the human brain and are a crucial component of AI. The script discusses artificial neural networks and their limitations compared to the human brain, suggesting that future AI architectures may incorporate additional features found in the brain for improved performance.

💡Transformers

Transformers are a type of neural network architecture that has proven highly effective in processing and understanding natural language. The script uses transformers as an example of current AI technology, comparing them to the early stages of electronic computing with vacuum tubes, suggesting that they are a stepping stone to more advanced AI systems.

💡Language Models

Language models are AI systems that are trained to understand and generate human language. The Large Language Models (LLMs) mentioned in the script are a subset of these, capable of impressive language-related tasks. The discussion highlights how these models have evolved and their potential to integrate with other tools and systems.

💡Consciousness

Consciousness, as a concept, refers to the quality or state of awareness, or the ability to experience thoughts, feelings, and surroundings. The script touches on theories of consciousness, suggesting that the loops present in the human brain might be necessary for subjective experiences, and how AI might one day replicate these aspects of cognition.

💡Recurrent Neural Networks (RNNs)

Recurrent Neural Networks are a class of neural networks that have a form of memory, allowing information to persist. The script contrasts RNNs with transformers, noting that the loops in RNNs could be a key ingredient for future AI systems, potentially contributing to their ability to mimic aspects of human cognition and consciousness.

💡Symbolic AI

Symbolic AI, also known as Good Old Fashioned AI (GOFAI), refers to AI systems that perform tasks through explicit, rule-based representations of knowledge. The script suggests a future where symbolic AI techniques are combined with modern neural networks to create more powerful and reasoning-capable AI systems.

💡Transfer Learning

Transfer learning is a machine learning method where a model developed for one task is reused as the starting point for a model on a second task. The script discusses how AI models can derive higher abstractions, allowing them to perform transfer learning between different domains, which is crucial for their ability to generalize knowledge.

💡Self-Driving Cars

Self-driving cars, or autonomous vehicles, are a practical application of AI where the technology can control a vehicle without human input. The script mentions self-driving cars as an example of how AI can be integrated with physical systems, suggesting that AI's role in robotics and automation is expanding.

💡AI Safety

AI safety refers to the research and practices aimed at ensuring that AI systems are developed and used in a way that minimizes harm and maximizes benefits to society. The script discusses the importance of AI safety, highlighting the need for research into safe and controllable AI systems, and the role of conferences and grants in promoting this field.

Highlights

The speaker's early fascination with the mysteries of the universe and intelligence, which directed their career path.

The unexpected advancements in AI, with chatGPT-4 being smarter than anticipated, reflecting the ease of creating thinking machines compared to understanding the brain.

The influence of cosmology on AI research, emphasizing the importance of looking for the bigger picture beyond current AI models.

The comparison between transformers and the human brain, suggesting future AI architectures may incorporate elements of how the brain works.

The potential for AI to learn from fewer examples and be more energy-efficient, taking inspiration from the human brain's efficiency.

The idea that future AI systems may combine symbolic reasoning with neural networks, similar to human cognitive processes.

The importance of analogies in AI, enabling transfer learning between different domains.

The discovery of patterns by AI systems, such as representing information in geometric patterns, contributing to their generalization capabilities.

The potential for AI to supersede human knowledge by drawing analogies across different disciplines.

The evolution of AI from language models to multimodal agents capable of performing actions to achieve goals.

The vision of using AI to revolutionize education by understanding learners' needs and tailoring teaching methods accordingly.

The concept of self-improving AI and the gradual transition towards systems that require fewer humans in the loop.

The importance of having a conversation about AI safety and the role of the 2015 AI Safety Conference in starting this dialogue.

The potential for AI to automate the process of understanding its own workings, contributing to increased trustworthiness.

The discovery of new scientific knowledge using AI, as demonstrated by the re-discovery of physics formulas and a new finding in climate chemistry.

The future potential of AI to meet safety standards, allowing for the development of more powerful and trustworthy systems.

The comparison of AI development to historical technological advancements, such as the nuclear reactor, signaling the need for caution and preparation.

The importance of establishing safety standards for AI to ensure beneficial outcomes and avoid potential risks.

Transcripts

play00:05

I remember when I was a teenager out on  Julittarvägen in Bromma lying in a hammock  

play00:11

between two apple trees. And I realized then  that I just loved thinking about big questions  

play00:17

and mysteries. And I always felt that the, the  two biggest mysteries of all were our universe  

play00:24

out there and the universe in here, intelligence  in the mind. And, and so throughout my career I  

play00:32

started with a, the outer universe and worked on  that a whole bunch. And then I got too excited  

play00:36

about artificial intelligence and neuroscience,  and that's what I've been researching here  

play00:41

at MIT for the past eight years. And, um,  it's just a crazy exciting time. You know,  

play00:47

just four years ago, most of my colleagues thought  that, uh, we were decades away from something as  

play00:55

smart as chatGPT-4, because they thought we  can never make machines that master language  

play01:00

and human knowledge until we figured out how  our brain works and we're nowhere near it still 

play01:05

. And it turned out there was an easier way to make  

play01:09

machines that think, well, and this reminds me a  lot of, uh, the situation with flight. You know,  

play01:17

in the year 1900, someone could have said, oh,  we're never gonna figure out how to make flying  

play01:21

machines until we figured out mechanical birds.  But that was completely wrong. 'cause there was a  

play01:26

much easier way to build machines that could fly  much faster than birds even. And I think that's  

play01:30

where we're seeing, um, that transformers that  are powering LLMs today are just incredibly simple  

play01:38

compared to the brain. And I, I think, um, we  should be very humble and never be too confident  

play01:44

that something is impossible to do soon, because,  um, there might be easier ways than we realized.

play01:52

How did those ideas from cosmology  influence your AI research?

play01:57

They influenced me by, I think make me always look  at the book for the bigger picture. I think today,  

play02:04

um, especially young people, almost put an  equal sign between artificial intelligence  

play02:08

and artificial neural networks. Some even put  an e equal sign between AI and transformers,  

play02:16

the particular kind of neural networks, the  powers, uh, chatGPT. I think that's way too  

play02:21

narrow-minded. I think, uh, transformers are gonna  be remembered as the vacuum tubes of AI. You know,  

play02:28

vacuum tubes was the first technology that  really let us build electronic computers.  

play02:33

But then we found better ones. And I'm pretty  confident that in not too many years we'll  

play02:38

have found much better AI architectures  than what we have now. Which let you do  

play02:43

the same thing with much less data, much  less power, much less electricity use.  

play02:49

Do you think the brain does something different  than a next token or next word prediction?

play02:55

For sure. Our brain, for example, is a, so-called  recurrent neural network where information flows  

play02:59

around in loops. There are no loops in, in a  transformer. And um, that's very interesting  

play03:06

actually, because one of the more detailed  theories of what causes consciousness,  

play03:11

the subjective experiences of colors and sounds  and love and so on, is that you need the loops for  

play03:19

it. But we can of course train recurrent neural  networks also. And I suspect that the new AI  

play03:28

architectures, which will end up being better than  today's transformers will probably combine a few  

play03:33

additional ingredients like this, that the brain  uses, that the transformers don't. Human brains  

play03:41

learn from often a lot fewer examples. We need  a lot less training data than the big AI systems  

play03:49

of today. And, uh, you see these data centers  going up using many, many millions of watts.

play03:57

You know, your brain uses 20 watts. So clearly  we have, we can get more ideas from the brain,  

play04:02

but we don't have to totally understand how  the brain works. When the brain was optimized  

play04:07

by evolution, it was constrained to  only develop a biological computer  

play04:12

that could self-assemble and that  could do it with only the most,  

play04:16

using the most common atoms in the periodic  table. You know, really weird constraints  

play04:21

that engineers just don't care about. Whereas  there was no limit on, there was no need for  

play04:28

it to be simple and easy to understand. Whereas  that's where we tend to be limited by engineers.

play04:35

And what, what do you think is these  missing ingredients? So you mentioned  

play04:40

recurrence. What do you think is required  to get these models to do more reasoning?

play04:48

I'm not gonna give you a glib answer because I  think, I don't know, and nobody knows exactly. But  

play04:53

if you very crudely think of the history of AI as  two stages. Stage one was gofi, good old fashioned  

play05:01

ai, where you have these symbolic logic-based  systems. And then stage two was self-learning  

play05:10

neural networks that largely crushed the old  stuff. This has made us think that the neural  

play05:17

networks are like the better cooler one, but  if you look in the live living world of animals  

play05:24

around us, it's kind of been the other way around.  Cats and dogs and eagles have often better vision  

play05:32

systems or all factor systems, et cetera, than we  do. What's special about humans is not that we can  

play05:39

see better than an eagle, it's that we can also  reason with symbols and communicate with human  

play05:46

language and mathematical language and programming  languages. And somehow we can combine the old AI  

play05:52

and the new AI in a seamless way. And that's how  I see the path to superhuman AI going as well.  

play06:00

People figure out how to take these powerful  new systems and merge them with a lot of these  

play06:06

more symbolic techniques so that, in a way  it's a bit more like how we humans do it

play06:12

What's really fascinating with this is  the analogies that the models are able  

play06:19

to take. So you train a model on code, it gets  better at human language as well. And so it's  

play06:29

able to derive some higher abstraction,  which allows it to do transfer learnings  

play06:34

between two entirely different domains  which speaks to sort of, you know,  

play06:40

how important analogies are. Analogies both for  the human brain, which seems to be the majority  

play06:47

of sort of what makes up how we reason. But  equally for these models. What do you think are  

play06:54

some of the most interesting sort of analogies  that you've come across when studying these  

play06:59

models? So how it's able to draw these conclusions  based on derived insights from different domains?

play07:06

Yeah, this is something we work a lot on in my  AI research group. Because eight years ago we  

play07:12

switched from doing physics to doing machine  learning. And we sit in this very office and,  

play07:16

and talk very much about these questions:  How do these analogies get discovered? How  

play07:21

do machine learning systems generalize  from one thing they've learned in one  

play07:26

domain to other things? And we published  some recent papers about it, which I find  

play07:31

really cool. And it's all about discovering  patterns, which neural networks are very,  

play07:35

very good at. You feed it massive amounts of data  and then it starts to see patterns like wait a  

play07:44

minute you know. For example, you have a large  language model read all the text on the internet,  

play07:52

and then it realizes that actually all these  places, it makes more sense if it chooses to  

play07:57

represent them in a two dimensional space, a  map, even if it's never seen a picture of a map.

play08:03

And then we wrote a paper where we looked inside  Llama 2, and there is a map sitting there.  

play08:09

Literally where Stockholm is here and Göteborg  is there. And once it's realized that it can  

play08:14

represent things like this, it can now start  answering questions that it never saw in the  

play08:18

training data. Like if you ask: is Göteborg west  of Katmandu? You know, it probably never saw that  

play08:25

question, but it has a map. It represented it this  way because to answer some other questions, right?  

play08:31

And now it can figure these things out. We see  same things when translating. For example, if you  

play08:39

look at how it represents words, it puts them in  a sort of high dimensional map, word embeddings.  

play08:45

And if you take, there was a nice paper recently  where someone took the word embedding of English  

play08:50

words from reading a bunch of English text and  word embeddings, and an Italian one from reading a  

play08:57

bunch of different Italian texts. And figured out  how do you need to rotate and shift these things  

play09:02

so that they kind of match well?

play09:04

And they got out a pretty good English Italian  dictionary just from these matching up these  

play09:12

patterns. So I think these AI systems  are learning a lot of these geometric  

play09:17

patterns and that's one of the key reasons  that they're often able to generalize and  

play09:25

actually answer whole new questions that you  have never, that they've never been trained on. 

play09:31

. That's, that seems to be the way that  

play09:35

which these models could also supersede current  human knowledge, where they're able to draw these  

play09:44

analogies, from very different disciplines in a  way which would've been infeasible for humans.  

play09:50

Maybe you're clever enough to sort of master all  human knowledge, but it seems like that's quite  

play09:55

a difficult thing to do today. But these models  can not only do that, but they can even fit it  

play10:00

into their short term memory soon with like the  context windows. So what, what's your perspective  

play10:07

on that? In these model's ability to supersede  the current human knowledge to go well beyond it?

play10:18

Yeah, I think it's absolutely going to  happen. No doubt about that. And I mean, if,  

play10:24

if you and I could just read all the Internet and  everything written in all the languages somehow,  

play10:29

and keep this in our minds, I'm sorry I cannot,  we would see a lot of patterns too. And when  

play10:34

I talk to colleagues of mine who are working on  training these models, they tell me themselves,  

play10:38

they were pretty shocked. Like, oh my gosh,  this thing can translate English to Chinese.  

play10:43

We never taught it to do that. Oh my gosh, it  can code in Python. Like how did it figure that  

play10:47

out? And what I think we might be seeing a lot  of in 2024 is patterns and connecting things,  

play10:57

not just in terms of the basic knowledge,  but also in terms of using tools.

play11:03

So right now these LLMs tend to output tokens, new  word or pieces of words, but people are realizing  

play11:11

you can also give them the option to output  commands, to a calculator that can do really good  

play11:18

math, or to a dictionary or to all sorts of other  tools or to a database. And this is something we  

play11:26

humans of course, do routinely. If you're driving,  you know, sending almost becomes like an extension  

play11:33

of your body. This can, this can start to bridge  and overcome a lot of the intrinsic weaknesses  

play11:39

in large language models. Because we have all  these old fashioned AI systems that can make very  

play11:44

powerful tools. So if the LLM learn to use them,  then they can get the best of both worlds. This  

play11:52

is also very powerful for building robots, whether  they be robots on wheels like self-driving cars or  

play12:01

more humanoid ones, where it can fundamentally  be controlled to some level by a large language  

play12:08

model that then some of the tokens and outputs  are just motor commands and things like that.

play12:15

And what do you think is the right way of  looking at these models? Because, you know,  

play12:20

initially they were language models, but  now they're increasingly multimodal. Would  

play12:25

you say they're sort of a compression  of all the world's knowledge? Are they  

play12:32

token predictors? Or what's the right way of  viewing the current state of these models?

play12:39

It's pretty diverse. I think it's almost easier to  look at the endpoint of where this is going rather  

play12:45

than trying to classify exactly what happens right  now. There's a huge commercial pressure to build  

play12:51

systems that are agents. traditional LLMs are more  like an oracle. You ask it a question and it gives  

play12:57

you an answer. But there's so much commercial  value to have agents which will actually do stuff,  

play13:02

take information and figure out what actions  to take to accomplish some goal and then go  

play13:07

out and do it. And I think 2024 will probably  also maybe be remembered as the year of the  

play13:14

agents when we start seeing a lot of more  autonomous systems. First purely software  

play13:19

ones on the internet. And then gradually we'll  get more and more physical agents too. This can  

play13:27

of course be great for many purposes, but  this is also obviously something we have  

play13:32

to be careful with. Because if we have a lot of  autonomous systems act in the world, you know,  

play13:38

it starts to feel less like just a new technology  like electricity and more like a new species.

play13:47

Do you think that we could radically  accelerate the rate at which we discover  

play13:51

new science? And if we do that,  what could that mean for the world?

play13:55

It could be amazing. I mean, I really like the  vision of your company to accelerate the growth  

play14:03

of knowledge and make it widely accessible to  everybody. This is to me one of the coolest  

play14:09

things in the history of life on Earth that  we've gone from being so disempowered, you know,  

play14:16

30 year life expectancy, knowing very little  about what was going on, to developing science  

play14:22

over thousands of years, becoming ffirst more  knowledgeable about what's actually happening.  

play14:28

And then through this knowledge, also developing  technology that put us more in charge and let us  

play14:34

control our destiny. Sort of be the captain of our  own ship. So I very much applaud this. I think if  

play14:41

you can have AI systems that can get very deep  insights and very broad knowledge themselves,  

play14:50

clearly that can revolutionize education. You  know, my job is not just out of a researcher, even  

play14:56

though that's what I spend most of my time on here  in this office, but also very much as an educator.

play15:02

And to me, the first thing I always have to do  before I start teaching something is to make  

play15:09

sure I really have a deep understanding of it.  Much deeper than what the students need to get  

play15:13

later. And then I need an understanding also  of the students, where are they? That's also  

play15:21

something AI can get better at. Understanding  what the people who are trying to learn,  

play15:25

actually know already and what they don't know  and what their misconceptions are. And then  

play15:29

finally, I cannot tackle this question. What  is the most helpful way for me, you know,  

play15:36

to convey this knowledge? Where do I start?  What are the metaphors and analogies that are  

play15:39

gonna work? And how should I present this so  that they keep being motivated and are curious  

play15:45

to learn more. I think AI can really help  revolutionize education in the broad sense  

play15:52

of the word. I certainly love the idea of  talking to some real expert in something  

play15:59

who's willing to give me some of their time and  answer questions in a way that makes sense to  

play16:04

me. It's one of the great joys of working at  this university. And if there are AI systems  

play16:11

that will play that role also and teach me things  the way I will really get it, I would love that.

play16:18

One of the assumptions with those  sort of historical reasonings has  

play16:22

been that they are self-improving.  The current state of these models is,  

play16:29

you know, you have to buy massive amounts  of compute, you need to spend six months  

play16:35

training them. There's alot of small  things that could go wrong when,  

play16:41

when you run these clusters. What's your intuition  there? Because they're not really self-improving  

play16:48

now. So how could we sort of hit that escape  velocity given how they're currently trained?

play16:55

So I think it's a mistake to think that now over  time now itself, it's not self-improving. Now  

play17:01

it's not self-improving and all of a sudden it is.  And boom, singularity. It's a gradual transition.  

play17:08

Technology has always been self-improving.  We always use today's technology to build  

play17:15

tomorrow's technology, which is why we've seen an  exponential growth in productivity in technology  

play17:22

of most measures. But what we also see in this  exponential growth, if you look at the world's GDP  

play17:29

over time or most measures of tech, the time it  takes to get twice as good to keep shortening. So  

play17:38

it's actually growing faster than exponentially.  And today, you have must have a lot of people  

play17:46

doing coding in your company right who are  using autopilot of some form. I don't know  

play17:53

if it's making them twice as productive or  1.5 times or three times as productive, but  

play17:59

this is already an example of how AI is  enabling AI development to go faster.

play18:06

So it is a kind of recursive self-improvement,  but there's still humans in the loop. Iit's just  

play18:12

that over time we get less and less humans  in the loop. When we had farming before,  

play18:18

like almost all people in Sweden were in that  loop, farming. Now it's like 1% in that loop.  

play18:23

Similarly with software development, there'll  gradually be fewer and fewer. If you look at  

play18:27

a car factory today, there are way fewer  people per car produced than there were 50  

play18:31

years ago. So I think that's how it's gonna  go. We'll see that, uh, you need fewer and  

play18:37

fewer and fewer humans in the loop. At some  point there might be no human in the loop,  

play18:42

and then things will take off even faster. But  you're going to start noticing the approach.

play18:49

Why do you have such a strong bias  towards humans compared to the AI models?  

play18:56

Because I am a human, you know, I have a  lovely little 1-year-old Leo. And when I  

play19:03

look into his eyes, you know,  I feel he's my son. Of course,  

play19:09

I should feel more loyal to him than to  some random machine, shouldn't I? And I  

play19:16

feel we humans have figured out collectively  how to start building AI. And so I think we  

play19:25

have the right to have some influence over  what future we build. I'm on Team Human here  

play19:31

not on Team Machine. I think we want to have the  machines work for us, not the other way around.

play19:36

I saw this awesome picture from the 2015  AI Safety Conference that you organized.  

play19:43

Can you tell us a bit more about who  was there and what you talked about?

play19:46

Oh yeah. That was quite fun. Kinda like  approaching the 10th anniversary now.  

play19:52

Since I like to think big, I felt it was really  important to start having a conversation with  

play20:01

all the leading players on how we can make AI get  used for good things rather than for bad things.  

play20:07

And I felt the situation was totally dysfunctional  in 2014 because on the one hand you had a small  

play20:14

group of people who were worrying about building  smarter than human AI that we would lose control  

play20:19

over. And then on the other hand, there were other  people actually driving the AI development in the  

play20:24

big companies and in academia who had never really  talked with those people who were worried and felt  

play20:31

that those are just a bunch of weirdos, maybe  they should ignore them because it could be  

play20:34

bad for funding. And so I had this vision that  we should actually try to bring them together.

play20:40

And, you know, one thing that Sweden is  strangely good at is throwing good parties  

play20:48

from the Nobel Fest to the more informal ones.  So I pulled out all the stops to try to entice  

play20:54

people to come to this conference. Put  it not in Sweden, but in Puerto Rico  

play20:59

in January, I sent out a invitation with a picture  of a guy digging his car out from one meter of  

play21:03

snow right next to the beach by the hotel and  be like, where would you rather be in January of  

play21:09

2015? And I went first after the most high profile  people I thought maybe I could actually persuade.  

play21:18

And then when you get them, you know, you can  start building a Swedish snowball that rolls  

play21:21

down the hill and you get more. And in the end  we got amazing people. We got Demis Hassabis, the  

play21:27

CEO of DeepMind, we got Elon Musk, we got really  top professors from academia and also all the top  

play21:33

people who had been worrying, you know, Elliot  Rutkowski, to Nick Bostrom and so many others.

play21:41

And to make sure they didn't kill each other. We  had a lot of wine and it was very pleasant. And I  

play21:45

was really happy with how this ended up. Everybody  came together and signed this statement saying,  

play21:53

yep, the risks are real. Let's deal with them.  Let's do research not just on how to make AI more  

play22:01

powerful, but also on how to make it safe and  transparent and controllable. Elon Musk said,  

play22:08

okay, I'll give you 10 million to do a grants  program to start getting a bunch of nerds  

play22:13

working on this. So we launched that and pretty  quickly this field stopped being taboo to talk  

play22:21

about AI safety and the technical AI conferences  would start having the nerd sessions basically  

play22:28

on these topics. So that also made me realize  that even very few people can often make a big  

play22:36

difference. Many times things that need to happen  don't happen just because of the bystander effect,  

play22:44

you know? And I think that's a message to  anyone listening to this. If they have an  

play22:48

idea for a new startup or just some new social  movement or anything, and they're like, ah,  

play22:54

this must be impossible because otherwise  someone else would've done it. No, it might  

play22:58

very well be that nobody else  actually tried very hard yet.

play23:02

What were some of the things that people  disagreed with the most at the conference?

play23:07

I mean, there were big differences in  the forecasts people had for how long  

play23:12

is it gonna take to get smarter than human AI?  There were also big disagreements whether people  

play23:19

thought it would probably be fine or great, or  whether it would probably suck. But everybody  

play23:26

pretty much agreed that, you know, it might  happen and it might happen soon. So you know,  

play23:35

with that humility, it makes sense to take some  precautions. Even the people who thought it was  

play23:41

very unlikely that humanity would go extinct,  you know, these are people who still buy fire  

play23:46

insurance on their house. That doesn't mean they  think their house is probably gonna burn down,  

play23:51

but just in case, you know, why not prepare  a little bit, maybe have also put in a smoke  

play23:56

alarm and have a fire extinguisher handy. And,  you know, given the enormous amounts of money  

play24:02

we're spending right now on training ever  bigger models and making AI more powerful,  

play24:08

it's pretty reasonable, even for people who are  sort of skeptical of the risks to say, well,  

play24:12

let's spend at least some substantial amount also  on figuring out how to make these systems safe.

play24:18

And one of the best boxes you have in your office  is the bananas box. What do you think is some  

play24:24

of the most bananas ideas that could work that  we are not paying enough attention to? For AI?

play24:32

Generally

play24:34

I think for AI in particular, if there is  something really bananas that we're not  

play24:40

paying enough attention to, it's probably  that we're thinking too small in terms  

play24:43

of architecture design. We are looking at the  transformer and thinking about like minor tweaks,  

play24:50

but there could easily be completely different  types of architectures. Even just look at the  

play24:54

history of computing, how many times we've had a  quantum leap in architecture. First Alan Turing  

play24:59

and Charles Babbage and so on, they were like  thinking about mechanical computers and there  

play25:04

was a pretty big shift to go from that to starting  to do electronic ones like eniac and and so on.

play25:13

And those computers, which is like when I was  a teenager and first learned to code, you know,  

play25:19

that was again, a completely different paradigm  of computation where you program everything in  

play25:23

and compile it into machine code than a  neural network that just teaches itself.  

play25:28

Our brain, you know the solution that  biology came up with is, again, very,  

play25:33

very different. And if you just take your  step back and say, I'm gonna give you a big  

play25:39

blob of atoms, you know, what's the best way  to arrange these to be really smart? You know,  

play25:45

it might be something we haven't thought about  at all. And the neat thing is if we start getting  

play25:52

ever more powerful AI to the point where we can  use AI to figure out how to make much better AI,  

play25:57

it will probably discover those really, really  clever things. So my prediction is actually that  

play26:03

even now we're getting these evermore ginormous  data centers the size of an airplane hangar where  

play26:10

you have to almost put a nuclear reaction next  to it soon to power the whole thing. That's just  

play26:14

temporary. I suspect that we're overcompensating  for a really poor software architecture with these  

play26:24

ridiculous amounts of hardware and training data.  Once we get over that hump, we'll realize that  

play26:30

you could do it all with much less hardware,  much less energy and much less training data.

play26:38

I mean, it's quite absurd that we weren't  using like mixture of experts and so on in  

play26:42

the prior design of some of these models  and, and also that we weren't curating  

play26:47

the data sets sufficiently. What do  you think are the most promising,  

play26:50

like new architectures, that you've seen  that you think we should be exploring more?

play26:57

So I mentioned tool use by LLMs. I think  more generally what's very promising is  

play27:03

scaffolding. Where you think of the  LLM as just one component in a bigger  

play27:11

architecture. Kahneman who sadly passed  away around age 90 just very recently,  

play27:21

used to talk about System One and System Two in  our human brains. System One, the fast thinking,  

play27:26

the intuitive thinking is a lot like neural  networks where System Two is the more logic  

play27:31

based, symbolic reasoning we have which lets us  speak Swedish and English and math and Python.  

play27:39

If you think about a future where the neural  networks, the transformers say is System One,  

play27:49

you can imagine that it's just part of this bigger  architecture which can enable clever, more clever  

play27:56

database structures, all sorts of tools, loops,  various other techniques. I could totally see  

play28:05

this kind of neural network scaffolding around  it being way more powerful than today's systems.

play28:13

Do you think it's necessary to change the  underlying model architecture or can this  

play28:17

be a heuristic on top? I mean, we already  see that with sort of the chainof thought  

play28:23

and similar approaches where you sort of  add reasoning on top of the current model.  

play28:28

So instead of stopping them as soon  as they start outputting tokens,  

play28:32

you let them set up a plan and then  recent sort of recurrently in that way.

play28:37

I think you're already seeing more things like  this. So the quiet star paper that just came out,  

play28:42

for example. A radical interpretation of it is  that it's stupid to force the neural network  

play28:48

to output one token, to basically say everything  that it thinks. That's not how you operate. When  

play28:53

you speak, sometimes you'll have several thoughts  before you decide to say the next word, right?  

play28:58

So the quiet star architecture has more of that  freedom and it performs way, way, way better. But  

play29:04

that's just a very small example of how messing  a little bit with the architecture can make big  

play29:10

improvements. And I'm fairly confident we're gonna  see huge improvements in the next year or two.

play29:17

Max, you've outputted some really interesting  

play29:19

stuff in the research around symbolic  regression. What's the latest there?

play29:24

So the technical research we focus on in my  group is very much about taking a black box  

play29:33

AI system that's doing something intelligent and  automating the process of figuring out what it  

play29:38

works, how it works, so we can make, see how  trustworthy it is and hopefully make it even  

play29:43

more trustworthy. So the simplest example  of this is if you have a neural network,  

play29:49

which is just computing some function of some sort  that it's learned somehow from data. The task of  

play29:55

figuring out what the formula actually is that  it's learned called symbolic regression. If it's,  

play30:02

for the nerds who are listening to this, if it's a  linear formula, it's just linear regression which  

play30:06

is super easy. But if it's some complicated  formula, like some of the physics formulas  

play30:12

there on my, by my window, it's generally  believed to be np hard. It could take longer  

play30:17

than the age of the universe just because there's  exponentially many formulas that are of length.

play30:23

And, but, but that, that one we managed to get  state of the art performance on. We used a lot  

play30:28

of ideas from physics where we were able to  discover automatically if this neural network  

play30:35

is actually modular and can be decomposed into  smaller pieces of two different parts. And we  

play30:41

actually were able to rediscover many of the  most famous physics formulas just from data.  

play30:47

So if you could go back in a time machine, we  could have scooped Einstein and others on some,  

play30:51

some cool stuff. And we actually even managed to  very recently discover a new physics result in  

play30:58

in climate chemistry about ozone, which  we actually got published. So that was the  

play31:04

first time we used these tools to advance  science a little bit. But more broadly,  

play31:10

what we wanna do is ultimately be able to  take any black box AI system and figure  

play31:17

out what is, what algorithm is learned and  what, what knowledge it's really learned.

play31:22

Our last paper on this, we was one where we,  we took about 60 different algorithms and  

play31:30

we trained a neural network to just learn to  do these tasks. So now you have a black box,  

play31:35

but what did it do? And then we had an automated  AI system which was able to figure out exactly  

play31:41

what it was doing and turn it into Python code  so you can like, ah, and we were quite excited  

play31:48

about this and we're scaling this up now in  a big way, see if we can make it work for,  

play31:54

for bigger systems. And the vision I have  here is that if you have a system, an AI  

play32:04

system that's gonna affect people's lives in some  way where you really want to have a high trust,  

play32:09

then it's a really good idea if you let  the machine learning do the learning,  

play32:14

we don't have any better way than that, but then  can distill out what it's learned into Python or  

play32:20

something else where you can actually prove  that the code meets the specs that you have.

play32:26

And I fundamentally believe that it's possible  because we humans can do it. You know,  

play32:32

like when you were a little kid, if your dad threw  you a tennis ball, you could catch it because your  

play32:38

brain had figured out how to compute parabolic  trajectories. But when you got older you're like,  

play32:45

oh, this is a parabola, Y equals X squared  you know formula. And this is how scientists  

play32:50

generally first intuitively figure out stuff, even  though they don't know how their brain works, and  

play32:56

then they learn to distill out the knowledge in  a way that you could, you could actually program  

play33:00

into a moon rocket, et cetera. And the fact that  we've managed to make so much progress on this  

play33:06

already and the fact that this field is rapidly  growing, I organize the biggest conference so  

play33:10

far here at MIT on this field called mechanistic  interpretability, makes me pretty hopeful actually  

play33:17

that we don't have to be resigned to the idea  that we'll never understand AI systems, the ones  

play33:24

that we really want to understand, I think we can  use other AI systems to help us understand them.

play33:31

Will the models derive those equations or  will they have a much simpler heuristic?  

play33:36

If you take the example, I throw a ball at  you and then you're not running the sort of  

play33:40

precise calculations, you just have a  very simple heuristic given the wins,  

play33:45

the size of the ball and so on. What do you  think these models will do? Will they have  

play33:50

that simple heuristic or will they actually  incorporate the exact equation as well?

play33:54

Our system actually does both. So  it actually finds many formulas.  

play34:00

We put them in a plot where on one axis is how  complicated they are and on the other axis is  

play34:06

how inaccurate they are. And it would find, like  in this case, the simplest one that is just the  

play34:14

parabola where there is no air resistance  whatsoever, but then it would find a more  

play34:20

complicated one, which is more precise. And  this is a lot like how we humans do it also.  

play34:28

And sometimes all you want is a  simple one, but in the big picture,  

play34:33

I think there's been too much pessimism on this  question about whether we'll able be able to  

play34:39

build systems that we can actually trust and  even prove things about. Because people have  

play34:44

made this mistake of assuming that all the work in  interpreting and proving has to be done by people.

play34:50

But AI systems are getting really good at that  stuff now and they can help us. And you might  

play34:55

think: Oh, how am I ever gonna trust a system  that was produced by an AI with a proof produced  

play35:02

by an AI if I don't understand if both the proof  and the code is too long to read? That's okay.  

play35:09

Because it turns out that just like it's much  harder to find a needle in a haystack than it  

play35:15

is to prove that it's a needle once you found  it, it's much harder to find the correct code  

play35:21

and proof that it meets your spec than it is  to verify that it works once you have it. So  

play35:27

all you have to do is actually understand  your proof checker, which can be 300 lines  

play35:31

of Python and now you can fully trust some very  powerful systems that have been made by an AI.  

play35:41

And you've had a lot of awesome folks at the  university and there's this one story about  

play35:49

Douglas Engelbart running into Marvin Minsky and  Marvin Minsky telling Engelbart about all of the  

play35:55

things he will make computers do, they will  reason and they will be conscious and so on  

play36:01

to which Engelbart replied: You're gonna do all  of that for computers. What are you gonna do for  

play36:06

humans? What are some of those stories? Did you  have any interactions with Minsky or other ones  

play36:14

this sort of OG AI folks and how did you see their  perception of AI change as the field advanced?  

play36:24

Most top AI thinkers and business leaders  that I know don't have as much time to step  

play36:35

back and reflect as they probably wished  they did. It's tough to run a company,  

play36:39

for example. And I get this as many of  them, most of had this idea of: Well,  

play36:45

first you just need to get this thing to work and  then I'm gonna figure out my strategy for how to  

play36:49

make sure this is good for society you know. And  then many of them were really taken by surprise  

play36:53

that chat GPTs and stable diffusion and so on  came decades before many expected. Like oh my God,  

play37:00

now what are we gonna do about this? What  I would love to see ultimately happen to  

play37:06

make sure we get a good future is actually  to give back some time to these people.  

play37:12

If you look at all other technologies that have  the potential to cause harm, we have a solution  

play37:17

for how we do it with all of them, whether it  be airplanes or medicines. We always have safety  

play37:23

standards. You know. If AstraZeneca comes up and  says: We have a new miracle drug that's gonna cure  

play37:30

cancer and we're gonna start selling it in ICA  tomorrow, Läkemedelsverket would be like: Where is  

play37:39

your clinical trial? Oh you couldn't be bothered  making, you haven't had time to make one yet? Just  

play37:45

come back when you've done the clinical trial  and then we will see if you meet the standards,  

play37:51

right? So that buys time for everybody involved.  It's the responsibility. The companies now have  

play37:58

an incentive to figure out what the impact on  society is gonna be. How many percent of people  

play38:02

are gonna get this side effect, that effect, They quantify everything in a very nerdy way,  

play38:07

and that way we can trust our biotech.  That's why ultimately AstraZeneca has  

play38:12

a pretty good reputation. Same thing  with airplanes, same thing with cars,  

play38:17

same thing with basically any tech that cause  harm except for AI where here in the US there's  

play38:24

basically no regulation. If Sam Altman wants GPT-5  tomorrow, he can, right? So I think the sooner we  

play38:35

switch over to treating AI the way we treat all  other powerful tech, the better. I think a very  

play38:43

common misconception is that somehow we have  to choose between quickly reaping all sorts  

play38:50

of benefits of AI on one hand, and on the other  hand avoiding going extinct. The fact is with

play39:00

99% of the things that the people I talk with are  excited about, which I think includes you with AI,  

play39:08

are things that are quite harmless. That have  nothing to do with building smarter than human  

play39:14

AI that we don't know how to control. We  can curate, we can help spread knowledge,  

play39:19

we can help make companies more efficient. We  can do great progress in science and medicine,  

play39:26

et cetera, et cetera. So if we put safety  standards in place that just end up slowing  

play39:32

down a little bit that that last percent the  stuff that we might lose control over, then we  

play39:36

can really have this age of abundance for a number  of years now where we can enjoy revolutions and  

play39:41

healthcare and education and so many other things  without having to lose sleep that this is gonna  

play39:48

all blow up on us. And then if we can get to the  point eventually where we can see that even more  

play39:54

powerful systems meet the safety standards,  great. If it takes a little longer, fine,  

play39:59

we're not in any great rush. We can have life  flourishing for billions of years if we get this  

play40:05

right. So there's no point in risking squandering  everything just to get it one year sooner.

play40:11

You're a big history nerd. What do you think is  most analogous to this in history? Is there any  

play40:19

historical inventions that we can learn  from that will behave quite similarly?

play40:25

Yeah, so in 1942, Enrico Fermi built the world's  first nuclear reactor in Chicago under a football  

play40:34

stadium. And when physicists found out about that,  they totally freaked out. Why? Was it because they  

play40:46

thought this reactor was really dangerous? No,  it was really small, low energy output. It was  

play40:51

because they realized that now we're only a few  years away from the bomb. And three years later,  

play40:56

Hiroshima, Nagasaki happened right? There's a nice  analogy there because around that time in 1951,  

play41:04

Alan Turing said that one day machines will become  as smart as people and then very quickly they'll  

play41:12

become way smarter than people because we're  biological computers and there's no reason  

play41:15

machines can't do much better. And then the  default is we lose control over the machines.  

play41:20

But I'll give you a little canary in the coal mine  warning so you know when you're close. The Turing  

play41:27

test, once machines become good enough  at language and knowledge that they can  

play41:32

fool a lot of people into thinking that  they are human, that's when you're close.  

play41:38

That's the Enrico for moment when you  might have a few years. And last year,  

play41:48

Joshua Bengio even one of the most sighted AI  researchers in the world argued that GPT-4 passes  

play41:52

the Turing test. You can squabble about whether  it's passed the Turing test or whether we'll  

play41:57

pass it next year. But we're roughly there at the  Enrico Fermi reactor now for AI where it's high  

play42:05

time to take it seriously. Big things are going to  happen soon and let's get it right. Let's prepare.

play42:16

So one final question. We're  we meet here at your campus,  

play42:21

you know, two decades from from now. What  do you think is the best case scenario?

play42:27

The best case is we're still alive and  we're happy and the humans are still in  

play42:33

charge of this planet. Except there's no more  famine, no more wars, the climate is all good,  

play42:42

and we have basically managed to solve  the biggest problems that have stumped  

play42:49

people throughout the ages. That's  the win case that I'm going for.

play42:54

And what do you think happens once we've solved  

play42:57

all problems? We've found treatments to all  diseases, what does that world look like?

play43:02

Well, Nick Bostrom just wrote a book on exactly  that, the solved world as he calls it. I'm not  

play43:09

spending too much time thinking about that  yet. I'd rather have the luxury of thinking  

play43:13

about that once we've actually gone a little  closer to solving it now. Right now there's  

play43:18

so much exciting stuff to be done. Both on  the nerd side, on figuring out how to make  

play43:23

systems the more trustworthy, and also on  the societal side of just making sure that  

play43:29

our politicians put good safety standards in  place. So that's where I'm spending my energy.

play43:34

So let's get to some of the easier questions.  What what do you believe is the meaning of life?

play43:43

The easy questions. You know, something we've discovered  

play43:51

by finding all these fundamental equations of  physics is that none of these equations have  

play43:59

any explicit mention of meaning put into them.  So I think it's rather up to us conscious living  

play44:09

beings to create our own meaning, frankly. I think  that consciousness and positive experiences is at  

play44:20

the heart of this because even beauty and love and  passion and kindness and hope are also conscious  

play44:28

experiences after all. I don't think my desk here  is having any of those experiences. And in other  

play44:35

words, I don't think that our universe gives  meaning to us. We give meaning to our universe.

Rate This

5.0 / 5 (0 votes)

Related Tags
Artificial IntelligenceNeural NetworksFuture PredictionsEthical ConcernsTechnological AdvancementEducational ImpactSelf-ImprovementAI SafetyHuman-AI InteractionInnovation History