How AI Will Step Off the Screen and into the Real World | Daniela Rus | TED

TED
19 Apr 202412:55

Summary

TLDRA robotics expert reflects on a key lesson learned during a student project—how the unpredictability of the physical world challenges digital systems. The speaker, now leading MIT's Computer Science and AI Lab, discusses the exciting future of combining AI with robotics. Introducing 'physical intelligence,' the fusion of AI's decision-making abilities with robots' physical capabilities, the talk explores how this breakthrough will transform industries, from designing robots using text prompts to creating adaptive AI. The speaker envisions a world where intelligent machines enhance human abilities, urging collaboration to shape the future of technology responsibly.

Takeaways

  • 🤖 Robotics requires careful planning to match the robot's capabilities with the physical world, as shown by the mishap with the cake-cutting robot.
  • 🎓 The speaker leads MIT's Computer Science and AI lab, where they're merging AI with robotics to bring intelligence into the physical world.
  • 🧠 AI and robotics are traditionally separate fields, but now, AI is moving from digital decision-making into real-world physical applications.
  • 🔧 Physical intelligence combines AI's decision-making with the mechanical prowess of robots to make machines smarter and more capable in the physical world.
  • 💡 Current AI systems are large, error-prone, and frozen after training, while 'liquid networks' offer a more efficient, adaptive, and understandable solution.
  • 🦠 Liquid networks are inspired by the simple neural structures of C. elegans, offering compact and explainable AI that continues to learn after deployment.
  • 🚗 Liquid networks lead to better performance in tasks like self-driving, where they focus on relevant features, unlike traditional AI, which can be easily confused.
  • 🔬 Text-to-robot and image-to-robot technology allows rapid prototyping, transforming digital designs into real-world machines with ease and precision.
  • 👨‍🍳 Robots can learn tasks from humans by collecting physical data from sensors, making them capable of mimicking human actions with grace and agility.
  • 🌍 The speaker emphasizes the potential of physical intelligence to amplify human capabilities, improve daily life, and help create a better future for humanity and the planet.

Q & A

  • What inspired the speaker's interest in robotics?

    -The speaker's interest in robotics began during their time as a student when they and a group of friends attempted to program a robot to cut a birthday cake for their professor.

  • What was the lesson learned from the cake-cutting robot incident?

    -The speaker learned that the physical world, with its laws of physics and imprecisions, is far more demanding than the digital world.

  • How does the speaker describe the current relationship between AI and robotics?

    -Currently, AI and robotics are largely separate fields. AI focuses on decision-making and learning in the digital realm, while robots are physical machines that can execute pre-programmed tasks but lack intelligence.

  • What is 'physical intelligence' according to the speaker?

    -'Physical intelligence' refers to the integration of AI's ability to understand data with the physical abilities of robots, enabling machines to think and interact intelligently in the real world.

  • What are some of the challenges in achieving physical intelligence?

    -Challenges include shrinking AI to run on smaller devices, such as robots, and ensuring that AI can adapt to physical environments without making mistakes.

  • How do liquid networks improve traditional AI models?

    -Liquid networks use fewer neurons with more complex mathematical functions, inspired by biological neurons like those in the C. elegans worm. These networks are adaptable and provide more explainable and focused decision-making compared to traditional AI.

  • What is the difference between traditional AI and liquid networks in adapting to real-world environments?

    -Traditional AI systems are fixed after training and cannot adapt, while liquid networks continue to learn and adapt based on the input they receive, even after deployment.

  • How does the speaker's lab turn text prompts into robots?

    -In the lab, a system starts with a language prompt, generates designs for a robot including its shape, materials, and control system, and then refines the design through simulation until it meets specifications.

  • What is the significance of human-to-robot learning in the speaker's research?

    -Human-to-robot learning allows machines to learn tasks from human behavior, such as food preparation and cleaning, by using physical data like muscle movements and gaze patterns.

  • What are the potential future applications of physical intelligence?

    -Physical intelligence could lead to personalized robots that assist in daily life, bespoke machines for work, and robots that can learn and perform tasks with agility and precision, extending human capabilities.

Outlines

00:00

🤖 A Cake-Cutting Robot Mishap and a Lesson in Robotics

The speaker shares a humorous anecdote about a project during their student days where a robot was programmed to cut a cake for their professor’s birthday. However, due to a lack of coordination, the robot, designed for a soft sponge cake, faced a square ice cream cake instead and failed spectacularly. Despite the failure, the professor appreciated the effort and referred to the incident as a ‘control singularity’—a robotics term. This experience taught the speaker that the physical world, with its unpredictable nature, poses far greater challenges than the digital realm. Now leading MIT's Computer Science and AI Lab, the speaker discusses how AI and robotics, previously separate fields, are converging, moving AI beyond digital screens and into the 3D physical world, thus enabling the next breakthrough—physical intelligence.

05:01

🚗 From Noisy AI to Streamlined Liquid Networks

This section contrasts traditional AI with liquid networks in the context of self-driving cars. The speaker highlights the inefficiencies of conventional AI, which relies on thousands of neurons that result in noisy, unfocused attention maps, causing the AI to misread its surroundings. In contrast, liquid networks, with only 19 neurons, provide a cleaner, more focused decision-making process. The speaker explains how traditional AI neurons are simple on/off units that struggle with complex real-world tasks, while liquid networks use differential equations inspired by the neural structure of worms, allowing for adaptability and continued learning after deployment. This innovation dramatically enhances AI's performance in dynamic environments, such as seasonal changes when navigating a forest.

10:03

🐰 Creating Robots from Text and Images

The speaker introduces the concept of transforming text and images into functional robots through a new AI-guided design process. By using physical constraints and simulations, robots can be designed from prompts like ‘make me a robot that can walk forward,’ and the system generates the required specifications for its construction. The speaker describes a specific example where an image of a bunny was turned into a physical, robotic version through 3D modeling and printing. This process dramatically speeds up prototyping and testing, leading to faster innovation cycles in robotics design and manufacturing.

👨‍🍳 Teaching Robots Human Tasks through Physical Intelligence

In this section, the speaker discusses how robots can be taught human tasks by observing physical data such as muscle movements, posture, and gaze. The speaker’s lab has created a kitchen environment to capture this data and train AI systems to replicate these tasks with robots. This approach allows machines to move with greater agility and grace, adapting to new situations. By utilizing liquid networks and physical intelligence, these machines can learn from human actions and become more versatile in completing tasks like cooking and cleaning, showcasing the far-reaching potential of physical intelligence in everyday life.

🌍 The Potential and Responsibility of Physical Intelligence

The speaker concludes by emphasizing the vast potential of physical intelligence, where AI-powered machines can transcend human limitations, assist with daily tasks, and innovate across industries. However, the speaker also underscores the importance of guiding AI responsibly, noting that current AI systems are unsustainable and prone to errors. By developing AI that interacts with the physical world and continues to learn, humanity can unlock unprecedented opportunities while ensuring a better future for both people and the planet. The speaker invites the audience to take part in shaping this future, either by developing, using, or inventing with physical intelligence.

Mindmap

Keywords

💡Physical intelligence

Physical intelligence refers to the integration of artificial intelligence (AI) with robotics to enable machines to interact with the physical world intelligently. In the video, it is a central theme, describing the shift from AI being confined to computers to it influencing real-world machines, allowing robots to perform tasks more effectively. The speaker explains that physical intelligence blends AI's data processing with the physical capabilities of robots, enabling them to make smarter decisions in real-world scenarios.

💡AI (Artificial Intelligence)

Artificial intelligence (AI) is the ability of machines to mimic human-like decision-making and learning. In the context of the video, the speaker highlights how AI is currently limited to digital systems, focusing on decision-making and learning without physical interaction. The talk discusses how AI is evolving to merge with robotics, allowing machines to not only make decisions but also act in the physical world, ushering in physical intelligence.

💡Liquid networks

Liquid networks are a new form of AI model that the speaker introduces as being more adaptable and efficient than traditional AI systems. Unlike typical neural networks, liquid networks continue to evolve and learn after training, adapting to new inputs. This innovation is inspired by the neural structure of the worm C. elegans and offers a more compact, explainable AI, essential for robots operating in dynamic, real-world environments.

💡Robotics

Robotics refers to the design and use of robots to perform tasks. In the video, the speaker discusses how robots traditionally perform pre-programmed tasks without intelligence. However, by combining AI with robotics, the field is moving toward physical intelligence, where robots can learn from humans and adapt to complex, real-world tasks like food preparation or object manipulation.

💡C. elegans

C. elegans is a type of worm with a simple nervous system, consisting of only 302 neurons, which biologists have mapped extensively. The speaker uses this organism as an analogy for liquid networks, explaining that AI systems can be made more efficient and explainable by using fewer neurons that perform more complex tasks, similar to the C. elegans's neural structure.

💡Control singularity

Control singularity is a technical robotics term referring to a point where a system becomes unstable or uncontrollable, as mentioned humorously in the story about the professor’s birthday cake. The robot's failure to cut the cake properly is described as a 'control singularity,' highlighting the challenges that arise when robotics systems are applied in unexpected real-world conditions.

💡Self-driving car

A self-driving car is a vehicle that uses AI to navigate without human intervention. In the video, the speaker contrasts traditional AI models, which rely on complex, noisy decision-making, with their lab’s liquid network solution. The liquid network AI performs better by focusing its attention on relevant elements like the road, demonstrating the improved performance of physical intelligence systems in real-world scenarios.

💡Differential equations

Differential equations are mathematical equations that describe how things change over time. In the video, the speaker explains that liquid networks use differential equations to model the behavior of neurons and synapses, allowing these AI systems to continue learning and adapting after deployment. This contrasts with traditional AI models, which are static after training.

💡Text-to-robot

Text-to-robot refers to the ability to design robots based on textual descriptions. In the video, the speaker demonstrates how AI can transform a simple command like 'make me a robot that can walk forward' into an actual physical machine. This approach drastically reduces the time and resources needed to create functional robots, combining language processing with physical intelligence.

💡Learning from humans

Learning from humans is the process by which robots acquire new skills by observing and mimicking human behavior. In the video, the speaker describes how robots are trained in their lab by monitoring humans performing tasks, such as food preparation. This data collection includes physical metrics like muscle movement and gaze, enabling robots to learn complex, dynamic tasks and move with greater agility and grace.

Highlights

A group of students attempted to program a robot to cut a cake for their professor's birthday, but the robot malfunctioned due to a mismatch between the soft cake they planned for and the hard ice cream cake they received.

Despite the failure, the professor praised the malfunction as a 'control singularity,' highlighting the unpredictable nature of robotics in real-world applications.

The speaker leads MIT's Computer Science and AI Lab, which focuses on combining AI with robotics to develop intelligent machines capable of operating in the physical world.

The current separation between AI and robotics is beginning to dissolve, as AI is transitioning from computer screens into the real world, paving the way for 'physical intelligence.'

'Physical intelligence' is defined as the combination of AI's ability to understand data with the mechanical execution of robots, allowing machines to perform tasks more intelligently.

For AI to integrate into the physical world, it must run on smaller, more efficient computers, enabling robots to carry out tasks based on real-time data.

The concept of 'liquid networks' is introduced as a new AI model inspired by the neurons of the worm C. elegans, making AI solutions more compact and understandable.

Traditional AI models rely on thousands of neurons to make decisions, but liquid networks use only 19 neurons, resulting in a cleaner, more focused decision-making process.

Liquid networks adapt to new data after training, unlike traditional AI systems which are frozen and cannot improve once deployed.

The speaker demonstrates how liquid networks outperform traditional AI in tasks such as identifying objects in different environments, like woods during summer and fall.

Liquid networks are not just limited to AI software; they are also being integrated into robotics, creating adaptable machines that can learn from humans.

The lab has developed systems that can design robots based on text or image prompts, drastically reducing the time required for prototyping and testing new machines.

The future of AI and robotics will involve creating machines that not only learn from humans but also perform tasks in physical spaces like food preparation and cleaning.

Physical intelligence enables machines to move with grace and agility, adapting to real-world situations and interacting with humans more seamlessly.

AI's future lies in its ability to extend human capabilities, amplify strengths, and refine precision, allowing us to build advanced tools that improve our interaction with the world.

The speaker calls for human guidance over AI development, emphasizing the responsibility we have to ensure technology benefits humanity and the planet.

Transcripts

play00:04

When I was a student studying robotics,

play00:06

a group of us decided to make a present for our professor's birthday.

play00:11

We wanted to program our robot to cut a slice of cake for him.

play00:16

We pulled an all-nighter writing the software,

play00:19

and the next day, disaster.

play00:22

We programmed this robot to cut a soft, round sponge cake,

play00:27

but we didn't coordinate well.

play00:29

And instead, we received a square hard ice cream cake.

play00:34

The robot flailed wildly and nearly destroyed the cake.

play00:38

(Laughter)

play00:39

Our professor was delighted, anyway.

play00:41

He calmly pushed the stop button

play00:45

and declared the erratic behavior of the robot

play00:48

a control singularity.

play00:50

A robotics technical term.

play00:52

I was disappointed, but I learned a very important lesson.

play00:56

The physical world,

play00:58

with its physics laws and imprecisions,

play01:01

is a far more demanding space than the digital world.

play01:05

Today, I lead MIT's Computer Science and AI lab,

play01:09

the largest research unit at MIT.

play01:11

This is our building, where I work with brilliant and brave researchers

play01:16

to invent the future of computing and intelligent machines.

play01:21

Today in computing,

play01:22

artificial intelligence and robotics are largely separate fields.

play01:27

AI has amazed you with its decision-making and learning,

play01:31

but it remains confined inside computers.

play01:34

Robots have a physical presence and can execute pre-programmed tasks,

play01:39

but they're not intelligent.

play01:42

Well, this separation is starting to change.

play01:45

AI is about to break free from the 2D computer screen interactions

play01:50

and enter a vibrant, physical 3D world.

play01:54

In my lab, we're fusing the digital intelligence of AI

play01:58

with the mechanical prowess of robots.

play02:01

Moving AI from the digital world into the physical world

play02:03

is making machines intelligent

play02:06

and leading to the next great breakthrough,

play02:08

what I call physical intelligence.

play02:11

Physical intelligence is when AI's power to understand text,

play02:16

images and other online information

play02:18

is used to make real-world machines smarter.

play02:21

This means AI can help pre-programmed robots do their tasks better

play02:27

by using knowledge from data.

play02:31

With physical intelligence,

play02:32

AI doesn't just reside in our computers,

play02:37

but walks, rolls, flies

play02:39

and interacts with us in surprising ways.

play02:42

Imagine being surrounded by helpful robots at the supermarket.

play02:47

The one on the left can help you carry a heavy box.

play02:51

To make it happen, we need to do a few things.

play02:54

We need to rethink how machines think.

play02:57

We need to reorganize how they are designed and how they learn.

play03:03

So for physical intelligence,

play03:05

AI has to run on computers that fit on the body of the robot.

play03:09

For example, our soft robot fish.

play03:13

Today's AI uses server farms that do not fit.

play03:17

Today's AI also makes mistakes.

play03:20

This AI system on a robot car does not detect pedestrians.

play03:25

For physical intelligence,

play03:27

we need small brains that do not make mistakes.

play03:31

We're tackling these challenges using inspiration

play03:34

from a worm called C. elegans

play03:37

In sharp contrast to the billions of neurons in the human brain,

play03:42

C. elegans has a happy life on only 302 neurons,

play03:47

and biologists understand the math of what each of these neurons do.

play03:53

So here's the idea.

play03:54

Can we build AI using inspiration from the math of these neurons?

play04:01

We have developed, together with my collaborators and students,

play04:05

a new approach to AI we call “liquid networks.”

play04:10

And liquid networks results in much more compact

play04:13

and explainable solutions than today's traditional AI solutions.

play04:17

Let me show you.

play04:19

This is our self-driving car.

play04:21

It's trained using a traditional AI solution,

play04:24

the kind you find in many applications today.

play04:28

This is the dashboard of the car.

play04:30

In the lower right corner, you'll see the map.

play04:32

In the upper left corner, the camera input stream.

play04:35

And the big box in the middle with the blinking lights

play04:38

is the decision-making engine.

play04:40

It consists of tens of thousands of artificial neurons,

play04:44

and it decides how the car should steer.

play04:48

It is impossible to correlate the activity of these neurons

play04:51

with the behavior of the car.

play04:53

Moreover, if you look at the lower left side,

play04:57

you see where in the image this decision-making engine looks

play05:01

to tell the car what to do.

play05:03

And you see how noisy it is.

play05:04

And this car drives by looking at the bushes and the trees

play05:09

on the side of the road.

play05:10

That's not how we drive.

play05:11

People look at the road.

play05:13

Now contrast this with our liquid network solution,

play05:16

which consists of only 19 neurons rather than tens of thousands.

play05:21

And look at its attention map.

play05:23

It's so clean and focused on the road horizon

play05:26

and the side of the road.

play05:28

Because these models are so much smaller,

play05:30

we actually understand how they make decisions.

play05:34

So how did we get this performance?

play05:38

Well, in a traditional AI system,

play05:41

the computational neuron is the artificial neuron,

play05:44

and the artificial neuron is essentially an on/off computational unit.

play05:48

It takes in some numbers, adds them up,

play05:50

applies some basic math

play05:52

and passes along the result.

play05:54

And this is complex

play05:55

because it happens across thousands of computational units.

play05:59

In liquid networks,

play06:01

we have fewer neurons,

play06:02

but each one does more complex math.

play06:05

Here's what happens inside our liquid neuron.

play06:08

We use differential equations to model the neural computation

play06:12

and the artificial synapse.

play06:14

And these differential equations

play06:16

are what biologists have mapped for the neural structure of the worms.

play06:22

We also wire the neurons differently to increase the information flow.

play06:27

Well, these changes yield phenomenal results.

play06:31

Traditional AI systems are frozen after training.

play06:34

That means they cannot continue to improve

play06:36

when we deploy them in a physical world in the wild.

play06:40

We just wait for the next release.

play06:43

Because of what's happening inside the liquid neuron,

play06:46

liquid networks continue to adapt after training

play06:49

based on the inputs that they see.

play06:51

Let me show you.

play06:53

We trained traditional AI and liquid networks

play06:56

using summertime videos like these ones,

play06:59

and the task was to find things in the woods.

play07:02

All the models learned how to do the task in the summer.

play07:06

Then we tried to use the models on drones in the fall.

play07:10

The traditional AI solution gets confused by the background.

play07:14

Look at the attention map, cannot do the task.

play07:17

Liquid networks do not get confused by the background

play07:20

and very successfully execute the task.

play07:24

So this is it.

play07:26

This is the step forward:

play07:27

AI that adapts after training.

play07:31

Liquid networks are important

play07:33

because they give us a new way of getting machines to think

play07:38

that is rooted into physics models,

play07:40

a new technology for AI.

play07:43

We can run them on smartphones, on robots,

play07:46

on enterprise computers,

play07:48

and even on new types of machines

play07:50

that we can now begin to imagine and design.

play07:53

The second aspect of physical intelligence.

play07:56

So by now you've probably generated images using text-to-image systems.

play08:02

We can also do text-to-robot,

play08:04

but not using today's AI solutions because they work on statistics

play08:08

and do not understand physics.

play08:11

In my lab,

play08:12

we developed an approach that guides the design process

play08:16

by checking and simulating the physical constraints for the machine.

play08:21

We start with a language prompt,

play08:23

"Make me a robot that can walk forward,"

play08:26

and our system generates the designs including shape, materials, actuators,

play08:32

sensors, the program to control it

play08:35

and the fabrication files to make it.

play08:37

And then the designs get refined in simulation

play08:41

until they meet the specifications.

play08:44

So in a few hours we can go from idea

play08:48

to controllable physical machine.

play08:51

We can also do image-to-robot.

play08:53

This photo can be transformed into a cuddly robotic bunny.

play08:58

To do so, our algorithm computes a 3D representation of the photo

play09:03

that gets sliced and folded, printed.

play09:08

Then we fold the printed layers, we string some motors and sensors.

play09:12

We write some code, and we get the bunny you see in this video.

play09:16

We can use this approach to make anything almost,

play09:20

from an image, from a photo.

play09:23

So the ability to transform text into images

play09:28

and to transform images into robots is important,

play09:31

because we are drastically reducing the amount of time

play09:35

and the resources needed to prototype and test new products,

play09:39

and this is allowing for a much faster innovation cycle.

play09:45

And now we are ready to even make the leap

play09:48

to get these machines to learn.

play09:50

The third aspect of physical intelligence.

play09:54

These machines can learn from humans how to do tasks.

play09:57

You can think of it as human-to-robot.

play09:59

In my lab, we created a kitchen environment

play10:02

where we instrument people with sensors,

play10:05

and we collect a lot of data about how people do kitchen tasks.

play10:09

We need physical data

play10:11

because videos do not capture the dynamics of the task.

play10:15

So we collect muscle, pose, even gaze information

play10:18

about how people do tasks.

play10:21

And then we train AI using this data

play10:24

to teach robots how to do the same tasks.

play10:28

And the end result is machines that move with grace and agility,

play10:34

as well as adapt and learn.

play10:36

Physical intelligence.

play10:39

We can use this approach to teach robots

play10:42

how to do a wide range of tasks:

play10:44

food preparation, cleaning and so much more.

play10:49

The ability to turn images and text into functional machines,

play10:54

coupled with using liquid networks

play10:56

to create powerful brains for these machines

play10:59

that can learn from humans, is incredibly exciting.

play11:02

Because this means we can make almost anything we imagine.

play11:07

Today's AI has a ceiling.

play11:09

It requires server farms.

play11:11

It's not sustainable.

play11:12

It makes inexplicable mistakes.

play11:15

Let's not settle for the current offering.

play11:18

When AI moves into the physical world,

play11:20

the opportunities for benefits and for breakthroughs is extraordinary.

play11:26

You can get personal assistants that optimize your routines

play11:31

and anticipate your needs,

play11:33

bespoke machines that help you at work

play11:36

and robots that delight you in your spare time.

play11:40

The promise of physical intelligence is to transcend our human limitations

play11:45

with capabilities that extend our reach,

play11:48

amplify our strengths

play11:50

and refine our precision

play11:52

and grant us ways to interact with the world

play11:55

we've only dreamed of.

play11:58

We are the only species so advanced, so aware,

play12:02

so capable of building these extraordinary tools.

play12:06

Yet, developing physical intelligence

play12:09

is teaching us that we have so much more to learn

play12:12

about technology and about ourselves.

play12:16

We need human guiding hands over AI sooner rather than later.

play12:20

After all, we remain responsible for this planet

play12:23

and everything living on it.

play12:26

I remain convinced that we have the power

play12:29

to use physical intelligence to ensure a better future for humanity

play12:34

and for the planet.

play12:36

And I'd like to invite you to help us in this quest.

play12:39

Some of you will help develop physical intelligence.

play12:43

Some of you will use it.

play12:45

And some of you will invent the future.

play12:48

Thank you.

play12:49

(Applause)

Rate This

5.0 / 5 (0 votes)

Étiquettes Connexes
AI roboticsphysical intelligencemachine learningliquid networksautomationMIT researchhuman-robotfuture techrobot designsmart machines
Besoin d'un résumé en anglais ?