How computers are learning to be creative | Blaise Agüera y Arcas

TED
22 Jul 201617:35

Summary

TLDRThis transcript outlines a presentation on machine intelligence and its connection to human brain functions, particularly in perception and creativity. The speaker delves into the history of neuroscience and machine learning, drawing parallels between neural networks and human cognition. By explaining how computers can now identify images and generate art through machine learning, the speaker highlights the shared mechanisms between perception and creation. The talk touches on the progress in AI, from early experiments to modern neural networks, showing how machines can both recognize and create, much like the human brain.

Takeaways

  • 🧠 The speaker leads a team at Google focused on machine intelligence, aiming to make computers capable of tasks similar to brain functions.
  • 🔍 Perception and creativity are central to both human intelligence and machine learning; perception turns sensory data into concepts, and creativity turns concepts into tangible outputs.
  • 🖼️ The team’s machine perception algorithms allow images to become searchable, as seen in Google Photos.
  • 🧑‍🎨 The speaker emphasizes the connection between perception and creativity, citing Michelangelo's insight that creativity is about discovery through perception.
  • 🧬 Santiago Ramón y Cajal’s groundbreaking 19th-century neuroanatomy work is still influential in understanding brain structures and neurons today.
  • 🖥️ Computers, modeled after the brain's neural networks, are now capable of recognizing images and even generating new ones using machine learning.
  • 🦅 Neural networks can recognize patterns in images and create artistic renditions, as seen in experiments where machines generate images of birds or morphing animals.
  • 🔢 Learning in neural networks mimics human learning, where the system iteratively reduces error by adjusting its internal weights based on training data.
  • 🎨 The speaker demonstrates how machine-generated images resemble creative works, with neural networks producing surreal, abstract forms based on familiar objects.
  • 💻 The talk concludes with the idea that computers, modeled after human brains, are helping us not only extend our intelligence but also gain a deeper understanding of how our minds work.

Q & A

  • What is the main focus of the team at Google that the speaker leads?

    -The team focuses on machine intelligence, specifically making computers and devices able to perform tasks that brains can do, such as perception and creativity.

  • Why is the team interested in studying real brains and neuroscience?

    -The team is interested in understanding how brains outperform computers in certain tasks, like perception, to enhance machine perception and intelligence systems.

  • What role does perception play in machine intelligence, according to the speaker?

    -Perception is the process by which sensory input, like sounds and images, is converted into concepts in the mind. In machine intelligence, it enables systems like Google Photos to recognize and make images searchable.

  • How does the speaker connect perception with creativity?

    -The speaker suggests that perception and creativity are closely linked because creating often involves perceiving, as exemplified by Michelangelo’s idea that a sculptor discovers a statue within a block of stone.

  • Who was Santiago Ramón y Cajal, and what was his contribution to neuroscience?

    -Santiago Ramón y Cajal was a 19th-century Spanish neuroanatomist who pioneered the study of brain cells (neurons) using microscopy and specialized staining techniques, leading to detailed drawings of neurons.

  • How did early computer scientists model the brain in their research?

    -Early computer scientists like Warren McCulloch and Walter Pitts modeled the brain’s visual cortex as a circuit diagram, viewing it as a series of computational elements that process information similarly to how computers work.

  • What challenge does the speaker highlight when it comes to machine perception?

    -The challenge for machine perception is to take an image and accurately classify it, such as recognizing a bird in a picture. This task, which is simple for the human brain, was once nearly impossible for computers.

  • What is the significance of 'x,' 'w,' and 'y' in the neural network model described by the speaker?

    -'x' represents the input pixels, 'w' the synaptic weights in the neural network, and 'y' the output, such as the classification of an object (e.g., 'bird'). These variables form the core of the neural network's computation.

  • How does the neural network ‘learn’ according to the speaker?

    -The neural network learns by minimizing error through iterative adjustments of the weights ('w'). This process mimics how humans learn by refining their understanding through trial and error based on feedback.

  • What does the speaker mean by solving for 'x' in the context of machine creativity?

    -Solving for 'x' means using a trained neural network to generate images based on its learned weights ('w') and a known concept ('y'). This process allows the machine to create visual outputs, like a picture of a bird, by reconstructing what it has learned about that concept.

Outlines

00:00

🤖 Machine Perception and Creativity

The speaker leads a team at Google working on machine intelligence, focusing on making computers capable of brain-like tasks. They emphasize that while computers have become proficient in perception, such as turning images into searchable content, creativity—transforming concepts into reality—remains a significant area of interest. Machine perception has evolved to the point where it unexpectedly connects to creativity, giving rise to machine art. The speaker references Michelangelo's quote about sculpture to illustrate how perception and creativity are interconnected. Furthermore, they highlight that studying brains and understanding perception enhances machine learning and vice versa, marking a significant crossover between neuroscience and artificial intelligence.

05:01

💡 The Historical Evolution of Brain Study

The speaker delves into the history of neuroscience, beginning with early brain anatomists and their superficial understanding of the brain's structure. The breakthrough came with Santiago Ramón y Cajal, a 19th-century neuroanatomist who used staining techniques to visualize individual brain cells. His detailed drawings of neurons were groundbreaking and still hold significance today. The speaker emphasizes the similarities between neurons and electrical wiring, noting how Ramón y Cajal's work laid the foundation for modern neuroscience, where scientists continue to explore and map the brain using advanced techniques like serial electron microscopy.

10:03

👀 Visual Processing and Neural Networks

The speaker transitions to discussing how brains and computers process visual information. They reference early computer scientists, like Alan Turing and Warren McCulloch, who saw brain structures as akin to circuit diagrams. By explaining the basic task of perception—recognizing a bird in an image—the speaker introduces neural networks, both biological and artificial, which handle complex layers of information to arrive at an output. They describe how modern neural networks can now perform tasks previously impossible for computers, such as identifying birds and even species from images. The discussion illustrates how perception works in both human and machine learning models.

15:03

🧠 Learning and Error Minimization in Neural Networks

The speaker dives into how machines learn, particularly through a process of error minimization. They explain that while it's relatively easy to solve for known variables in a simple equation, the challenge lies in figuring out the unknowns, such as the weights (w) in neural networks. By using error minimization, machines take guesses, adjust, and iteratively improve their predictions, similar to how humans learn through experience. The speaker illustrates this process with an example of a neural network recognizing birds on a mobile phone, showcasing how modern technology can perform trillions of operations to identify objects in real-time.

🔄 Reverse Engineering Neural Networks for Creativity

The speaker introduces the concept of reverse-engineering neural networks to generate images instead of recognizing them. By solving for the input (x) rather than the output (y), neural networks trained to recognize birds can generate images of birds. The speaker shares examples of neural networks creating art, including an 'Animal Parade' video morphing from one animal to another, highlighting how machine learning can produce creative outcomes. The speaker also demonstrates how adjusting parameters in a neural network can produce new, surreal images, further merging the worlds of machine learning and creativity.

🎨 Machine Art and Perception: Hallucination and Creativity

The speaker discusses how machine learning models can generate images by 'hallucinating' or interpreting random noise or existing images. They describe how neural networks, when left to iterate over images like clouds, begin to perceive recognizable objects, similar to how humans see patterns in ambiguous visuals. The speaker highlights experiments in which networks are pushed into a sort of 'free association,' creating continuous loops of generated images. This capability, the speaker notes, demonstrates how machines, like humans, use perception to fuel creativity.

🎭 Expanding Creativity Beyond Visual Art

The speaker concludes by pointing out that machine learning isn't limited to visual art. They mention an experiment where a neural network generates poetry based on images, highlighting a collaboration with artist Ross Goodwin. The poetry, informed by 20th-century works, demonstrates that machines can extend their creative capabilities to other forms of expression. The speaker reiterates Michelangelo's idea that perception and creativity are deeply linked, suggesting that machines, like human brains, can perform both tasks, thus expanding the horizons of artificial intelligence in creative fields.

Mindmap

Keywords

💡Machine Intelligence

Machine intelligence refers to the capability of computers and devices to replicate some of the functions that human brains perform, such as processing, analyzing, and interpreting information. In the video, the speaker discusses how his team at Google focuses on developing systems that improve machine perception and creativity, making devices more brain-like in their operations.

💡Perception

Perception is the process by which the brain interprets sensory information, such as sounds and images, turning them into concepts. In the video, it is highlighted as an area where human brains are traditionally superior to machines. The speaker explains how their team works on algorithms that enable computers to process and understand visual information, such as identifying objects in images.

💡Creativity

Creativity is the ability to transform concepts or ideas into something new, often seen in artistic or problem-solving contexts. The speaker relates it to machine learning by discussing how neural networks, initially designed for perception, can also be used to generate creative outputs, like images or poetry, by reversing the perception process.

💡Neural Networks

Neural networks are computational models inspired by the structure of the brain's neurons, designed to recognize patterns and process data through interconnected nodes. The video explains how these networks are used to process visual data in layers, simulating how the human brain interprets images, such as identifying a bird in a photo.

💡Synapses

Synapses are the connections between neurons that transmit signals in the brain. In the context of machine learning, synapses are modeled as 'weights' between nodes in a neural network, which are adjusted during the learning process. The speaker describes how the strength of these connections determines the network's ability to correctly identify objects, such as a bird in an image.

💡Santiago Ramón y Cajal

Santiago Ramón y Cajal was a 19th-century neuroanatomist known for his pioneering work in neuroscience, particularly in mapping out the structure of neurons. The speaker references his work to illustrate how early discoveries about brain cell structures laid the foundation for modern understanding of neural networks, both biological and artificial.

💡Error Minimization

Error minimization refers to the process of gradually improving a model’s accuracy by reducing the difference between its predicted output and the actual result. The speaker describes how neural networks are trained by making guesses and adjusting their weights to minimize errors, enabling the model to 'learn' and improve over time.

💡Inference

Inference is the process of drawing conclusions from known information. In machine learning, it refers to the ability of a model to make predictions or recognize patterns based on input data. The speaker explains that the goal of perception in neural networks is to infer the correct label (e.g., identifying a bird) from the input data (the image).

💡3D Reconstructions

3D reconstructions refer to the process of creating a three-dimensional model from two-dimensional data, such as slices of brain tissue. In the video, the speaker talks about how modern neuroscience uses 3D reconstructions of neurons to visualize the brain’s structure, which was pioneered by early scientists like Ramón y Cajal.

💡Artificial Neural Networks

Artificial neural networks (ANNs) are computational systems modeled after biological neural networks in the brain, designed to process data in a way similar to human cognition. The speaker describes how these networks are used to recognize patterns in images, such as identifying species of birds, and how they can also be used to generate creative outputs.

Highlights

The team at Google is working on machine intelligence to make computers perform brain-like functions, especially focusing on perception and creativity.

Machine perception is crucial for both human brains and computers, enabling tasks like image recognition in Google Photos.

Creativity, the ability to turn concepts into reality, is closely linked with perception, as Michelangelo’s famous quote on sculpture suggests.

Santiago Ramón y Cajal's groundbreaking 19th-century work on neuroanatomy laid the foundation for modern neuroscience, helping us understand brain cells (neurons).

Modern technologies, like those from the Max Planck Institute, allow us to image tiny brain tissue, advancing our understanding of neurons and their connections.

Neurons use electricity, which led to groundbreaking studies in the mid-20th century on how neurons communicate, forming the basis for artificial neural networks.

Warren McCulloch and Walter Pitts's work on visual cortex laid the groundwork for understanding how neurons process images in a computational manner.

The brain's process of image recognition is mirrored in artificial neural networks, where pixel data (input) is processed layer by layer to identify objects (output).

Training neural networks to recognize images involves solving for the unknown variables (synapse weights), similar to how humans learn through trial and error.

Artificial neural networks can now generate images based on trained data, such as creating bird images by solving for input variables in the network.

‘Animal Parade’ is an example of machine creativity, generating morphing images of animals using neural networks trained to recognize them.

Neural networks can generate surreal images of faces by reconstructing them based on trained data, highlighting the challenges of ambiguity in perception.

Machines can perceive and create; generating images or concepts from neural networks shows the link between recognition and creation, a process similar to human imagination.

Computer models of perception and creativity suggest that these processes aren't unique to humans, as computers are computational models of the mind.

The development of computing, initially inspired by the human brain, is now fulfilling the early visions of pioneers like Alan Turing and John von Neumann by extending human cognitive abilities.

Transcripts

play00:12

So, I lead a team at Google that works on machine intelligence;

play00:15

in other words, the engineering discipline of making computers and devices

play00:20

able to do some of the things that brains do.

play00:23

And this makes us interested in real brains

play00:26

and neuroscience as well,

play00:27

and especially interested in the things that our brains do

play00:32

that are still far superior to the performance of computers.

play00:37

Historically, one of those areas has been perception,

play00:40

the process by which things out there in the world --

play00:43

sounds and images --

play00:45

can turn into concepts in the mind.

play00:48

This is essential for our own brains,

play00:50

and it's also pretty useful on a computer.

play00:53

The machine perception algorithms, for example, that our team makes,

play00:57

are what enable your pictures on Google Photos to become searchable,

play01:00

based on what's in them.

play01:03

The flip side of perception is creativity:

play01:07

turning a concept into something out there into the world.

play01:10

So over the past year, our work on machine perception

play01:13

has also unexpectedly connected with the world of machine creativity

play01:18

and machine art.

play01:20

I think Michelangelo had a penetrating insight

play01:23

into to this dual relationship between perception and creativity.

play01:28

This is a famous quote of his:

play01:30

"Every block of stone has a statue inside of it,

play01:34

and the job of the sculptor is to discover it."

play01:38

So I think that what Michelangelo was getting at

play01:41

is that we create by perceiving,

play01:44

and that perception itself is an act of imagination

play01:47

and is the stuff of creativity.

play01:50

The organ that does all the thinking and perceiving and imagining,

play01:54

of course, is the brain.

play01:57

And I'd like to begin with a brief bit of history

play01:59

about what we know about brains.

play02:02

Because unlike, say, the heart or the intestines,

play02:04

you really can't say very much about a brain by just looking at it,

play02:08

at least with the naked eye.

play02:09

The early anatomists who looked at brains

play02:12

gave the superficial structures of this thing all kinds of fanciful names,

play02:16

like hippocampus, meaning "little shrimp."

play02:18

But of course that sort of thing doesn't tell us very much

play02:21

about what's actually going on inside.

play02:24

The first person who, I think, really developed some kind of insight

play02:28

into what was going on in the brain

play02:30

was the great Spanish neuroanatomist, Santiago Ramón y Cajal,

play02:34

in the 19th century,

play02:35

who used microscopy and special stains

play02:39

that could selectively fill in or render in very high contrast

play02:43

the individual cells in the brain,

play02:45

in order to start to understand their morphologies.

play02:49

And these are the kinds of drawings that he made of neurons

play02:52

in the 19th century.

play02:54

This is from a bird brain.

play02:56

And you see this incredible variety of different sorts of cells,

play02:59

even the cellular theory itself was quite new at this point.

play03:02

And these structures,

play03:03

these cells that have these arborizations,

play03:06

these branches that can go very, very long distances --

play03:08

this was very novel at the time.

play03:10

They're reminiscent, of course, of wires.

play03:13

That might have been obvious to some people in the 19th century;

play03:17

the revolutions of wiring and electricity were just getting underway.

play03:21

But in many ways,

play03:23

these microanatomical drawings of Ramón y Cajal's, like this one,

play03:26

they're still in some ways unsurpassed.

play03:28

We're still more than a century later,

play03:30

trying to finish the job that Ramón y Cajal started.

play03:33

These are raw data from our collaborators

play03:36

at the Max Planck Institute of Neuroscience.

play03:39

And what our collaborators have done

play03:41

is to image little pieces of brain tissue.

play03:46

The entire sample here is about one cubic millimeter in size,

play03:49

and I'm showing you a very, very small piece of it here.

play03:52

That bar on the left is about one micron.

play03:54

The structures you see are mitochondria

play03:57

that are the size of bacteria.

play03:59

And these are consecutive slices

play04:00

through this very, very tiny block of tissue.

play04:04

Just for comparison's sake,

play04:06

the diameter of an average strand of hair is about 100 microns.

play04:10

So we're looking at something much, much smaller

play04:12

than a single strand of hair.

play04:14

And from these kinds of serial electron microscopy slices,

play04:18

one can start to make reconstructions in 3D of neurons that look like these.

play04:23

So these are sort of in the same style as Ramón y Cajal.

play04:26

Only a few neurons lit up,

play04:27

because otherwise we wouldn't be able to see anything here.

play04:30

It would be so crowded,

play04:31

so full of structure,

play04:33

of wiring all connecting one neuron to another.

play04:37

So Ramón y Cajal was a little bit ahead of his time,

play04:40

and progress on understanding the brain

play04:42

proceeded slowly over the next few decades.

play04:45

But we knew that neurons used electricity,

play04:48

and by World War II, our technology was advanced enough

play04:51

to start doing real electrical experiments on live neurons

play04:54

to better understand how they worked.

play04:56

This was the very same time when computers were being invented,

play05:01

very much based on the idea of modeling the brain --

play05:04

of "intelligent machinery," as Alan Turing called it,

play05:07

one of the fathers of computer science.

play05:09

Warren McCulloch and Walter Pitts looked at Ramón y Cajal's drawing

play05:14

of visual cortex,

play05:15

which I'm showing here.

play05:17

This is the cortex that processes imagery that comes from the eye.

play05:22

And for them, this looked like a circuit diagram.

play05:26

So there are a lot of details in McCulloch and Pitts's circuit diagram

play05:30

that are not quite right.

play05:31

But this basic idea

play05:32

that visual cortex works like a series of computational elements

play05:36

that pass information one to the next in a cascade,

play05:39

is essentially correct.

play05:41

Let's talk for a moment

play05:43

about what a model for processing visual information would need to do.

play05:48

The basic task of perception

play05:50

is to take an image like this one and say,

play05:55

"That's a bird,"

play05:56

which is a very simple thing for us to do with our brains.

play05:59

But you should all understand that for a computer,

play06:02

this was pretty much impossible just a few years ago.

play06:05

The classical computing paradigm

play06:07

is not one in which this task is easy to do.

play06:11

So what's going on between the pixels,

play06:13

between the image of the bird and the word "bird,"

play06:17

is essentially a set of neurons connected to each other

play06:20

in a neural network,

play06:22

as I'm diagramming here.

play06:23

This neural network could be biological, inside our visual cortices,

play06:26

or, nowadays, we start to have the capability

play06:28

to model such neural networks on the computer.

play06:31

And I'll show you what that actually looks like.

play06:34

So the pixels you can think about as a first layer of neurons,

play06:37

and that's, in fact, how it works in the eye --

play06:39

that's the neurons in the retina.

play06:41

And those feed forward

play06:43

into one layer after another layer, after another layer of neurons,

play06:46

all connected by synapses of different weights.

play06:49

The behavior of this network

play06:50

is characterized by the strengths of all of those synapses.

play06:54

Those characterize the computational properties of this network.

play06:57

And at the end of the day,

play06:59

you have a neuron or a small group of neurons

play07:01

that light up, saying, "bird."

play07:03

Now I'm going to represent those three things --

play07:06

the input pixels and the synapses in the neural network,

play07:11

and bird, the output --

play07:13

by three variables: x, w and y.

play07:16

There are maybe a million or so x's --

play07:18

a million pixels in that image.

play07:20

There are billions or trillions of w's,

play07:23

which represent the weights of all these synapses in the neural network.

play07:26

And there's a very small number of y's,

play07:28

of outputs that that network has.

play07:30

"Bird" is only four letters, right?

play07:33

So let's pretend that this is just a simple formula,

play07:36

x "x" w = y.

play07:38

I'm putting the times in scare quotes

play07:40

because what's really going on there, of course,

play07:43

is a very complicated series of mathematical operations.

play07:47

That's one equation.

play07:48

There are three variables.

play07:50

And we all know that if you have one equation,

play07:52

you can solve one variable by knowing the other two things.

play07:57

So the problem of inference,

play08:00

that is, figuring out that the picture of a bird is a bird,

play08:03

is this one:

play08:04

it's where y is the unknown and w and x are known.

play08:08

You know the neural network, you know the pixels.

play08:10

As you can see, that's actually a relatively straightforward problem.

play08:14

You multiply two times three and you're done.

play08:16

I'll show you an artificial neural network

play08:19

that we've built recently, doing exactly that.

play08:21

This is running in real time on a mobile phone,

play08:24

and that's, of course, amazing in its own right,

play08:27

that mobile phones can do so many billions and trillions of operations

play08:31

per second.

play08:32

What you're looking at is a phone

play08:34

looking at one after another picture of a bird,

play08:37

and actually not only saying, "Yes, it's a bird,"

play08:40

but identifying the species of bird with a network of this sort.

play08:44

So in that picture,

play08:46

the x and the w are known, and the y is the unknown.

play08:50

I'm glossing over the very difficult part, of course,

play08:53

which is how on earth do we figure out the w,

play08:56

the brain that can do such a thing?

play08:59

How would we ever learn such a model?

play09:01

So this process of learning, of solving for w,

play09:04

if we were doing this with the simple equation

play09:07

in which we think about these as numbers,

play09:09

we know exactly how to do that: 6 = 2 x w,

play09:12

well, we divide by two and we're done.

play09:16

The problem is with this operator.

play09:18

So, division --

play09:19

we've used division because it's the inverse to multiplication,

play09:23

but as I've just said,

play09:24

the multiplication is a bit of a lie here.

play09:27

This is a very, very complicated, very non-linear operation;

play09:30

it has no inverse.

play09:32

So we have to figure out a way to solve the equation

play09:35

without a division operator.

play09:37

And the way to do that is fairly straightforward.

play09:39

You just say, let's play a little algebra trick,

play09:42

and move the six over to the right-hand side of the equation.

play09:45

Now, we're still using multiplication.

play09:47

And that zero -- let's think about it as an error.

play09:51

In other words, if we've solved for w the right way,

play09:53

then the error will be zero.

play09:55

And if we haven't gotten it quite right,

play09:57

the error will be greater than zero.

play09:59

So now we can just take guesses to minimize the error,

play10:02

and that's the sort of thing computers are very good at.

play10:05

So you've taken an initial guess:

play10:06

what if w = 0?

play10:08

Well, then the error is 6.

play10:09

What if w = 1? The error is 4.

play10:10

And then the computer can sort of play Marco Polo,

play10:13

and drive down the error close to zero.

play10:15

As it does that, it's getting successive approximations to w.

play10:19

Typically, it never quite gets there, but after about a dozen steps,

play10:22

we're up to w = 2.999, which is close enough.

play10:28

And this is the learning process.

play10:30

So remember that what's been going on here

play10:32

is that we've been taking a lot of known x's and known y's

play10:37

and solving for the w in the middle through an iterative process.

play10:40

It's exactly the same way that we do our own learning.

play10:44

We have many, many images as babies

play10:46

and we get told, "This is a bird; this is not a bird."

play10:49

And over time, through iteration,

play10:51

we solve for w, we solve for those neural connections.

play10:55

So now, we've held x and w fixed to solve for y;

play10:59

that's everyday, fast perception.

play11:01

We figure out how we can solve for w,

play11:03

that's learning, which is a lot harder,

play11:05

because we need to do error minimization,

play11:07

using a lot of training examples.

play11:08

And about a year ago, Alex Mordvintsev, on our team,

play11:12

decided to experiment with what happens if we try solving for x,

play11:15

given a known w and a known y.

play11:18

In other words,

play11:19

you know that it's a bird,

play11:20

and you already have your neural network that you've trained on birds,

play11:24

but what is the picture of a bird?

play11:27

It turns out that by using exactly the same error-minimization procedure,

play11:32

one can do that with the network trained to recognize birds,

play11:35

and the result turns out to be ...

play11:42

a picture of birds.

play11:44

So this is a picture of birds generated entirely by a neural network

play11:48

that was trained to recognize birds,

play11:50

just by solving for x rather than solving for y,

play11:53

and doing that iteratively.

play11:55

Here's another fun example.

play11:57

This was a work made by Mike Tyka in our group,

play12:01

which he calls "Animal Parade."

play12:03

It reminds me a little bit of William Kentridge's artworks,

play12:06

in which he makes sketches, rubs them out,

play12:08

makes sketches, rubs them out,

play12:10

and creates a movie this way.

play12:11

In this case,

play12:12

what Mike is doing is varying y over the space of different animals,

play12:16

in a network designed to recognize and distinguish

play12:18

different animals from each other.

play12:20

And you get this strange, Escher-like morph from one animal to another.

play12:26

Here he and Alex together have tried reducing

play12:30

the y's to a space of only two dimensions,

play12:33

thereby making a map out of the space of all things

play12:37

recognized by this network.

play12:38

Doing this kind of synthesis

play12:40

or generation of imagery over that entire surface,

play12:43

varying y over the surface, you make a kind of map --

play12:46

a visual map of all the things the network knows how to recognize.

play12:49

The animals are all here; "armadillo" is right in that spot.

play12:52

You can do this with other kinds of networks as well.

play12:55

This is a network designed to recognize faces,

play12:58

to distinguish one face from another.

play13:00

And here, we're putting in a y that says, "me,"

play13:03

my own face parameters.

play13:05

And when this thing solves for x,

play13:06

it generates this rather crazy,

play13:09

kind of cubist, surreal, psychedelic picture of me

play13:14

from multiple points of view at once.

play13:15

The reason it looks like multiple points of view at once

play13:18

is because that network is designed to get rid of the ambiguity

play13:22

of a face being in one pose or another pose,

play13:24

being looked at with one kind of lighting, another kind of lighting.

play13:28

So when you do this sort of reconstruction,

play13:30

if you don't use some sort of guide image

play13:32

or guide statistics,

play13:33

then you'll get a sort of confusion of different points of view,

play13:37

because it's ambiguous.

play13:39

This is what happens if Alex uses his own face as a guide image

play13:44

during that optimization process to reconstruct my own face.

play13:48

So you can see it's not perfect.

play13:50

There's still quite a lot of work to do

play13:52

on how we optimize that optimization process.

play13:55

But you start to get something more like a coherent face,

play13:57

rendered using my own face as a guide.

play14:00

You don't have to start with a blank canvas

play14:03

or with white noise.

play14:04

When you're solving for x,

play14:05

you can begin with an x, that is itself already some other image.

play14:09

That's what this little demonstration is.

play14:12

This is a network that is designed to categorize

play14:16

all sorts of different objects -- man-made structures, animals ...

play14:19

Here we're starting with just a picture of clouds,

play14:22

and as we optimize,

play14:24

basically, this network is figuring out what it sees in the clouds.

play14:28

And the more time you spend looking at this,

play14:31

the more things you also will see in the clouds.

play14:35

You could also use the face network to hallucinate into this,

play14:38

and you get some pretty crazy stuff.

play14:40

(Laughter)

play14:42

Or, Mike has done some other experiments

play14:45

in which he takes that cloud image,

play14:49

hallucinates, zooms, hallucinates, zooms hallucinates, zooms.

play14:52

And in this way,

play14:53

you can get a sort of fugue state of the network, I suppose,

play14:57

or a sort of free association,

play15:01

in which the network is eating its own tail.

play15:03

So every image is now the basis for,

play15:06

"What do I think I see next?

play15:08

What do I think I see next? What do I think I see next?"

play15:11

I showed this for the first time in public

play15:14

to a group at a lecture in Seattle called "Higher Education" --

play15:19

this was right after marijuana was legalized.

play15:22

(Laughter)

play15:26

So I'd like to finish up quickly

play15:28

by just noting that this technology is not constrained.

play15:33

I've shown you purely visual examples because they're really fun to look at.

play15:36

It's not a purely visual technology.

play15:39

Our artist collaborator, Ross Goodwin,

play15:41

has done experiments involving a camera that takes a picture,

play15:44

and then a computer in his backpack writes a poem using neural networks,

play15:49

based on the contents of the image.

play15:51

And that poetry neural network has been trained

play15:54

on a large corpus of 20th-century poetry.

play15:56

And the poetry is, you know,

play15:57

I think, kind of not bad, actually.

play15:59

(Laughter)

play16:01

In closing,

play16:02

I think that per Michelangelo,

play16:04

I think he was right;

play16:05

perception and creativity are very intimately connected.

play16:09

What we've just seen are neural networks

play16:12

that are entirely trained to discriminate,

play16:14

or to recognize different things in the world,

play16:16

able to be run in reverse, to generate.

play16:20

One of the things that suggests to me

play16:21

is not only that Michelangelo really did see

play16:24

the sculpture in the blocks of stone,

play16:26

but that any creature, any being, any alien

play16:30

that is able to do perceptual acts of that sort

play16:34

is also able to create

play16:35

because it's exactly the same machinery that's used in both cases.

play16:38

Also, I think that perception and creativity are by no means

play16:43

uniquely human.

play16:44

We start to have computer models that can do exactly these sorts of things.

play16:48

And that ought to be unsurprising; the brain is computational.

play16:51

And finally,

play16:53

computing began as an exercise in designing intelligent machinery.

play16:57

It was very much modeled after the idea

play17:00

of how could we make machines intelligent.

play17:03

And we finally are starting to fulfill now

play17:05

some of the promises of those early pioneers,

play17:08

of Turing and von Neumann

play17:09

and McCulloch and Pitts.

play17:12

And I think that computing is not just about accounting

play17:16

or playing Candy Crush or something.

play17:18

From the beginning, we modeled them after our minds.

play17:21

And they give us both the ability to understand our own minds better

play17:24

and to extend them.

play17:26

Thank you very much.

play17:27

(Applause)

Rate This

5.0 / 5 (0 votes)

Related Tags
Machine LearningNeuroscienceAI CreativityPerceptionGoogle AINeural NetworksArt and TechComputational ModelsMichelangeloTuring