Architecture All Access: Neuromorphic Computing Part 2

Intel Technology
22 Nov 202211:12

Summary

TLDRIn this episode of Architecture All Access, Mike Davies from Intel's Neuromorphic Computing Lab explores designing silicon chips that mimic the brain's efficiency. Discussing the challenges of replicating biological neural networks in CMOS circuits, Davies highlights the advantages of CMOS technology, such as speed and reliability. He outlines key neuromorphic principles like asynchronous communication and packetized spike routing, which Intel's Loihi 2 chip exemplifies. Davies emphasizes the progress in neuromorphic computing, particularly in solving complex optimization problems, and the ongoing challenge of programming these chips for practical applications.

Takeaways

  • 🧠 The goal of neuromorphic engineering is to design chips that mimic the brain's computational efficiency and structure.
  • 🔄 Despite decades of research, directly copying biological neural structures into CMOS circuits is impractical due to differences in design tools and manufacturing technology.
  • 🏎️ CMOS technology offers significant advantages in speed, with the ability to build circuits that operate at nanoseconds and gigahertz, compared to the brain's milliseconds and kilohertz.
  • 🔗 Neuromorphic chips focus on principles like sparse, distributed, and asynchronous communication to process sensory input quickly and minimize energy consumption.
  • 💡 The concept of asynchronous activation of neurons, where only the most activated neurons communicate, is crucial for efficient neuromorphic computing.
  • 🌐 Unlike the brain's 3D routing, neuromorphic chips use packetized spikes sent over shared time-multiplexed wiring channels to simulate more layers.
  • 🚀 Intel's Loihi 2 is a fully digital neuromorphic chip that uses an asynchronous design style, aligning with the brain's massive asynchronous circuitry.
  • 🔍 The neuromorphic core of Loihi 2 includes an asynchronous signal processing pipeline and internal memory, which is a critical resource in these chips.
  • 📉 Neuromorphic chips like Loihi 2 use time multiplexing to reduce the effective area per neuron, allowing for higher neural density within a small footprint.
  • 🤖 Implementing spike-based neuron models in neuromorphic chips results in lower activity and communication levels, leading to reduced power consumption.
  • 📚 Programming neuromorphic chips to perform useful computations remains a significant challenge, but progress is being made, with applications in optimization and pattern recognition.

Q & A

  • What is the main challenge in designing neuromorphic chips that mimic the brain's functionality?

    -The main challenge is the significant difference in design tools and manufacturing technology between silicon-based chips and biological systems. Specifically, the transistor area and wiring resources in silicon are much less efficient compared to the DNA-based molecular 3D self-assembly techniques used in biological systems.

  • How do neuromorphic chips differ from traditional chips in terms of speed?

    -Neuromorphic chips can operate at nanoseconds and gigahertz, which is significantly faster than biological neurocircuits that operate at milliseconds and kilohertz scales.

  • What is the advantage of CMOS semiconductor manufacturing technology in neuromorphic computing?

    -CMOS technology allows for the construction of fast and reliable circuits that operate precisely and deterministically, unlike the brain where precision and reliability come at the expense of redundancy or more neuro-resources.

  • What principle of brain computation is adapted in neuromorphic chip design to minimize energy consumption and manufacturing cost?

    -The principle of sparse, distributed, asynchronous communication is adapted, which allows for the processing of the most important inputs quickly with minimal latency.

  • How do neuromorphic chips handle the routing of spikes compared to the brain's 3D routing?

    -Neuromorphic chips use packetized spikes sent over shared time multiplexed horizontal wiring channels, effectively creating more vertical layers and allowing for faster distribution over a routed mesh network.

  • What is the significance of the speed advantage in neuromorphic chips for spike routing?

    -The speed advantage allows neuromorphic chips to distribute thousands of spikes over the same wires in biological time scales without interference, simulating the effect of each neuron having its own dedicated axon wiring network.

  • How does the neuromorphic core in chips like Loihi 2 reduce its effective area per neuron?

    -The neuromorphic core in Loihi 2 continues to time multiplex its circuitry, which allows for a significant reduction in the effective area per neuron, enabling the implementation of several thousand neurons in a small area.

  • What is the role of memory in neuromorphic chips, and how is it optimized in Loihi 2?

    -Memory is a critical resource in neuromorphic chips, and its effective area cannot be reduced using time multiplexing. In Loihi 2, the core circuitry is optimized to use this fixed pool of memory efficiently, with features like convolutional features and a granular array of memory banks.

  • How does the learning process in neuromorphic chips like Loihi 2 work?

    -The learning process in Loihi 2 operates in the background, updating the parameters of the core's neural network, particularly the synaptic state variables, based on a combination of input and output side activity at each synapse.

  • What are some of the applications where neuromorphic chips like Loihi have shown significant gains over conventional solvers?

    -Neuromorphic chips like Loihi have shown significant gains in solving optimization problems such as railway scheduling and QUBO from quantum computing, compared to the best conventional solvers.

  • What is the current challenge in the field of neuromorphic computing regarding the chips' programming?

    -The current challenge is understanding how to program neuromorphic chips to perform useful computations effectively, as it is still a formidable task despite the progress being made in the field.

Outlines

00:00

🧠 Introduction to Neuromorphic Computing

The first paragraph introduces the concept of neuromorphic computing and the challenges of designing silicon-based chips that mimic the brain's functionality. It discusses the limitations of directly copying biological neural structures due to differences in design tools and manufacturing technology. The narrator, Mike Davies, highlights the advantages of CMOS technology, such as speed and reliability, compared to biological systems. The principle of sparse, distributed, asynchronous communication is introduced as a fundamental concept in neuromorphic design, emphasizing the importance of processing important inputs quickly and efficiently. The paragraph concludes with a discussion on the need to adapt brain computation principles to the design tools available in chip manufacturing.

05:01

🔍 Deep Dive into Neuromorphic Chip Design

The second paragraph delves into the specifics of neuromorphic chip design, focusing on the Loihi 2 chip developed by Intel's Neuromorphic Computing Lab. It explains how the chip uses a fully digital approach and asynchronous design style to mimic the brain's massive asynchronous circuit. The paragraph discusses the importance of memory in neuromorphic chips and how the core circuitry is optimized to use memory efficiently. It details the process of spike routing, neuron activation, and the learning process that operates in the background to update the neural network's parameters. The summary also touches on the challenges of programming neuromorphic chips for useful computation and the progress being made in the field, with examples of applications in optimization problems.

10:03

🚀 Conclusion and Future of Neuromorphic Computing

The final paragraph concludes the discussion on neuromorphic computing by emphasizing the progress made in the field and the challenges that still lie ahead. It acknowledges that while building neuromorphic chips is achievable, programming them for practical applications remains a significant challenge. The paragraph highlights some of the successes, such as solving complex optimization problems more efficiently than traditional methods. It encourages viewers to follow the progress of neuromorphic computing as it moves from the lab to real-world applications, suggesting that this technology has the potential to revolutionize various industries.

Mindmap

Keywords

💡Neuromorphic Computing

Neuromorphic Computing refers to the development of computer systems that mimic the neural structure and functioning of the human brain. In the context of the video, it is the central theme, with the discussion revolving around designing chips that behave more like a brain than traditional silicon chips. The video explains how neuromorphic chips can replicate some of the form and functions of biological neural networks, aiming to process sensory input quickly while minimizing energy consumption and manufacturing cost.

💡CMOS Circuits

CMOS, which stands for Complementary Metal-Oxide-Semiconductor, is a technology used for constructing integrated circuits. The video mentions that CMOS circuits can replicate some aspects of biological neural networks, but there are practical challenges in directly copying the structure and dynamics of biological neurons due to differences in design tools and manufacturing technology.

💡Sparse, Distributed, Asynchronous Communication

This concept refers to a communication method where only the most strongly activated neurons transmit their signals asynchronously, rather than all neurons sending signals simultaneously. The video explains that this principle can lead to quicker processing of important inputs with minimal latency, which is a key aspect of neuromorphic chip design aimed at emulating the efficiency of the brain's communication.

💡Spike-based Neuron Models

In the video, spike-based neuron models are discussed as a way to implement biological neurons' behavior in neuromorphic chips. These models use spikes, or brief electrical signals, to transmit information, similar to how biological neurons communicate. The video emphasizes that these models allow for lower activity and communication levels, leading to more efficient routing and lower power consumption.

💡Time Multiplexing

Time multiplexing is a technique where multiple signals are transmitted over a single communication channel by dividing the signal into time slots. The video explains that neuromorphic cores use time multiplexing to reduce their effective area per neuron, thus increasing the density of neurons that can be implemented on a chip.

💡Routing

Routing in the context of neuromorphic chips refers to the way spikes are directed through the chip's architecture. The video discusses how neuromorphic chips cannot directly copy the three-dimensional routing of the brain but instead use packetized spikes sent over shared time multiplexed channels to achieve efficient routing.

💡Loihi 2

Loihi 2 is a neuromorphic chip developed by Intel, as mentioned in the video. It is a fully digital chip implemented in a standard CMOS process and is designed to support many of the computational principles found in brains. The video provides an overview of Loihi 2's architecture, highlighting its asynchronous design style and internal memory structures.

💡Asynchronous Circuits

Asynchronous circuits are electronic circuits that do not rely on a global clock signal to synchronize operations. The video explains that neuromorphic chips, including Loihi, use asynchronous design styles, which is well-suited for spiking neural networks that are only sparsely active, leading to power efficiency.

💡Learning Rules

In the context of neuromorphic computing, learning rules refer to the algorithms or equations that govern how the synaptic state variables of a neural network evolve over time. The video discusses the importance of getting these rules right to enable the network to adapt and learn new patterns, which is a challenging and active area of research.

💡Optimization

Optimization in the video refers to the application of neuromorphic chips like Loihi in solving complex problems more efficiently than traditional methods. The video gives examples such as railway scheduling and QUBO problems from quantum computing, where neuromorphic chips can provide significant gains in performance.

Highlights

Mapping neuromorphic concepts into silicon chips is a challenge that requires adapting brain-like computation to CMOS technology.

CMOS circuits can replicate some form and functions of biological neural networks, but direct copying is impractical due to differences in design tools and manufacturing.

Transistor area and wiring resources in chip design are significantly different from biological systems, leading to a need for innovative approaches.

CMOS technology offers advantages in speed, with circuits operating at nanoseconds and gigahertz, compared to biological systems' milliseconds and kilohertz.

The principle of sparse, distributed, asynchronous communication is fundamental in neuromorphic chip design, mimicking the efficiency of biological neural networks.

Neuromorphic chips aim to implement neurons that represent their activations as events in time, similar to biological spikes.

Routing spikes in neuromorphic chips requires innovative solutions due to the limitation in vertical wiring compared to the brain's three-dimensional routing.

Packetizing spikes and sending them over shared time multiplexed channels can overcome the wiring disadvantage in chip design.

The speed advantage of CMOS allows for efficient spike distribution, simulating the effect of dedicated axon wiring networks.

Modern semiconductor process technology can pack metal in a one-micron strip, approaching the efficiency and density of the brain's myelinated axons.

Neuromorphic cores need to time multiplex their circuitry to reduce the effective area per neuron, achieving high neural density.

Spike-based neuron models in neuromorphic chips result in lower activity and communication levels, leading to lower power consumption.

Loihi 2, a neuromorphic chip developed by Intel, is a fully digital chip implemented in a standard CMOS process, optimized for asynchronous operation.

Asynchronous circuits in neuromorphic chips consume power only when computing, matching the sparse activity of spiking neural networks.

Loihi 2's neuromorphic core features an asynchronous signal processing pipeline with internal memory, optimized for efficient memory usage.

The learning process in Loihi operates in the background, updating the neural network's parameters based on input and output activity.

Programming neuromorphic chips for useful computation remains a challenge, but the field is making progress with exciting results in optimization problems.

Loihi has demonstrated significant gains in solving complex problems like railway scheduling and QUBO from quantum computing, compared to conventional solvers.

Transcripts

play00:00

(bright music)

play00:12

- [Narrator] How do we map neuromorphic concepts

play00:14

into chips built from silicon?

play00:17

Here's Mike Davies, Senior Principal Engineer

play00:20

and Director of Intel's Neuromorphic Computing Lab.

play00:25

This is Architecture All Access.

play00:30

- Let's now talk about how to architect a chip

play00:33

that behaves more like a brain

play00:35

than the chips we're familiar with today.

play00:37

If we step back and look at the state of modern neuroscience

play00:40

and chip design,

play00:41

we see that we can clearly build CMOS circuits

play00:44

that replicate some of the form and functions

play00:46

of biological neural networks.

play00:48

People have been experimenting with this

play00:50

for over three decades,

play00:52

but an approach of directly copying the structure

play00:55

and circuit dynamics of biological neurons

play00:58

quickly runs into practical trouble.

play01:00

The design tools and manufacturing technology

play01:03

we're working with are simply too different from biology,

play01:06

specifically in terms of transistor area

play01:09

and wiring resources.

play01:11

We are many orders of magnitude

play01:13

behind what evolution has achieved

play01:15

with its DNA-based molecular 3D self-assembly techniques.

play01:20

As a chip designer,

play01:21

I'd love to get my hands on nature's circuit design toolbox,

play01:25

but those capabilities

play01:26

are going to be a very long time to come.

play01:28

On the other hand,

play01:29

today's CMOS semiconductor manufacturing technology

play01:33

does give us some important advantages.

play01:36

In speed, for example,

play01:37

we can construct extraordinarily fast circuits

play01:40

compared to biological neurocircuits.

play01:42

While biology operates at milliseconds and kilohertz scales,

play01:46

our CMOS circuits operate at nanoseconds and gigahertz,

play01:50

and we can construct reliable circuits

play01:52

that operate precisely and deterministically.

play01:55

Whereas in the brain, precision and reliability

play01:59

come at the expense of redundancy or more neuro-resources.

play02:03

So this leads us away

play02:04

from the idea of literally copying

play02:06

or exactly replicating the neural structures

play02:09

we find in the brain.

play02:10

Instead, we have to understand

play02:12

and adapt the underlying principles of brain computation

play02:16

to the design toolbox we're working with.

play02:18

But despite the different design tools,

play02:20

the same principles surely apply

play02:23

since we are targeting the same objectives,

play02:25

intelligently process sensory input quickly

play02:28

while minimizing both energy consumption

play02:30

and manufacturing cost.

play02:32

So let's consider some of these principles.

play02:36

One of the most fundamental is the virtue

play02:38

of sparse, distributed, asynchronous communication.

play02:42

Consider the problem of communicating the activations

play02:45

of a set of neurons to some downstream neuron.

play02:48

If somehow only the most strongly activated neurons

play02:51

could asynchronously announce themselves in time,

play02:54

then the most important inputs can be processed quickly

play02:57

with minimal latency.

play02:59

On the other hand, if the entire set of input neurons

play03:02

has to be processed as one dense matrix,

play03:04

which is the standard way

play03:05

in today's mainstream architectures,

play03:07

then the important inputs will get congested

play03:10

and delayed behind the herd of less important inputs.

play03:13

This may seem obvious, but to truly exploit this principle,

play03:17

a hardware implementation needs to respond

play03:20

to unpredictable neuron activations immediately,

play03:23

which implies extremely fine grain parallelism

play03:27

and neurons that asynchronously represent their activations

play03:30

as events in time rather than as numbers

play03:33

in a synchronously processed vector.

play03:35

This is what biological neurons do using spikes,

play03:39

and it's what we aim to implement in neuromorphic chips.

play03:43

Now, if we turn to the problem

play03:45

of routing those spikes in a chip we can build,

play03:48

we're gonna find a clear example

play03:49

where we can't just copy biology.

play03:52

Brains use truly three-dimensional routing

play03:55

with tens of thousands of overlapping wires

play03:57

through any cross section.

play03:59

In contrast, our chips, vertically,

play04:02

we can only stack about 20 metal layers.

play04:04

This is a huge wiring disadvantage.

play04:07

But remember, we have an enormous speed advantage.

play04:11

So what we can do is packetize the spikes and send them

play04:14

over shared time multiplexed horizontal wiring channels,

play04:18

giving us effectively many more vertical layers.

play04:21

The packets can then be distributed

play04:23

over a routed mesh network

play04:24

at speeds of up to a million times faster

play04:27

than what biology uses for its spikes.

play04:29

At those speeds,

play04:31

the routers can distribute thousands of spikes

play04:33

over the same wires

play04:34

in biological time scales without interference,

play04:37

as if each neuron had its own dedicated axon wiring network

play04:41

as in the brain.

play04:43

Now, you may think this looks incredibly inefficient.

play04:46

Spikes sent as packets with 32 or more bits

play04:49

sent over bundles of wires.

play04:51

How could this possibly reach the level

play04:53

of efficiency and density in the brain?

play04:56

But actually, we're not that far off.

play04:58

The brain's long distance wires are myelinated axons,

play05:01

which have a diameter of about one micron.

play05:04

In modern semiconductor process technology,

play05:06

we can easily pack all the metal we need

play05:09

in a one micron strip of planar wires.

play05:12

So our brain-inspired spiking interconnect is

play05:15

in the ballpark of biology.

play05:17

And once we have these routers,

play05:19

we can now connect them

play05:20

to neuromorphic cores that produce and consume these spikes

play05:24

by modeling the essential behaviors

play05:26

of biology's temporal spike-based neurons.

play05:29

Each core needs to continue to time multiplex its circuitry

play05:33

so it can reduce its effective area per neuron

play05:36

by orders of magnitude.

play05:38

This way, each core with a fraction of a square millimeter

play05:41

can implement several thousand neurons,

play05:43

which actually approaches the neural density

play05:45

of the neocortex.

play05:46

By implementing spike-based neuron models,

play05:49

the activity and communication levels are far lower

play05:52

than what you'd have on a conventional chip.

play05:54

The cores send their spikes

play05:56

as sparse peer-to-peer distributions,

play05:58

leaving the routing network relatively idle

play06:01

compared to the buses in a conventional chip.

play06:03

In this architecture,

play06:04

there's no off-chip memory interface

play06:06

that has to stay saturated

play06:08

with predictable vectorized memory patterns.

play06:10

This all translates to lower power.

play06:13

Now moving on to a real neuromorphic chip.

play06:15

Here we have Loihi 2, developed by my group,

play06:18

and one of the most recently developed neuromorphic chips.

play06:22

The chip plot frankly looks rather boring

play06:24

because, like a memory chip,

play06:26

it's mostly a single repeated core instantiated in a mesh

play06:29

and the core is itself dominated

play06:31

by internal memory structures.

play06:34

Loihi 2 is a fully digital chip

play06:36

implemented in a standard CMOS process.

play06:39

Historically, many neuromorphic research chips

play06:42

have used analog circuits to replicate the membrane dynamics

play06:45

of biological neurons, but for many reasons

play06:48

relating to circuit density, precision, reliability,

play06:52

the field has generally moved to fully digital designs,

play06:55

at least for now.

play06:56

The brain is one massive asynchronous circuit,

play06:59

and a lot of neuromorphic chips, including our Loihi chips,

play07:02

are implemented using an asynchronous design style.

play07:06

Asynchronous circuits don't use free running clocks

play07:09

and automatically stop consuming power

play07:11

when there's nothing to compute.

play07:12

Of course, this is a great match

play07:14

for spiking neural networks that are only sparsely active.

play07:18

If we now dive a bit deeper

play07:19

into the details of Loihi 2's neuromorphic core,

play07:23

you'll find an asynchronous signal processing pipeline

play07:26

with a lot of internal memory.

play07:28

Memory is the scariest resource in a neuromorphic chip,

play07:31

and unlike compute circuitry,

play07:33

its effective area cannot be reduced

play07:35

using time multiplexing.

play07:37

So the rest of the core circuitry is highly optimized

play07:40

to use this fixed pool of memory as efficiently as possible.

play07:45

For example, convolutional features allow a

play07:47

repeated pattern of synaptic weights to be stored only once

play07:51

and then reused across multiple neurons.

play07:54

The core's memory is a granular array

play07:56

of many banks with varying sizes.

play07:59

Different processes in the core

play08:01

require different levels of access parallelism.

play08:04

The memory can be allocated

play08:05

in a flexible way across the core's pipeline stages,

play08:08

although the performance and storage density are optimized

play08:11

for the typical configurations.

play08:13

As spikes arrive at the core,

play08:15

table lookups determine what neurons in the core

play08:18

the spikes connect to.

play08:20

The lookups return data structures that specify weights

play08:23

and propagation delays for each destination neuron.

play08:27

As each spike is applied to its fanout neurons,

play08:30

the associated weights are absorbed

play08:32

by a dendritic accumulation process

play08:35

which maintains the total received synaptic stimulation

play08:38

for each of the core's neurons.

play08:40

In a completely decoupled asynchronous process,

play08:42

the dendrite stage reads

play08:45

each neuron's accumulated synaptic input

play08:47

and then updates the neuron's state variables

play08:50

for the current time

play08:51

following the neuron's programmed model.

play08:54

Typically, synaptic input causes a neuron state variable

play08:57

to jump in a positive or negative direction,

play09:00

while the absence of input causes a gradual decay

play09:03

of the variable to zero.

play09:05

If a neuron becomes sufficiently activated

play09:08

and crosses a threshold, then it generates an output spike.

play09:11

Further table lookups in an egress process

play09:14

determine which core is in the network

play09:16

the spiking neuron connects to,

play09:18

and then spike packets are sent

play09:19

through the routing infrastructure

play09:21

to those destination cores.

play09:23

In Loihi, we also have a learning process that operates

play09:26

in the background,

play09:27

updating the parameters of the core's neural network,

play09:31

particularly the synaptic state variables.

play09:34

Based on a combination of the input side activity

play09:36

and the output side activity at each synapse,

play09:39

its state variables, such as its weight,

play09:41

can evolve over time according to programmed equations.

play09:45

Getting these so-called learning rules right

play09:47

so that the large scale behavior of the network

play09:50

is to adapt online and usefully learn new patterns

play09:54

is one of the trickiest and most actively researched areas

play09:57

of neuromorphic computing.

play09:59

And that concludes our quick tour

play10:00

of neuromorphic architecture.

play10:02

Hopefully, I provided an understandable glimpse

play10:05

into the world of neuromorphic computing

play10:07

and you're leaving with a greater appreciation

play10:09

for this rather different and exotic architecture.

play10:12

The architecture is actually the relatively easy part.

play10:15

We can build these chips today

play10:17

that support many of the computational principles

play10:19

we find in brains,

play10:21

but understanding how to program them

play10:23

to perform useful computation,

play10:25

that's still admittedly a formidable challenge.

play10:28

The field is making a lot of progress on that though

play10:31

with some exciting results.

play10:33

For example, in the domain of optimization,

play10:35

Loihi can solve problems like railway scheduling,

play10:38

and QUBO from quantum computing

play10:40

with orders of magnitude gains

play10:42

compared to the best conventional solvers.

play10:45

So please check out some of our recent publications

play10:47

and follow our progress as we get this technology

play10:50

out of the lab

play10:51

and into game-changing products.

play10:54

This has been

play10:54

Architecture All Access, Neuromorphic Computing.

play10:57

Thanks for joining us.

play10:59

(bright music)

Rate This

5.0 / 5 (0 votes)

Связанные теги
NeuromorphicChip DesignIntel LabsBrain-InspiredCMOS CircuitsAsynchronousSparse CodingEnergy EfficiencyAI InnovationTech Advancement
Вам нужно краткое изложение на английском?