Lecture 1.2 — What are neural networks — [ Deep Learning | Geoffrey Hinton | UofT ]

Artificial Intelligence - All in One
24 Sept 201708:30

Summary

TLDR本视频深入探讨了真实神经元的工作原理,这些神经元启发了人工神经网络的设计。视频中首先强调了研究神经元网络的三个主要原因:理解大脑工作方式、启发新型并行计算方式以及解决实际问题。接着,视频详细描述了单个神经元的结构,包括细胞体、轴突和树突,以及神经元之间通过突触进行通信的过程。突触通过释放神经递质来改变后神经元的去极化状态,从而实现信号传递。此外,视频还讨论了突触如何通过改变神经递质的数量或受体分子的敏感性来适应和学习。最后,视频指出大脑皮层的模块化特性,以及大脑的灵活性和通用学习算法,这些特性使得大脑能够通过经验学习新的功能。

Takeaways

  • 🧠 真实神经元启发了人工神经网络的设计,它们在大脑中的网络结构是并行计算的基础。
  • 🧪 通过计算机模拟帮助理解大脑,因为大脑复杂且不易直接实验。
  • 💡 研究神经元网络的计算方式有助于开发更好的并行计算机,尤其是在视觉等大脑擅长的领域。
  • 🚫 人工神经网络与大脑的实际工作方式可能不同,但它们解决实际问题的能力依然非常有用。
  • 🌐 一个典型的皮层神经元由细胞体、轴突和树突构成,通过突触与其他神经元通信。
  • ⚡️ 神经元通过在轴突上传递电位尖峰来发送消息,这些尖峰能够触发突触释放神经递质。
  • 🔄 突触的适应性是学习过程的关键,通过改变突触的效能来实现。
  • 🔬 突触的结构包含小囊泡,它们包含传递信号的化学物质,影响后续神经元的激活状态。
  • 📉 突触的响应速度虽然慢,但它们小而节能,并且能够适应,这是它们的主要优势。
  • 🧬 大脑皮层是模块化的,不同的区域负责不同的功能,这种分工是通过经验形成的。
  • 🔄 大脑具有通用学习算法,能够灵活适应不同的任务,类似于FPGA(现场可编程门阵列)。
  • 🧬 如果早期大脑受损,功能可以转移到其他部位,表明大脑具有可塑性,可以根据经验重新分配功能。

Q & A

  • 为什么我们需要通过计算机模拟来理解大脑的工作方式?

    -大脑非常复杂且体积庞大,直接实验操作存在困难,因为大脑在被操作时可能会受损甚至死亡。计算机模拟可以帮助我们理解大脑的功能和结构,尤其是在进行实证研究时。

  • 为什么研究大脑的并行计算对我们设计更好的并行计算机有帮助?

    -大脑通过大量相对较慢的神经元构成的并行网络进行计算,这种计算方式与常规串行处理器截然不同。理解这种并行计算模式有助于我们设计出更适合进行视觉等大脑擅长任务的并行计算机。

  • 神经网络的灵感来源于哪里,它们在解决实际问题中有什么作用?

    -神经网络的灵感来源于大脑的工作机制。这些受大脑启发的新型学习算法在解决实际问题时非常有用,即使它们并非大脑实际工作的方式。

  • 典型的大脑皮层神经元的物理结构包括哪些部分?

    -典型的大脑皮层神经元的物理结构包括细胞体、轴突(用于向其他神经元发送信息)和树突(用于接收来自其他神经元的信息)。

  • 突触是如何工作的,它在神经元之间传递信息中扮演什么角色?

    -突触是神经元之间传递信息的结构。当一个神经元的轴突与另一个神经元的树突接触时,形成突触。轴突上的电位变化(称为动作电位或“spike”)会在突触处引起电荷注入,从而传递信号。

  • 突触中的神经递质是如何影响神经元的?

    -突触中的神经递质通过在突触前神经元的动作电位到达时释放,然后扩散穿过突触间隙,与突触后神经元的受体分子结合。这种结合改变了受体分子的形状,从而在细胞膜上形成通道,允许特定离子流入或流出,改变突触后神经元的去极化状态。

  • 突触如何适应并进行学习?

    -突触通过改变其效能来适应和学习。这可以通过改变每次动作电位到达时释放的囊泡数量,或者通过改变对释放的神经递质敏感的受体分子数量来实现。

  • 大脑皮层的神经元数量大约有多少,每个神经元平均有多少个突触连接?

    -大脑皮层大约有10^11个神经元,每个神经元平均有10^4个突触连接,因此总共大约有10^15到10^14个突触权重。

  • 大脑皮层是如何实现模块化的,模块化对学习有什么好处?

    -大脑皮层通过不同的部分执行不同的功能来实现模块化。这种模块化使得大脑能够更有效率地处理信息,并且具有更好的适应性,因为如果早期大脑受到损伤,功能可以重新定位到大脑的其他部分。

  • 大脑皮层的灵活性如何体现在其学习和功能定位上?

    -大脑皮层的灵活性体现在它可以通过经验将通用的计算硬件转变为特定任务的专用硬件。例如,对小猫进行的实验表明,如果听觉皮层的输入被视觉输入取代,那么原本处理声音的听觉皮层可以学会处理视觉输入。

  • 大脑的并行计算与FPGA(现场可编程门阵列)有何相似之处?

    -大脑的并行计算与FPGA相似,因为它们都构建了标准的并行硬件,然后通过输入信息来指定特定的并行计算任务。这种结构允许快速的并行计算,并且具有在学习新功能时的灵活性。

  • 为什么传统的计算机需要非常快的中央处理器?

    -传统的计算机需要非常快的中央处理器来访问存储在程序中的线条,并执行长时间的顺序计算。这是因为它们通过存储的顺序程序来获得灵活性,这要求快速的中央处理器来访问程序中的线条并执行计算。

Outlines

00:00

🧠 大脑神经元与人工神经网络的启发

本段介绍了真实的大脑神经元如何启发了人工神经网络的设计。强调了研究神经元网络的三个主要原因:理解大脑的工作原理、探索大脑并行计算风格以提升并行计算机性能、以及开发受大脑启发的新型学习算法来解决实际问题。详细描述了神经元的物理结构,包括细胞体、轴突、树突以及突触的工作机制。突触通过改变其效能来适应,这是学习过程的关键,突触的适应性通过改变释放的囊泡数量或受体分子的敏感度来实现。

05:00

🤖 大脑的并行计算与学习算法

这段内容深入探讨了大脑的并行计算能力,以及大脑如何通过调整突触权重来学习执行各种任务,如识别物体、理解语言、制定计划和控制身体运动。描述了大脑中约有10^11个神经元,每个神经元大约有10^4个突触权重,形成了一个巨大的突触权重网络。这些权重可以在几毫秒内影响正在进行的计算,展现了比现代工作站更高的带宽到存储知识的比率。此外,还讨论了大脑皮层的模块化特性,不同的皮层区域负责不同的功能,这种模块化是后天经验形成的,而非遗传决定。通过实验,如切断幼年雪貂的听觉皮层输入并重新路由视觉输入,证明了大脑皮层具有通用学习算法,能够将通用硬件转变为针对特定任务的专用硬件。

Mindmap

Keywords

💡神经元

神经元是神经系统的基本工作单元,负责接收、处理和传递信息。在视频中,神经元的结构和功能是理解人工神经网络的基础,因为人工神经网络的设计灵感来源于生物神经元的工作方式。例如,视频中提到典型的皮层神经元由细胞体、轴突和树突构成,它们通过突触与其他神经元进行信息交流。

💡突触

突触是神经元之间传递信号的结构,它包含有传递化学物质的小囊泡。在视频中,突触的作用是通过释放神经递质来改变另一个神经元的电位,从而实现神经信号的传递。突触的适应性是学习过程的关键,它通过改变释放囊泡的数量或受体分子的敏感度来调整其效能。

💡神经递质

神经递质是突触中传递信号的化学物质,它们可以是兴奋性的或抑制性的。视频中提到,当一个神经冲动到达轴突时,神经递质会从囊泡中释放到突触间隙,并与后突触神经元的受体分子结合,改变细胞膜的形状,从而影响后突触神经元的去极化状态。

💡去极化

去极化是神经元细胞膜电位的变化过程,当神经元接收到足够的电荷时,会导致细胞体的某个部分(轴突结节)去极化。视频中解释说,去极化会导致神经元沿着其轴突发送一个冲动,即去极化波。这个过程是神经元传递信号的关键步骤。

💡人工神经网络

人工神经网络是受生物神经系统启发而设计的计算模型,它们模仿生物神经元的连接和处理信息的方式。视频中提到,人工神经网络可以解决实际问题,如视觉识别、语言理解等,尽管它们可能并不完全模拟大脑的工作方式。

💡并行计算

并行计算是一种计算方式,它允许多个计算任务同时进行,以提高计算效率。视频中提到,大脑通过大量并行的神经元网络进行计算,这种计算方式与常规的串行处理器不同,特别适合于大脑擅长的任务,如视觉处理。

💡

💡学习算法

学习算法是人工神经网络中用于训练模型以执行特定任务的算法。视频中指出,这些算法受到大脑的启发,可以用于解决实际问题,即使它们可能并不完全模拟大脑的工作机制。学习算法通过调整网络中的权重来优化网络的性能。

💡权重

权重是人工神经网络中连接神经元的边的属性,它们决定了信号在网络中的传递强度。在视频中,权重可以是正的或负的,并且可以通过学习过程进行调整。权重的调整是神经网络学习的基础,它们影响着网络对输入信号的响应。

💡大脑皮层

大脑皮层是大脑的外层,负责处理复杂的认知功能,如感觉知觉、语言和决策。视频中提到,大脑皮层是模块化的,不同的区域负责不同的功能,但它们看起来非常相似,表明它们可能具有通用的学习算法。

💡模块化

模块化是指系统的不同部分具有特定的功能,并且可以独立于其他部分工作。在视频中,大脑皮层的模块化意味着不同的皮层区域负责处理不同类型的信息,如视觉、听觉或语言信息。这种模块化有助于提高大脑处理信息的效率。

💡灵活性

灵活性是指系统适应新情况或学习新任务的能力。视频中提到,大脑的灵活性表现在它能够重新分配功能到不同的区域,特别是在早期损伤后,这表明大脑具有通用的学习算法,可以根据经验调整其功能。

Highlights

真实神经元为人工神经网络提供了灵感,这些网络将在本课程中学习。

研究神经元网络如何计算的三个主要原因:理解大脑工作原理、启发并行计算风格、解决实际问题。

大脑的复杂性和脆弱性使得计算机模拟成为理解大脑工作的重要工具。

大脑的并行计算风格与常规串行处理器的计算方式截然不同。

人工神经网络算法即使不完全模仿大脑工作方式,也非常有用。

皮层神经元的物理结构包括细胞体、轴突和树突。

突触是神经元间传递信息的关键结构,通过释放神经递质来实现信号传递。

神经递质通过改变膜上受体分子的形状来影响神经元的去极化状态。

突触的适应性是学习过程中最重要的特性,通过改变突触的效率实现。

大脑中有大约10^11个神经元,每个神经元平均有10^4个突触权重。

大脑的带宽到存储知识的能力远超现代工作站。

大脑皮层是模块化的,不同的皮层区域负责不同的功能。

大脑功能的定位可以通过观察血流变化来了解。

大脑皮层看起来大致相同,暗示它拥有一个灵活的通用学习算法。

大脑的灵活性允许功能在早期损伤后重新定位到其他区域。

实验表明,大脑皮层能够将通用硬件转变为针对特定任务的专用硬件。

大脑的快速并行计算与学习新功能的灵活性相结合,类似于FPGA。

传统计算机通过存储的顺序程序获得灵活性,但需要非常快的中央处理器。

Transcripts

play00:00

in this video I'm going to tell you a

play00:03

little bit about real neurons on the

play00:05

real brain which provide the inspiration

play00:07

for the artificial neural networks that

play00:10

we're going to learn about in this

play00:11

course in most of the course we won't

play00:14

talk much about really ions but I wanted

play00:17

to give you a quick overview at the

play00:19

beginning there are several different

play00:22

reasons to study how networks of neurons

play00:25

can compute things the first is to

play00:28

understand how the brain actually works

play00:31

you might think we could do that just by

play00:33

experiments on the brain but it's very

play00:36

big and complicated and it dies when you

play00:38

poke it around and so we need to use

play00:40

computer simulations to help us

play00:43

understand what we're discovering in

play00:45

empirical studies the second is to

play00:49

understand a style of parallel

play00:50

computation that's inspired by the fact

play00:52

that the brain can compute with a big

play00:55

parallel network of relatively slow

play00:57

neurons if we can understand that style

play00:59

of parallel computation we might be able

play01:01

to make better parallel computers it's

play01:04

very different from the way computation

play01:06

is done on a conventional serial

play01:08

processor it should be very good for

play01:11

things that brains are good at like

play01:12

vision and it should also be bad for

play01:16

things that brains are bad at like

play01:17

multiplying two numbers together a third

play01:21

reason which is the relevant one for

play01:23

this course is to solve practical

play01:25

problems by using novel learning

play01:27

algorithms that were inspired by the

play01:29

brain these algorithms can be very

play01:32

useful even if they're not actually how

play01:34

the brain works so most of this course

play01:37

we won't talk much about how the brain

play01:38

actually works it's just used as a

play01:40

source of it for inspiration to tell us

play01:43

that big parallel networks of neurons

play01:44

can compute very complicated things I'm

play01:49

going to talk more in this video though

play01:52

about how the brain actually works

play01:54

a typical cortical neuron has a gross

play01:58

physical structure that consists of a

play02:00

cell body and an axon where it sends

play02:03

messages to other neurons and a

play02:06

dendritic tree where it receives

play02:08

messages from other neurons where an

play02:11

axon from one neuron

play02:13

contacts a dendritic tree of another

play02:15

neuron there's a structure called the

play02:16

synapse

play02:17

and a spike of activity traveling along

play02:21

the axon causes charge to be injected

play02:25

into the postsynaptic neuron at a

play02:28

synapse

play02:29

a neuron generates spikes when it's

play02:34

received enough charge in its dendritic

play02:36

tree to depolarize a part of the cell

play02:39

body called the axon hillock and when

play02:42

that gets depolarized the neuron sends a

play02:44

spike edge along its axon when the

play02:46

spikes just a wave of depolarization

play02:48

that travels along the axon synapses

play02:53

themselves have interesting structure

play02:58

they contain little vesicles of

play02:59

transmitter chemical and when a spike

play03:01

arrives in the axon it causes these

play03:05

vesicles to migrate to the surface and

play03:07

be released into the synaptic cleft

play03:09

there's several different kinds of

play03:11

transmitter chemical there's ones that

play03:13

implement positive weights and ones that

play03:15

implement negative weights the

play03:18

transmitter molecules diffuse across the

play03:19

synaptic cleft and bind to receptor

play03:22

molecules in the membrane of the

play03:23

postsynaptic neuron and by binding to

play03:26

these big molecules in the membrane they

play03:28

change their shape and that creates

play03:31

holes in the membrane these holes are

play03:34

life specific ions to flow in or out of

play03:37

the postsynaptic neuron and that changes

play03:39

their state of depolarization synapses

play03:45

adapt and that's what most of learning

play03:48

is changing the effectiveness of a

play03:50

synapse they can adapt by varying the

play03:52

number of vesicles that get released

play03:54

when a spike arrives or by varying the

play03:58

number of receptor molecules that are

play04:00

sensitive to the released transmitter

play04:02

molecules synapses are very slow

play04:06

compared with computer memory but they

play04:09

have a lot of advantages over the random

play04:10

access memory on a computer they're very

play04:13

small and very low power and they can

play04:17

adapt that's the most important property

play04:19

they use locally available signals to

play04:21

change their strengths and that's how we

play04:23

learn to perform

play04:25

any computations the issue of course is

play04:28

how do they decide how to change their

play04:31

strengths what is the what are the rules

play04:33

for how they should adapt so all on one

play04:38

slide this is how the brain works each

play04:41

neuron receives inputs from other

play04:42

neurons a few of the neurons receive

play04:45

inputs from the receptors it's a large

play04:48

number of neurons but only a small

play04:49

fraction of them and the neurons

play04:52

communicate with each other in the

play04:54

cortex by sending these spikes of

play04:56

activity the effect of an input line on

play05:00

the neuron is controlled by synaptic

play05:01

weight which can be positive or negative

play05:04

and these synaptic weights adapt and by

play05:08

adapting these weights the whole network

play05:10

learns to perform different kinds of

play05:11

computation for example recognizing

play05:13

objects understanding language making

play05:16

plans controlling the movements of your

play05:18

body you have about 10 to the 11 neurons

play05:22

each of which has about 10 to the 4

play05:25

weights so you probably have 10 to the

play05:28

15 or maybe only 10 to the 14 synaptic

play05:31

weights and a huge number of these

play05:34

weights quite a large fraction of them

play05:36

can affect the ongoing computation in a

play05:39

very small fraction of a second in a few

play05:42

milliseconds that's much better

play05:44

bandwidth to stored knowledge than even

play05:46

a modern workstation has one final point

play05:51

about the brain is that the cortex is

play05:53

modular or at least it learns to be

play05:55

modular different bits of the cortex end

play05:58

up doing different things genetically

play06:01

the inputs from the senses go to

play06:04

different bits of the cortex and that

play06:06

determines a lot about what they end up

play06:07

doing if you damage the brain of an

play06:11

adult local damage to the brain causes

play06:15

specific effects damage to one place

play06:17

might cause you to lose your ability to

play06:20

understand language damage to another

play06:22

place might cause you to lose your

play06:23

ability to recognize objects we know a

play06:28

lot about how functions are located in

play06:30

the brain because when you use a part of

play06:32

the brain for doing something it

play06:34

requires energy

play06:36

and so it demands more blood flow and

play06:37

you can see the blood flow in a brain

play06:40

scanner so that allows you to see which

play06:42

bits of the brain you are using for

play06:43

particular tasks but the remarkable

play06:47

thing about cortex is it looks pretty

play06:49

much the same all over and that strongly

play06:51

suggests that it's got a fairly flexible

play06:55

Universal learning algorithm in it

play06:57

that's also suggested by the fact that

play06:59

if you damage the brain early on

play07:01

functions will relocate to other parts

play07:03

of the brain so it's not genetically

play07:06

predetermined at least not directly

play07:08

which part of the brain will perform

play07:10

which function this convincing

play07:12

experiments on baby ferrets that show

play07:16

that if you cut off the input to the

play07:19

auditory cortex it comes from the ears

play07:21

and instead reroute the visual input to

play07:24

auditory cortex than the auditory cortex

play07:27

that was destined to deal with sounds

play07:29

will actually learn to deal with visual

play07:32

input and create neurons that look very

play07:35

like the neurons in the visual system

play07:40

this suggests the cortex is made of

play07:42

general-purpose stuff that has the

play07:44

ability to turn into special purpose

play07:46

hardware for particular tasks in

play07:47

response to experience and that gives

play07:50

you a nice combination of rapid parallel

play07:53

computation once you've learned plus

play07:55

flexibility so you can put you can learn

play07:58

new functions so you're learning to do

play08:02

the parallel computation it's quite like

play08:04

an FPGA where you build some standard

play08:07

parallel hardware and then after it's

play08:09

built you put in information that tells

play08:13

it what particular parallel computation

play08:15

to do conventional computers get their

play08:18

flexibility by having a stored

play08:20

sequential program but this requires

play08:22

very fast central processors to access

play08:25

the lines in the scheduled program and

play08:27

perform long sequential computations

Rate This

5.0 / 5 (0 votes)

Related Tags
神经科学人工智能并行计算学习算法大脑结构认知功能神经网络模拟实验计算效率模块化学习生物启发
Do you need a summary in English?