2. Three Basic Components or Entities of Artificial Neural Network Introduction | Soft Computing

Mahesh Huddar
8 Oct 202209:31

Summary

TLDRThis video delves into the three fundamental components of artificial neural networks (ANNs): synaptic interconnections, learning rules, and activation functions. It explains various network models, including single and multi-layer feed-forward, single node with feedback, and single and multi-layer recurrent networks. The video also discusses learning methods, such as parameter and structural learning, and covers several activation functions like identity, binary step, bipolar step, ramp, and sigmoidal functions, which are crucial for ANN operations.

Takeaways

  • 🧠 The models of artificial neural networks are specified by three basic components: synaptic interconnections, learning rules, and activation functions.
  • 🔗 Synaptic interconnections can be of five types: single layer feed forward, multi-layer feed forward, single node with feedback, single layer recurrent, and multi-layer recurrent networks.
  • 🌐 In a single layer feed forward network, each neuron in the input layer is directly connected to the output layer neurons.
  • 🌀 Multi-layer feed forward networks include one or more hidden layers between the input and output layers.
  • 🔁 Single node with feedback involves sending back the output as feedback to modify the computation of the input neuron for better output.
  • 🔄 Single layer recurrent networks update weights considering both the current input and the previous output.
  • 🎓 Learning in ANNs is categorized into parameter learning, which updates connection weights, and structural learning, which changes the network's structure.
  • 📈 Activation functions determine the output of a neuron based on its input. Common functions include identity, binary step, bipolar step, ramp, and sigmoidal functions.
  • 📊 The sigmoidal function, particularly the binary sigmoid function, is widely used in backpropagation neural networks due to its desirable mathematical properties.
  • 🔍 The derivative of the binary sigmoid function is key for backpropagation as it helps in calculating the gradient during the learning process.
  • 🔑 The video aims to clarify the concepts of the components of artificial neural networks and encourages viewers to engage with the content for further learning.

Q & A

  • What are the three basic components of an artificial neural network?

    -The three basic components of an artificial neural network are the synaptic interconnections, the training or learning rules, and the activation functions.

  • What are the different types of interconnections in artificial neural networks?

    -The different types of interconnections in artificial neural networks include single layer feed forward networks, multi-layer feed forward networks, single node with its own feedback, single layer recurrent networks, and multi-layer recurrent networks.

  • How does a single layer feed forward network differ from a multi-layer feed forward network?

    -A single layer feed forward network has only an input layer and an output layer, whereas a multi-layer feed forward network includes one or more hidden layers in addition to the input and output layers.

  • What is the purpose of a single node with its own feedback in a neural network?

    -In a single node with its own feedback, the output of the input neuron is used as feedback to validate and modify the computation of the input neuron to improve the output.

  • How does a single layer recurrent network differ from a multi-layer recurrent network?

    -A single layer recurrent network updates weights considering the previous iteration's outputs along with the current input, while a multi-layer recurrent network has multiple layers, including hidden layers, and updates weights similarly across these layers.

  • What are the two main types of learning in artificial neural networks?

    -The two main types of learning in artificial neural networks are parameter learning, which involves updating connection weights, and structural learning, which focuses on changing the network structure such as the number of processing elements or layers.

  • What is an identity function in the context of activation functions?

    -An identity function is an activation function where the output is equal to the input, defined as f(x) = x, and it operates linearly on the input values.

  • What is a binary step function and how does it work?

    -A binary step function is an activation function that outputs 1 if the input value is greater than or equal to 0, and 0 if the input value is less than 0, resulting in a binary output range.

  • What is a bipolar step function and how does it differ from a binary step function?

    -A bipolar step function outputs 1 if the input value is greater than or equal to 0, and -1 if the input value is less than 0, providing a bipolar output range, unlike the binary step function which only outputs 0 or 1.

  • What is the range of output for the RAM function in activation functions?

    -The RAM function outputs 1 for input values greater than 1, linearly increases from 0 to 1 for input values between 0 and 1, and outputs 0 for input values less than 0.

  • How is the binary sigmoid function defined and what is its range of output?

    -The binary sigmoid function is defined as f(x) = 1 / (1 + e^-λx), where λ is the stiffness parameter and x is the summation term. Its output ranges from 0 to 1.

  • What is the derivative of the binary sigmoid function and what does it represent?

    -The derivative of the binary sigmoid function is f'(x) = λ * f(x) * (1 - f(x)), which represents the rate of change of the function and is used in backpropagation for learning.

Outlines

00:00

🧠 Fundamental Components of Artificial Neural Networks

This paragraph introduces the three fundamental components of artificial neural networks (ANNs): synaptic interconnections, learning rules, and activation functions. It explains the various models of ANNs based on these components. The synaptic interconnections are detailed with examples of different network architectures, such as single-layer feedforward, multi-layer feedforward, single node with feedback, single-layer recurrent, and multi-layer recurrent networks. The paragraph also briefly touches on the learning aspect of ANNs, differentiating between parameter learning (adjusting connection weights) and structural learning (changing network structure).

05:02

📈 Activation Functions in Neural Networks

This paragraph delves into the various types of activation functions used in ANNs, which are crucial for determining the output of neurons. It describes the identity function, binary step function, bipolar step function, ramp function, and sigmoidal functions. Each function is defined and explained with its mathematical representation and graphical depiction. The paragraph highlights the importance of these functions in shaping the behavior of neural networks, particularly in backpropagation networks, where derivatives of these functions play a significant role in learning.

Mindmap

Keywords

💡Artificial Neural Network (ANN)

An Artificial Neural Network is a computational model inspired by the way biological neural networks, such as the human brain, process information. It is composed of interconnected nodes or 'neurons' that process information through a connectionist system. In the video, ANNs are discussed as models specified by three basic components, highlighting their complexity and adaptability in learning from data.

💡Synaptic Interconnections

Synaptic interconnections refer to the way neurons are connected to each other within a neural network. These connections are crucial as they determine the flow of information and the network's overall architecture. The video script mentions five types of interconnections, including single-layer feedforward and multi-layer recurrent networks, which are fundamental to understanding how ANNs are structured and function.

💡Feedforward Network

A feedforward network is a type of ANN where the connections between the neurons do not form a cycle. Information moves in one direction, from the input layer, through hidden layers, to the output layer. The video describes single-layer and multi-layer feedforward networks, emphasizing the straightforward flow of data which is a key feature of these networks.

💡Recurrent Network

Recurrent networks are a class of ANNs where connections between neurons form a cycle, allowing information to be 'passed back' through the network. This type of network is capable of 'remembering' previous inputs, which is crucial for tasks involving sequences or time-series data. The video script explains single-layer and multi-layer recurrent networks, showcasing their dynamic nature compared to feedforward networks.

💡Learning Rules

Learning rules in ANNs refer to the algorithms or methods used to update and adjust the connection weights within the network. These rules are essential for the network's ability to 'learn' from data. The video mentions parameter learning, which involves updating weights using logic like gradient descent, and structural learning, which focuses on changing the network's structure.

💡Activation Functions

Activation functions are mathematical equations that determine the output of a neuron given an input or set of inputs. They introduce non-linear properties to the network, allowing it to model complex relationships. The video script discusses various activation functions, such as the identity function, binary step function, and sigmoid functions, illustrating their role in shaping the network's output.

💡Identity Function

The identity function is a type of activation function where the output is equal to the input. It is defined as f(x) = x. This function is used in ANNs to maintain the linearity of the input-output relationship. The video script uses the identity function as an example to explain the simplest form of an activation function.

💡Binary Step Function

The binary step function is an activation function that outputs a binary value (0 or 1) based on whether the input is above or below a certain threshold. This function is used to introduce binary decision-making capabilities within a neural network. The video script describes how this function can be used to create a clear distinction between active and inactive states in neurons.

💡Sigmoid Function

The sigmoid function, also known as the logistic function, is an S-shaped activation function that outputs values ranging between 0 and 1. It is widely used in ANNs, particularly in backpropagation networks, to introduce non-linearity while maintaining a smooth gradient. The video script explains both the binary and bipolar forms of the sigmoid function, highlighting their use in different types of neural networks.

💡Backpropagation

Backpropagation is a learning algorithm used in ANNs to adjust the weights of the network by minimizing the difference between the predicted output and the actual output. It involves the calculation of gradients and the use of these gradients to update the weights in the opposite direction of the information flow. The video script mentions backpropagation in the context of sigmoid functions, emphasizing its importance in training neural networks.

Highlights

The video discusses the three basic components of artificial neural networks.

The first component is the model's synaptic interconnections.

There are five types of interconnections: single layer feed forward, multi-layer feed forward, single node with feedback, single layer recurrent, and multi-layer recurrent networks.

A single layer feed forward network has a direct connection between the input and output layers.

A multi-layer feed forward network includes hidden layers between the input and output layers.

A single node with its own feedback validates the output and adjusts computations if necessary.

In a single layer recurrent network, weights are updated considering both the input and previous outputs.

A multi-layer recurrent network has multiple layers with recurrent connections.

The second component is training or learning rules for updating connection weights.

There are two types of learning: parameter learning and structural learning.

Parameter learning updates connection weights using algorithms like gradient descent.

Structural learning focuses on changing the network's structure, such as the number of neurons or layers.

The third component is activation functions.

Identity function is an activation function where the output equals the input.

Binary step function is an activation function that outputs 1 if the input is greater or equal to 0, otherwise 0.

Bipolar step function outputs 1 if the input is non-negative and -1 if it is negative.

The ramp function outputs the input value if it's between 0 and 1, and 1 if it's greater than 1.

Sigmoidal functions are widely used in backpropagation neural networks.

Binary sigmoid function outputs a value between 0 and 1 and is defined by a logistic equation.

Bipolar sigmoid function outputs a value between -1 and 1 and has a specific derivative form.

The video aims to clarify the concepts of artificial neural networks' components.

Encouragement to like, share, subscribe, and turn on notifications for more content.

Transcripts

play00:00

hi welcome back in the previous video I

play00:03

have discussed what is artificial neural

play00:06

network and how it works

play00:08

in this video I will discuss what are

play00:10

the three basic components of artificial

play00:12

neural network

play00:15

the models of artificial neural networks

play00:18

are specified by three basic components

play00:21

or entities

play00:24

the first one is the models synaptic

play00:27

interconnections

play00:28

second one is the training or learning

play00:31

rules adopted for updating and adjusting

play00:35

the connection weights

play00:36

third one is the activation functions we

play00:40

will try to discuss each of these

play00:42

components one by one

play00:43

the first one is the connections

play00:46

in artificial neural network we have

play00:48

multiple number of neurons

play00:50

interconnected with one another in

play00:53

multiple number of layers like input

play00:56

layer hidden layer and output layer and

play00:58

so on

play01:00

there exists basically five types of

play01:04

interconnections

play01:06

the first one is a single layer feed

play01:08

forward Network

play01:10

second one is multi-layer feed forward

play01:12

Network

play01:13

third one is single node with its own

play01:15

feedback

play01:17

fourth one is single layer recurrent

play01:19

Network fifth one is multi-layer

play01:21

recurrent Network

play01:24

this is how the single layer feed

play01:25

forward Network looks like in this case

play01:28

we have two layers the first one is the

play01:30

input layer and the second one is the

play01:32

output layer the input layer has one or

play01:35

more neurons and output layer can have

play01:37

one or more neurons in this case

play01:40

and each of the neurons are present in

play01:42

the input layer are directly connected

play01:44

with the output layer neurons that is X1

play01:48

is connected to y1 Y2 and Y M in this

play01:51

case similarly X2 is connected to y1 Y2

play01:55

ym and so on

play01:57

so this is also called as fully

play01:59

connected single layer feed forward

play02:01

Network in this case the input layer

play02:04

receives the inputs from the input input

play02:07

and then that input is forwarded to the

play02:09

output layer and the output will be

play02:11

received at this particular layer

play02:14

this is how the multi-layer field

play02:16

forward Network looks like it has input

play02:19

layer and output layer along with input

play02:22

and output layer it has one or more

play02:23

hidden layers in this case

play02:25

the input layer has one or more input

play02:28

neurons the output layer has again one

play02:31

or more output neurons based on the

play02:33

problem what we are solving and hidden

play02:36

layer can have again one or more the

play02:38

neurons in this case

play02:40

the input layer and neurons are

play02:42

connected to the hidden layer neurons

play02:43

and the hidden layer neurons are

play02:45

connected to Output layer neurons in

play02:47

this case

play02:48

coming back to the third one that is a

play02:50

single node with the own feedback in

play02:53

this case we have only one layer that is

play02:54

the input layer and the output of input

play02:58

neuron is nothing but the output here

play03:01

so uh we we want to validate whether

play03:05

this particular output is acceptable or

play03:08

not if it is acceptable it's okay if it

play03:10

is not acceptable we will send back the

play03:12

feedback and then we will modify the

play03:14

computation part of this particular

play03:16

input neuron so that we will get the

play03:18

better output in this case

play03:20

the next one is uh the single layer

play03:23

recurrent Network in this case we have

play03:25

input layer and output layer

play03:27

while modifying or updating this

play03:30

particular weights rather than just

play03:32

considering the input input as well as

play03:36

the previous weights we consider the

play03:39

outputs as well in this case the output

play03:42

of the previous iteration and then the

play03:45

input will be considered to update this

play03:48

particular weights over here so that is

play03:49

what is called a single layer recurrent

play03:51

Network

play03:52

multi-layer recurrent network has

play03:54

similar architecture but we have what is

play03:56

that called as multiple number of layers

play03:58

in this case along with input and output

play04:00

layer we have what is that called as the

play04:02

hidden layer also in this case

play04:04

so this these are the different type of

play04:07

connections what we can have in the

play04:10

artificial neural network

play04:12

coming back to the second component that

play04:14

is the learning

play04:15

the main part of artificial neural

play04:17

network is its capability to learn

play04:19

through the past experience

play04:23

there are mainly two kind of learnings

play04:25

in artificial neural network the first

play04:27

one is called as parameter learning and

play04:29

second one is called as structural

play04:31

learning parameter learning is nothing

play04:34

but here we update the connection

play04:36

weights with some Logic for example we

play04:40

use either gradient descent or backward

play04:42

protection algorithm and we update this

play04:44

particular bits that is what is called

play04:46

parameter learning structure learning

play04:49

focuses on the change in network

play04:51

structure which includes the number of

play04:53

processing elements or the number of

play04:55

layers and so on so that is nothing but

play04:57

the structure learning the third

play04:59

component of the artificial neural

play05:02

network is the activation functions

play05:04

there are several activation functions

play05:06

exists one is the identity function

play05:10

this identity function is defined like f

play05:12

of x is equal to X

play05:14

here x is nothing but the summation term

play05:17

I think we have already discussed in the

play05:19

previous video like at each neuron there

play05:22

are two things takes place one is called

play05:24

as summation and another one is called

play05:26

as applying the activation function

play05:29

so once you get the summation term on

play05:31

the top of summation term we apply the

play05:32

activation function here so once you

play05:34

apply the activation function we if you

play05:37

get the same value that is nothing but

play05:38

the identity function here so as and

play05:41

when the value of X increases the f of x

play05:43

also increases with the same value here

play05:47

it's a linearly you can say

play05:49

the second one is a binary step function

play05:51

which is defined as something like this

play05:53

if the value of x is greater than

play05:55

equivalent to 0 the value is 1 if the

play05:57

value of x is less than 0 it is 0 in

play06:00

this case

play06:01

the value of this particular function

play06:04

will range uh will be either 0 or 1 in

play06:07

this case

play06:07

that is shown in this particular graph

play06:10

over here

play06:11

the third activation function is the

play06:13

bipolar step function

play06:16

which is defined as something like this

play06:18

if the value of x is greater than or

play06:19

equal to 0 it is 1 if the value of x is

play06:22

less than 0 it is minus 1 so there are

play06:25

two possible outputs are there one is

play06:26

plus one another one is minus one in

play06:28

this case if the value of x is greater

play06:31

than 0 it is 1 if it is value of x is

play06:34

less than 0 it is minus 1 in this case

play06:39

the next activation function is the ram

play06:41

function which is defined something like

play06:43

this if the value of x is greater than 1

play06:46

it is 1 you can see here from here

play06:48

onwards the value of x is greater than

play06:49

one so as and when the value of x is

play06:52

greater than 1 the f of x is equal to 1

play06:54

here that is noted here

play06:56

if the value of x is in the range of 0

play06:58

to 1 the f of x is equal to X so as and

play07:02

when the value of X increases the f of x

play07:05

increases linearly that is f of

play07:09

0.5 is equal to 0.5 here so similarly it

play07:13

will be increased increased in linearly

play07:16

if the value of x is less than 0 the f

play07:19

of x is equivalent to 0 here that is the

play07:22

x is less than 0 f of x is equal to 0 in

play07:25

this case

play07:27

coming back to the last activation

play07:28

function that is called as sigmoidal

play07:31

functions they are widely used in Back

play07:33

propagation artificial neural networks

play07:36

there are two types of sigmoidal

play07:38

functions are there one is called as

play07:40

binary sigmoid function another one is

play07:43

called as bipolar sigmoid function

play07:46

binary Sigma function is defined

play07:48

something like this that is f of x is

play07:50

equivalent to 1 divided by 1 plus e

play07:53

raised to minus Lambda X where Lambda is

play07:56

the stiffness parameter and X is the

play07:59

summation term here

play08:00

and it is also called as the logistic

play08:02

sigmet function or unipolar sigmoid

play08:05

function

play08:06

The Peculiar property of this by binary

play08:09

Sigma function is when you do the

play08:11

derivative or when you take the

play08:12

derivative of f of x that is f Dash X is

play08:15

equal into Lambda that is a stiffness

play08:17

curve stiffness value multiplied by f of

play08:21

x into 1 minus f of x that is what we

play08:23

will get when you do the derivative of

play08:25

this particular sigmoid function in this

play08:28

case and the value of f of x will range

play08:31

in the range from 0 to 1 in this case

play08:34

bipolar sigmoid function is defined

play08:37

something like this that is f of x is

play08:39

equivalent to 2 divided by 1 plus e

play08:41

raised to minus Lambda X minus 1 in this

play08:44

case which is equivalent to 1 minus E

play08:47

raised to minus Lambda x divided by 1

play08:49

plus C raised to minus Lambda X here

play08:52

uh the value of this particle f of x

play08:54

will range in minus 1 to plus 1 and the

play08:57

derivative of this particle f of x is

play08:58

nothing but Lambda divided by 2

play09:01

multiplied by 1 plus f of x multiplied

play09:04

by 1 minus f of x over here

play09:07

so in this video I have discussed uh

play09:09

what are the different components of

play09:11

artificial neural network that is the

play09:13

first one is connection second one is

play09:15

the learning and third one is the

play09:18

activation functions

play09:20

I hope the concept is clear if you like

play09:22

the video do like and share with your

play09:24

friends press the Subscribe button for

play09:26

more videos press the Bell icon for

play09:28

regular updates thank you for watching

Rate This

5.0 / 5 (0 votes)

関連タグ
Artificial IntelligenceNeural NetworksMachine LearningDeep LearningAI ComponentsBackpropagationSigmoid FunctionsLearning AlgorithmsConnection WeightsNeuron Models
英語で要約が必要ですか?