Week 2
Week 2
Network Topology
A network topology is the arrangement of a network along with its nodes and connecting
lines. According to the topology, ANN can be classified as the following kinds −
Feedforward Network
It is a non-recurrent network having processing units/nodes in layers and all the nodes
in a layer are connected with the nodes of the previous layers. The connection has
different weights upon them. There is no feedback loop means the signal can only flow
in one direction, from input to output. It may be divided into the following two types −
Single layer feedforward network − The concept is of feedforward ANN having
only one weighted layer. In other words, we can say the input layer is fully
connected to the output layer.
Feedback Network
As the name suggests, a feedback network has feedback paths, which means the signal
can flow in both directions using loops. This makes it a non-linear dynamic system,
which changes continuously until it reaches a state of equilibrium. It may be divided into
the following types −
Recurrent networks − They are feedback networks with closed loops. Following
are the two types of recurrent networks.
Fully recurrent network − It is the simplest neural network architecture because
all nodes are connected to all other nodes and each node works as both input
and output.
Jordan network − It is a closed loop network in which the output will go to the
input again as feedback as shown in the following diagram.
Reinforcement Learning
As the name suggests, this type of learning is used to reinforce or strengthen the
network over some critic information. This learning process is similar to supervised
learning, however we might have very less information.
During the training of network under reinforcement learning, the network receives some
feedback from the environment. This makes it somewhat similar to supervised learning.
However, the feedback obtained here is evaluative not instructive, which means there
is no teacher as in supervised learning. After receiving the feedback, the network
performs adjustments of the weights to get better critic information in future.
Figure 3: Different Training methods of Artificial Neural Network
Every input pattern that is used to train the network is associated with an output pattern
which is the target or the desired pattern.
5
1.3.1.2 Unsupervised learning
In this learning method the target output is not presented to the network.It is as if there
is no teacher to present the desired patterns and hence the system learns of its own by
discovering and adapting to structural features in the input patterns.
In this method, a teacher though available, doesnot present the expected answer but
only indicates if the computed output correct or incorrect.The information provided helps the
network in the learning process.
It is computed as
Here is the transposeof the associated output vector .Numerous variants of the rule have been
proposed.
This is based on the minimization of error E defined in terms of weights and activation
function of the network.Also it is required that the activation function employed by the network
is differentiable, as the weight update is dependent on the gradient of the error E.
Thus if ∆ is the weight update of the link connecting the ℎ and ℎ neuron of the two neighbouring layers, then ∆ is defined as,
∆ =ɳ
In this method, those neurons which respond strongly to input stimuli have their
weights updated.
6
When an input pattern is presented, all neurons in the layer compete and the winning
neurons undergoes weight adjustment.Hence it is a winner-takes-all strategy.
The different learning laws or rules with their features is given in Table1 which is given
below
Table 1: Different learning laws with their weight details and learning type
7
1.4 TYPES OF ACTIVATION FUNCTIONS
Linear functions are simplest form of Activation function.Refer figure 4 . f(x) is just an
identity function.Usually used in simple networks. It collects the input and produces an output
which is proportionate to the given input. This is Better than step function because it gives
multiple outputs, not just True or False
1.4.2. Binary Step Function (with threshold) (aka Heaviside Function or Threshold
Function)
1 if x
f (x)
0 ----------- eq (4)
if x
Binary step function is shown in figure 4. It is also called Heaviside function. Some
literatures it is also known as Threshold function. Equation 4 gives the output for this function.
8
1.4.3. Binary Sigmoid
9
1.5 PERCEPTRON MODEL
1.5.1 Simple Perceptron for Pattern Classification
----------- eq (7)
Equation 7 gives the bipolar activation function which is the most common function used
in the perceptron networks. Figure 7 represents a single layer perceptron network. The inputs
arising from the problem space are collected by the sensors and they are fed to the aswociation
units.Association units are the units which are responsible to associate the inputs based on their
similarities. This unit groups the similar inputs hence the name association unit.
10
A single input from each group is given to the summing unit.Weights are randomnly fixed
intially and assigned to this inputs. The net value is calculate by using the expression
This value is given to the activation function unit to get the final output response.The
actual output is compared with the Target or desired .If they are same then we can stop training
else the weights haqs to be updated .It means there is error .Error is given as δ = b-s , where b
is the desired / Target output and S is the actual outcome of the machinehere the weights are
updated based on the perceptron Learning law as given in equation 9.
Step 1: Initialize weights and bias.For simplicity, set weights and bias to zero.Set
learning rate in the range of zero to one.
11
1.5.3 Multi-Layer Perceptron Model
Figure 8 is the general representation of Multi layer Perceptron network.Inbetween
the input and output Layer there will be some more layers also known as Hidden layers.
12
1.6 LINEARLY SEPERABLE & LINEAR IN SEPARABLE TASKS
Perceptron are successful only on problems with a linearly separable solution sapce. Figure
9 represents both linear separable as well as linear in seperable problem.Perceptron cannot handle,
in particular, tasks which are not linearly separable.(Known as linear inseparable problem).Sets of
points in two dimensional spaces are linearly separable if the sets can be seperated by a straight
line.Generalizing, a set of points in n-dimentional space are that can be seperated by a straight
line.is called Linear seperable as represented in Figure 9.
Single layer perceptron can be used for linear separation.Example AND gate.But it cant
be used for non linear ,inseparable problems.(Example XOR Gate).Consider figure 10.
13
Here a single decision line cannot separate the Zeros and Ones Linearly.At least Two
lines are required to separate Zeros and Onesas shown in Figure 10. Hence single layer
networks can not be used to solve inseparable problems. To over come this problem we go for
creation of convex regions.
Convex regions can be created by multiple decision lines arising from multi layer
networks.Single layer network cannot be used to solve inseparable problem.Hence we go for
multilayer network there by creating convex regions which solves the inseparable problem.
Select any Two points in a region and draw a straight line between these two points. If
the points selected and the lines joining them both lie inside the region then that region is
known as convex regions.