NN1 PDF
NN1 PDF
Net wor ks
1.1 Neural Processin g
Artificial neural networks are the result of academic investigation s that use mathematical formulations
to model nervous system operations. The resulting techniques are being successfu lly applied in a variety
of everyday business applications.
Neural networks (NNs) represent a meaningfully different approach to using computers in the \vork-
place. A neural network is used to learn patterns and relationships in data. The data may be the resl!lts of
2 !111mductw11 to Neural N<'fll'or/...,
of te n ca
·\-. Th l', ca n irn d la ye re d_'~.e ~-
lle
p\erncnt :lrb1tt d t
on s.
y
.. X
x2(t)
•
- Y2(t), i y
('(,m p'1e· "- in • pu t/o ut ~~~s 0 r dc - \ Xn(t) '- - '
pu t · o •
.... , ~11..111 ~mfa
rn ap ~,
cc ~ paratmg. d1
~e Yn~t)
tern :-- A th re e- la tte re nt pa~-
) er M N N is sh ow
Fi ~ 1. 2. and a ~i n m
mphfied bl ock rii
agram ~ ~ ~
repre~ em at io n in In pu
Fi g. 1.3 . ln a M . a t La yer Hidden La
la\ er of in pu t un N N ye r O ut pu t La
its is co nn ye r
la~er of hi dd en un ec te d to a
its. which is conn
lo· the la, er of ou ec te d A Densely Intercon
tput units . Th e ac nected Th re e- la
of neur~n~ tivity Fig. 1. 2 ye re
m th e in pu t la ye Ne ur al Ne tw ork. Each Sh d St at ic
the ra~ in form at r represen ts ad ed
io n that Circle , or Node
ne n, or k. Th e ac is fe d in to th e \ Re pr es en ts an Artificial Ne ur ,
tiv ity of neurons in on
hi dd en la ve r i~ de the
te nn ined by th e ac
tie<; of th e in tiv i-
pu t ne uron s and N1
ne ct in g "v.eigh ts th e co n- N2 N3 y_
be tw ee n the inpu
hi dd en un
t and
iu .. Si m ila rly, th e be ha vi or In pu t La yer Hidden La
the ou tp ut un of ye r O ut pu t La
its de pe nd s on the activ ity \ ye r
of th e ne ur on s . _
in th e hi dd en la A Bl oc k D ia gr am R ep re
th e co nn ectin g w ye r an d Fig. 1·3 Thre se nt at 10 n of a
eights be tw ee n th e-layered MNN
de n an d e hi d-
th e ou tp ut
are fre e to com.tr layers. Th is simple neural structu
uct their ow n re pr re is in te re st in g
es en ta tio n of th e input. be ca us e ne ur on s in th e hi
MNN s pr ov id e dd en layers
an in crease in co m pu ta
a no nl in ea r ac tiv tio nal po w er ov er
ation fu nc tio n be a si ng le -l ay er
tunct10na1 pr ox tw ee n la ye rs . M an y ca pa ne ur al ne tw or k un le
tion tu nc uuap im at io
n of each ne n, le ar ni ng , ge ne raliz ation,
bi lit ie s of ne ur al ne tw or ks su
ss th er e is
ur on . et c ar e in fa ct ch as no nl in ea r
pe rf or m ed du e
to th e no nl in ea
A1 ':N ,. ha ve be co m r activa-
su ftv .a re and ha
e a te chni ca l fo lk le ge
rd w ar e prod nd . The m ar ke t
uc ts, an d m is fl oo de d w ith
implementat1om an y m or e are su ne w , incr ea si ly te
ng
, ar e H op fle ld , Mu ltilaye re to co m e. A m on g th e ch ni :
Q ua nu 1.a t1o n. Ra r Pe rc ep tro n, m os t po pu la r ha rd w
dial l\a ,is Functio Se lf- or gani zi ng
Co un te r Pr op ag n, C ellular N eu ra Fe at ur e M ap . Le
au on ne tw or l, an d A da pt iv e R am in g V ecto r
en ce oJ all 1ile,e k~ , Back Propagati es on an ce T he or y
ne 1work1,, the ap on ne tw or ks , N eo (AR T) networks
pl ic at io n of -c og ni tro n . et c. .
th e ne ur al A s a re su lt of th e
eraTh u, a11Ific1al neu ral ne1w ne tw or k is in creasi ng
tre m en do us ly .
eX tS -
tion -. ork re prese nis th t
'>ll11 1la 1 \o th at ot1 th e hu m a
e m aj or ex te ns io n
un derst n hr ur n . He nc e co m pu ta tio n. T
ca tio n an opJ1 pu11 t1g u1 ar uh c1
u ni ll e~
-
a. ne ur
it is re ·ts on 10 ,. tb le to
he y pe rf or m th e op -
. al ne 1w o rk , le a,h•ng. lo i01p • •d •n cr ea se in ou
ro'vc·
li •ne tw or k ex pe ct a ra r
pa ra di gm s pt
an d 1a ho st of ap .
1. 3 pll-
T h e R is e o f N cu
ro co mputing
A m aJ·or ity of in fo rm at i on proc essing
w id el y he l d m is to day is ca n - d
pe rc ep tt. on t h at 111
- t·or . n or
m atio r,,," .. .1 e ou t h, ,
/11t roductio11 to A11ifi,cial Neural Network.\ 11
2.1 Introduction
The bas ic pre limin aries in vo lved in the /\rti t'i tia l Nc urn l Ne twork (ANN) arc de~cribed in thi s chapter.
A brie f summ ary nf the hi story o f ne ural nc lwo rk s, in te rm :-. nf Lhc deve lopment of a rchitecture.., and
algorithm s. the stru c tu re of the bi, ,log ica l ne uro n is di sc ussed a nd co mpared with th e artificial neuron .
The basic build ing bloc ks a nd the vari ous te rmin o logies of the a11ificia l neural network are ex plained
rownrds th!.! e nd o f ihe c ha pt e r. The chapter conc lu des by g ivi ng the s um mary or notation<.,, which are
used in a ll the netwo rk a lgorithm s, arc hitectures, e tc. di sc ussed in the fo rthcoming c hapter\ .
2.2 Artificial Neural Networks
~ rtific ial neural networks are nonlin ear information (sig na l) ~!ocessi_ng device'>, which are built fro~
interconnected ele me nta ry process ing de vices called Eeuro nv
An Artific ia l Neural Ne twork (ANN) is an information-processing p aradi g m that is in spired by the
way bi o log ical nervous systems .. such as the brain" process inform ation. Th e key element onS~para..:
dig m is the nove l structure of the information process ing systeni . ft is composed of a la_Ege nu mber of
hi__ghly interconnected processing elementS_ (neu~ ns) working in uni ontosol ve specifi c problems. A.i~s.
like people, learn by exampl e . An ANN is configuredTor a spec ific applicati o n, such as pattern recogni -
tion or data classification, through a learning proce_SSJ Learning in biologi cal systems involv es a ~-
ments to th e synaptic connections that exi st bdween the neuron s. This is true of ANNs as well .
ANN 's :ue a type of artificial intelligence that atte mpts to imitate the w ay a human brain ,rnrk-..
Rather than using a digital model, in which all computations marupulate zeros and ones. a neural ne t-
work works by creating connections between processing elements, the computer equivalent of neuro n!:).
The organization and weights of the connections determine the output.
A neural network is a massively parallel-distributed processor that has a natural propensity for ~rt.r-
ing experimental knowledge and making it available for use. It resem~les the brain in rwo re~peL"t::-.:
1. Knowledge is acquired by the network through a learning process, and.
2. Inter-neuron connection strengths known as synaptic weights are used to store the knov, ledge
Neural networks can also be defined as parameterized computational nonlinear algorith m~ fr,r nll -
merical) data/signal/image processing. These algorithm s are either implemented on a general-pu.rpt.',e
computer or are built into a dedicated hardware.
Artificial Neural r•:etworks thus is an information-process ing syste m . In this inforn1ation-prnL°e"~ •1g
system , the elements called as neurons, process the information. The sign a ls are transmitted b~ mean~ o f
connection links. The links possess an associated we ig ht, whi ch is multiplied along\\ ith the rnconung
signal (net input) for any typical neural net. The output s ig nal is o btained hy applying acti, ation:- tu the
net input.
An artificial ne uron is characterized by :
1. Architecture (connection between neuron s)
2. Trainin g or learning (de te rmining we ights o n rhc connectio ns)
3. Activation fun cti o n
,. \ , \1, •I ~ I
J: hrr. , •fu, -n,,r: • ' • ' x 1- ( ~ w1 (weights)
,r fl.irthro1ni11g
.., kt\\ l ll I tt I
-\': r''c'' c' 3~' J,,~'tl'-''C".l',,t. th'
\ ' l' ' 1itid :d 11~ura
, 11\ll' l ,ll .$
:::,
:
I X1 :: -------- .............. ,----\
\ y
, u~, c'\.' tlf\n, Th' , mi,'tu"\: , \. .'
: I
0. :
•. E l -. ___ )l (O UtpU\)
~ . tl n\'ural net'',orr,.
I II
nc'f\\,,r\ ;, , 1\'1\\11 Hi F·f - II I
Fi.£ur<'
~
~ ! , fh."' ' J , utq' k ,uul\l.) J t,)l\t' LHl!pll t I1t:·uron
x2 1
3 1"- , W2 (weights)
Output L
ayer
\ , d \ \I 1 •
~
\\ 1th i " \) JflfUl (K'Ul\.'Il~ \ \ · \ : Jl\, ,•1H'll ln' ,,·, ,111 ,_ .,.,I
l ~ ). Th<' tnt<'r ,-,,nnc'('tc<l "tlf h~ ,U\' .:-
. ,[ . ,,utpU! ..
· ,1,rtkl, I-pro-_
.\n 3.rtlficial neuron 1~ ar-mput sn~ t - , ' ;;: . x2
t. •\s ·1 simple
ce.--..-,ing element. ,, hich ,·an ti,;:' tIwug lH O ' ' • . 2 1 Input Layer
«ll"ldel of a non-branchin £ bi\)h}fi( ~tl neumn . In Fig · ·
,.m-ous inputs to the nerwZ,rk Mt' n:>pR'St'_nted by the mat~-
2
Fl.g. •1
. Artificial Neural Net
I A Simple
ernatical s, mbol. \ tnt Each ,,f thest' inputs :u-e rnultt -
phed by •.~om"'--0'"' ,-,,gtu. Th<S< ,veights are represented by w(n). In the simplest case, these Prod·
ucts are ~1'.11-pl~ :-umm~. ~~ d~ugh a_trru~ster functJo~ to generate a resu~t, and then delivered as
oucpuc. !11~s pr<X"c....;.s lends !csd_t w ph~·s1l·:1l ~mplementat10n on a large scale _m a small package. .
electromc 1mplemcnrnoon is soil possi ble wnh other network structures, which utilize diffe rent sum.
This
mmo funcu·on~- as- "c II as dift.erem rranskr · · •
;;: fu nctions.
Why Artificial. Neural Net1u•orks ?
The long course of e,·olution has given the human brain many desirable characteris tics not present in
Von Neumann or modern parallel computers. These include
• Massive parallelism.
• Distributed representation and computation ,
• Leaming ability.
• Generalization ability.
• Adaptivity,
• Inherent contextual infonnation processinoe
• Fault tolerance, and
• Low energy consumption.
.
Ir is hoped that devices based on b'101oo1cal
. . n I .
charactensncs. Modem digital com e. eura networks - 1 1
and related symbo l manipulat'1011 Hputers outperform huma .· Wt posses some of these deslf8~le
- owever h ns in the do · o f numenc · uon
!ems (li ke recognizing a man in a crowd f • umans can effort! 1 main computa
as to dwarf the world's fastest com put ro; a m~re glimpse of h~ss Y_ solve complex perceptual prob-
15
mance? The biological neural system ar~~- hy 1s there such . face) at such a high speed and ext~nt
1ecrure (see Table 2. 1). This difference ste~t~re is cornpletel a ;~I~arkabl e difference in their perfo~·
model can bes t perform . gnificantly affects t{
•fferent from the Von Neumann arrh1
Num erous efforts to develop "intellige ., e type of functions each computatiLllll
ture ha ve not res ulted in any general -purpnt . P~ogran1s based
. I ose inte11 · on Vo N c t ..... d architc'\'
works, ANNs are mass ive Y•parallel comput·1ng sy ,gent Pro grarns n eurnann's
1
. . . en ra112"
simple processors with many interconnection sten1s co . . · 1nsp1rect by b ' . . ur:al th'
. s. ANN nsistt 0 O of an extr
10 1og1ca 1 ne
ciples believed to be used in the human brain rnocteJ s . c .l. \~r .
· attempt to erne 1y arge nun
Use Sorn e " orgamut10 . . naf' .-r
Table 2.1 Von Neum ann Comp uter Versus Biolog ical Neura
l System
Von Ne11111 an11 Riolog ;caf
Co111p ura Ne11 rol System
Proces sor c, ,mpl cx Simpl e
I ligh speed Low :-. peed
One or a few /\ large numhe r
Memo ry Separate from a proccs 1-or Integrated in to
Locali zed Proccs<;or
Nnn cnntent addre:-. :-.ab lc D1stnhut cd
Conte nt addre..,..,ahl e
Computing C1,;. ntrali zed Distrihutcd
Seque ntial Parall d
Stored progra m s Se lf-learn ing
Rali ability Ve ry vulner ab le lfobu st
Expertise Nume rical and sy mbo lic Percep tua l
m anipul ations proble ms
Opera ting Well-d efined , Poorly define d,
enviro nment well-c onstra ined uncon straine d
Either human s or other comp uter techni ques can use
neural netwo rks, with thein emark able ab1bry to
deri ve meani ng from comp licate d or impre cise data,
to extrac t patter ns and detect trends that a:e too
compl ex to be notice d. A traine d neura l netwo rk can
be t~ugh t of as an --expe rt.. in the catego ~ of
inform ation it has been given to analyz e. This exper t
can then be used to provid e proj ection s gi ven ne~
situati ons of intere st and answe r "what if' questi ons.
Other advan tages includ e:
1. Adapt ive Leaming: An ability to learn how to do
tasks based on the data giYe n for trainin g or
initial exper ience.
2. Self-organization: An ANN can create its own organ
isation or repres entation of the infom uuur: l l
receiv es during learni ng time.
3. Real-time operation: Al'.TN compu tation s may be
carrie d out in paralk L using spa.:i:il h m:i~, Mt'
device s desig ned and manuf acture d to take advan tage
of this capab ility.
4 . Fault tolera nce via redundant information codin g: Partia l destru ction
of a ne twork. k,.1Js h.' ..1
corres po nding degra dation of perfor mance . Howe ver, some netwe rk capab ilities m:1~ be n-ta.incJ
even af:er major netwo rk damaq;e due to thi s featur e
.
,_.......2.3 Historical Dev elop men t of N eural N etworks
The hi storic al devel opme nt of the ne ural netwo rks can
be traced as follo \\ s:
• ~943 - McCulloch and Pitts: start of the modern era
of nl'Ut'Ul uehH>rks
Thi s forms a logica l ca lc.: ulus of ne ural netwo rks. A
netwo rk l;onsists of sutfi~tetH mnn hc r of neu-
rons (using a simpl e mode l) and prope rl y sl.! t sy nnpt
il' C\)nnec.: tions ~nn l'1)\l\l) utt:' ~\il~ ,:ompu t~lbk
,. , ~,. \,H'i ::.d OJJ'JO tJ,,,; .i/';lg_J;1-:, •.w
"
N •111 1 ' ' " \ . . fill '' I '
111 ,,, N, ,,,,,, ' rw1ir1,,1 , ~
t,,,,,11/111
/II 111 J • av= ,w1:1 t,t,; reprtl~ •lfYJ t:,.,
,., 1111
,,I t, y 11 . HI I 91 1, h "'
, ,1' 1I111 , t rut 0 11 ,';
I JIHI I I 1 tll'fl 1 ' 11
,,. t,f neur,,r, 1'-1 t c u,rJl.•.t'"JfA ~
\ . 11111111 lt1)tll 111111 f ,,,,.., 1YJ' , -y• '~
.t , ,. ,,..
11111, f111li I
1lh ,u111.rl jll' l .,,1 h .uftW p
II Hlllll p1 ,r l/1 I •f 1,1:,11
~r,t.;t.ificd thf~'<-,h<A,J b-11 -
Ill llh' ~, , ( 1111th II I 1lf • I 11111 \ IIH 111n!tl
jlll 11 •111011 Ill i~1t•H t
, .. )';
,.. n1 ,,,; ur,,n ,► CYW:l hl'/t)y,
'1111,lrn1,1t11111 ,,1 1,1111, 1111111,,\,"11 " ' ,1 p,11tH 11h11 '101111d
10 1h1 1rt tyf w,,.
tl11t ,l,,1ld ' Ih II 111,· 1h. . Il l ,,. I Iii II ,t i. illl' ,,
1111 1111111HI 1,1, "' I , ,,
,,,, I I 1h ,1 ' ,,
I
O
If OIi . /ti 1110111/11 {)1/IJn V/tJ ' J)fC'L1l1!;Jj for
r ,,,,huvt or f . \Vl ffl /1 ,
• 1••.i" ltc-hh'._ lnhil, " l'h•• o•U'" ; : ·1'1L ul l1•1un1111' tllk ',"1 !Jilli •• , ,,n1,111
wJI; dwng rng a5, a.n 'ng;::;,-.
1
' ' ,,,111 ,1 ,mt, ,1h i11 t1f n pin•' ,11 ', '", n111u·, 11 v1t y nf ,11, ,r,l lt<'h •JI (; crcW.:
' 1 1
,
JJ b1 (,uch ch,mg t\ .
11 111 \\ 1'(111 1111 I ll
1111 111 ,, 1mh 11, h P l 1
• ., 1I ltlf .. (' lll 1 '
111
,, ~, 11 11d thiit n c ' . , '.l 'J
1
I If · t l 111111 \!(lilt 11 1' · I I iti l!i
c ,·on<.:c pt t,chind the Hebb U.lefJr(
'"'" l,•1n1,,' ' ', '• 1 "'
·h 1111 Ill l ' ll i II ,1i111wni,• p1,yd10 o,~ I I Ltr•·n"th of cc,nnect1<m bet·;,~.,;e::,
• J l1 y I )(: . ., " ~
Iii'''''',,,,,~ ,,.,. . i1 1rn1111 , 1 . I . . :.'.to
I1 • ,,, ,,w i, 1111 1t11 11 111,1H • ti. '..ft r,f the correJatJ<JT1 rr~y
,, th;11 i1 ,,~ ,, ,11·111,rn, ,11,·• 1111111
' 111 l ' • 11hr ,,, ''"
I ·11 t·1111,.;t 111 '• " ~111 ' "
'thi rn ,, \li 111,1n, , h,m 1i '11 ,, • ·••lli l'I
IIH t \, ll '-1
• The postsynaptic signals are aggregated and u·ansferTed along the dendrites to the nerve cd l b'--xi: .
• The cell body generates the output neuronal signal. a spike. which is tr~msferred along the :.1., t.")n to
the ~ynaptic terminals of other neurons.
• The frequency of firi ng of a neuron is proponionaJ to the totaJ synaptic acriYities ~md is \.-ontrolkd
b) the ~ynaptic parameters (weights).
• The P) ramidal cell can receive 104 sy naptic inputs and ir can fan -our the output sign:tl f l) thl)U!-ands
of target cells - a connecti vi ry difficult to ac hieve in the artificial neuml nct\h)rks.
In general the function of the main elements can be given :ts,
Dendrite Rece ive~ !> ignal s from other neurons
Soma Sum s all the incoming signal s
Axon When a pani rnlar :111wunt of input is rt~ceh c.~d , tlwn rh~ 1..' \'ll fire.~~- lt transmirs sign:il
th, ougll axon to other ce lb.
A
1fl /11tn 1tl11, tu m
10 N, ,,m ,I N('{ 1W
1'h,· 111nda111r >l'b
·" ' ar,· nc ss c, w ntal pn,cc-sinµ clement ol
nm pa ss c, a h' a ne ural ne tw
,,1h('1 , c,\1r ·' W µc ncni l ca ork is a ne ur
, , , . "' ""1·
'" "' ' them . pa hi lities . B as ic on
ih ,·n nlllt" ' ,ome way, pe 1· all y, a bi olo g1 . T hi s bu il di ng bl
1lw 1mal resu r orrn s a ge ne ra II y . c al ne ur on re
ce iv :' k of hu
11
lt. Fi111gu re 2.2 no n 1·,n ear o pe
, how s the re la ra ti.on o n sh 1npu lll
tion sh ip o f th
es e fo ur pa rt t, fr an
s. t e re sult, ~ :
4 Parts of a
Ty pi ca l N er
ve C el l
1
I
--------...._
______ Dendrite
s: A cc ep t in pu ts
@ _,_ _ _ __ so m a: P ro ce
ss th e in pu ts
~ k' - - - -- Axon: Tu rn th
e pr oc
into ou tp ut s es se d in pu ts
& -- --
/2 \\ )\\ \, - -- - Synapses: Th e el ec
tr oc he
co nt ac t be tw m ic al
ee n ne ur on s
Fig. 2.2 \ A
Biological N
eu ro n
T he pr op er ti
es of th e bi ol
og ical ne ur on
I . Si gn ah ar po se so m e fe
e re ce iv ed by at ur es on th e
th e pr oc es sing ar ti fi ci al ne
2 . T he w ei
ght at th e rece el em en ts . T hi s el em ur on . T he y
are:
en t su m s th
ivi ng end ha s
th e ca pa e w ei
3. T he ne ur
on fires (t ra ns bi li ty to m od if y th
g h te d in pu ts
.
mit s output ), e in co m in g
4 . T he ou tp w he n su ff ic ie si gn al .
ut pr oduc ed nt in pu t is ob
fro m on e ne ur ta in ed .
5. T he pr oc es on may be tr
si ng of in fo rm an sm it te d to
at io n is fo un d to ot he r ne ur on s
6. T he w e ig
ht s can be m be lo ca l. .
od if ie d by ex
7 . N eu rn tra n~ pe ri ence.
m it le rs fo r th
e sy na ps e may
8 B oth artif be ex c it at or y
ic ial and bi o or · h "b•
lo gica l ne ur on 1n 1
r 1g.u1 e 2 ".\ a.n s have in bu 'ilt 1to ry .
ne t. d T ab le 2.2 in f'\ lt t l
d icate ho w th
e bi olog ic ·\\ ' u o e ra nc e.
ne t . l .
' 11
a ne t is as so
ci at ed w it h
th e mtificial neur 1tl
Cell B od y
Dendrites
\ \ T hr r hol d
-\
\
t
Sumrnatlon t
Fig. 2.3 , A xo n
A ss ocia tion
of Bio /ogica
/ N et With A
rt ifi ci al N et
/t//fl)(I//( f/()/'1 /(J / \rt/JU UII ,Y (' llff l/ ,'l f. 'H',, I Jfl',._\ J7
. . .0 I Networks
N1111
18 /11rrod11ct1011 m . logical Neural Network
., . ·8 1and B10
. n Between Artr,1c1
Table 2.31 Companso Biological Neural Network
Size and These do not involve as much Neural networks have large number of
Complexity computational neurons. Hence it is difficult computing elements, and the computing
to perform complex pattern recognition. is not restricted to within neurons. The
number of neurons in the brain is estima-
11
ted to about 10 and the total number
of interconnections to be around 10 15•
The size and complexity of connections
gives the brain the power of performing
complex pattern recognition tasks, which
cannot be realized on a computer.
Storage ln a computer, the information is stored in
the memory, which is addressed by its Neural networks store information in the
location. Any new information in the same st
rengths of the interconnec tions. Infor-
location destroys the old information. mati~n in the brain is adaptable, because
Hence here it is strictly replaceable. new lllformation is added by adjusting
th
e interconnection strenoths without
0 . '
Fault tolerance destroy· th .
Artificial nets are inherently not fault ing e old information .
tolerant, since the information corrupted They exh·b· 1
in the memory cannot be retrieved. . lt fault tolerance since the
tnformar 10 · •
. n is distributed in the connec-
~~ fur h h
· f i: oug out the network. Even thoug h
t tew con .
. nections are not working t e
inf'°rrnatto ·
Control distnb n is still preserved due to the
mechanism
There is a control unit, Which monitors an
the activities of computing. Utect nature of the encoded infomiation
:here is no .o
tnforrn . central control for processino
ation in th cs
based 00 . e brain . The neuron ac
r-n,,. . n locally avai·1able.
nct
a tran fue. lnfio·"uatio
Conn L, _ , srnns its ()111.,... . . -- - c
/111mrluction to /\rtifiria/ Neural Networks 1
~
J~U Ii \__~ l_~ ~)
Input Input Input
Single Layer Feedforward Multi Layer Feedforward Fully Recurrent Network Competitive Network
Output
Output
,c~ /)
?} f J ....,
• /4-(
~id_de_n___ __
Hidden
:'c J?~ ~
~~~g Copy Back
1. Supervised Training
Supervised training is the process of prov iding the netwo rk with a seri es of sample inputs and compar-
ing the output with the expected responses. The training continues until the network is able to provide
the expected response. In a neural net, fo r a sequence of trainin g input vectors there may exi st target
output vectors. The weights may then be adjusted accordin g to a learnin g algorithm . Th is process is
called supervised training.
In a logic circuit. we might have the target output as' + l ', if the necessary logic conditi on is satisfied,
or ·- 1·. if the logic condition is not satisfied . These type of logical nets are trai ned using supervised
algorithm. The same criterion is applicable for pattern classification nets also.
Supervised training is adopted in pattern association as well. If a neural net is trained to associate a set
of input Yectors with a con-esponding set of output vectors , then it is called associative memory net. If
the output is same as the input, then it forms auto-associative memory, if the output is different from the
input then it is hetero-associative.
Some of the supervised learning algorithms include Hebb net, Pattern association memory net. Back
propagation net. counter propagation net, etc.
2. Unsupervised Training
In a neural net. if for the training input vectors, the target output is not known, the training method
adopted is called as unsupervised training. The net may modify the weight so that the most similar input
vector is assigned to the same output unit. The net is found to form a exemplar or code book vector for
each cluster formed.
Unsupervised networks are far more complex and difficult to implement. It involves looping connec-
tions back into feedback layers and iterating through the process until some sort of stable recall can be
achieved. Unsupervised networks are also called self-learning networks or self-organizing networks
because of their ability to carry out self-learning. This is the method adopted in the case of self-orgmiz-
ing feature maps, adaptive resonance theory, etc. The training process extracts the statistical properties
of the training set and groups similar vectors into classes.
3. Reinforcement Training
In this method , a teac her is also assumed to be present, but the right answer is not presented to the
network. Instead, the network is only presented with an indication of whether th e output answer is right
or wrong. The net work mu st then use this information to improve its perfo rmance. Reinforct?ment learn-
ing is a very general approach to learning that can be applied when the knowledge r~quirt'd w apply
supervi sed learn ing is not availabl e. If sufficient information is available, the reinforcement kaming can
readil y handl e a specific problem. However, it is usually better to use other methods such as superYised
and unsupervised learnin g, becau se they are more direct and the ir underlying analytical basis is usually
well understood.
Reinforcement training is related to supervi sed training. T he output in thi s case may not be indicated
as the desired output, but the condition whether it is ·success ' (+ l ) or ·failure' (0) may be indicated.
Based on this, error may be calculated and the training process may be continued. The enor signal
produced from reinforcement training is found to be binary. Reinforcement learning attempts to learn
the input-output mapping through trial and error with a view to maximi ze a performance index called
the reinforcement signal. The system knows whether the output is correct or not, but does not know th e
correct ou tn11 t
H 1t11,t d11, , , ,I
110111r1 1'Nom Nr11Vork,1 .h certain (cl ass of) network top 1
1
·lo"cly connected wit a o ogy,
• 1 Js are (; ·'
!I.Ian) ,11 tlll'M' ka111111 g 11\1: t w .
. . , . so lllC names...
No" here,~ the 11~1. Jll ~t givin g · .,,
. . 'thout a ''teache1 )..
1. Unsupcn iscd tcm·nmg (1.c· w, ·
1 l) Fcl·dbad, Ne ts:
\a) Binary Adapti ve Resonance Theory (A RTI )
· . . . (ART2, ART2a)
(b) Analog Adapuve Resonance Theoiy
LI I
2.0. Wc:io·ht
I"'
s
A~ discussed in the prev iou s sec tions, a neura l 11c 1work <.: on
sists of a lnrgc num ber of simp le process i ng ckmcnl~ cal led
ncmxms. These neuron s arc con nec ted to the eac h oth er hy di
rccted communication li nks , whi<.:11arc assoc iated with w eig ht s.
"Weight is an information used by the neural net to solve a
problem".
Figure 2.5 indicates a simp le neural network . Th e wei ght~
that cruTy information are denoted by w I and w2. They may be
fixed, or can take random va lues. Weights can be set to zero, or
Fig. 2.5 A Simple Neural Net I
can be calculated by some methods. lnitiali zation of weight~~ is an imrortant c.:ritcria in a neural net. 'l he
weight changes indic~te the overall perfo rmance of the neural net. From Fi g. 2.5,
x1 = Activation of neuron l (i nput signal)
x2 = Acti vati on of neuro n 2 (input signa l)
y= Output neuron
w =1
Weight connecting neuro n I to output
w2 = Weight connecting neuron 2 to output
Based on all these parameters, the net input ' Net' is ca lculated. The Net i'> the ~ummauon of the
products of the weights and the input signals.
Net=x 1 w 1 +x 2 w2
Generally, it can be written as,
Net input = Net = I X; w;
i
From the calculated net input, applying the acti vation function s, the output may oc cakulaktl.
2.8.2 Activation Functions
The acti vati on function is used to calcul ate the output response of a neuron . The :-.u m of tht' \\ t~1~hrcJ
...
input \ignal is applied with an ac ti vation to obtain the response. +
For neuron\ in same layer, same ac tivation fu nction s are used. There
ma y be linear as well as no nlinear activation function s. 'The non -
linear activation fun<.:tiom, arc used in a multi layer net. A few tin - f(x)
ear and non li near ac ti vation functi ons are di sc ussed here.
Identity Function
The function i.., given by,
X
I' (. ,) = x ; for all r .
f (x) = - - - I
--
1 + exp (- cr x)
wh ere , a is cal led the ste
ep ne ss parameter. If f (x ) is
differentiated we ge t.
f ' (x ) = cr f (x) [ l - f (x )).
Fig ure 2.8 sh ow s the bin
ary sigmoidal function
f(x )
a = 2
-- -- -- -
a= 1
h{ \) = 2 x
I + ex p(- a x)
2 I - ex p(- <J x )
=
I + ex p( - CJ x )
b(.,) = I ~ exp( - CJ x)
I + ex p(- a x)
on diffrrr ntiating the function b (x), we get ,
f(x)
- -- - - - - - - --Jc_ __ _ _ _ _ _ _ _ _ _
_
X
-1
II
Yi11; = L X;W ,1
1=!
Hence net input can be calcu lated using matri x multiplication method .
2.8.5 Bias
A_bias acts exactl y as a weight on a connec t inn from n unit whnse ac ti\
~1tmn is alw:1y ~ l. lnrreasing the
? 1as increas es the
net input to the unit (b = w 0 ). Fi gurt: 2. \ l) :-.hn\\ s :1 simple nt'u r~1l nr.:.'t with
the ~i:1s
1nclucled.
., \ 'eflrorks
26 !nm '<-Im ·n~ •11 , • \ , ,ft
w, y )
\
Net = b+ _L\ 1
\\ '
1
f (Net) = {+ l: if net ~ O:
- ] : if net
if bias is includ ed.
< 0:
2.8.6 Threshold
The thresh old ·e· i~ a /actor which ·
v al ue o f rhre-,h o l<.J the output may bis used in . I
en cu
e~ ca lcularcd . .lari 11 ~1., ttic, ac11va
. ti - , . . , . , I i>II th,,
For examp le, the activa ti o n ru . . · 1
-1.: . !he ac liv · •. . ons ot !It<: ~lV~Jl
nc11ons mny he .i11on lu11L·tio11 is b: 1s•'lf t )II chcB.t~n
ncr.
, ~illKl)I
ti
+ I if . •- ' . . _
U) y= f(NeL) = n et >O :
{
- J if ncr < 0;
if Y,n1 > ()
(ii) y == f(Net J : { ; 1 ti
1 Y,, ,1 0'
1 lt \e<f 1·01
- J jf
Y,ry 'l I.
< O " 11uircc1j
1 ona 1 ass ,·. .
H e nce 0 and 01 indi cate the rh r,_, ,1 · ou.ir1v e nicnu>r y ,wr
, - --~ IO/d
old va lue is d e f 1ncd by the user. ' · due lo w 1)(.(_' ~,
I !he\ ,,n
Y\ ten 1, r
cspons l' is ,·,tk 11t111~·,f J1h' tlll'
C
H
A
p What You Will Learn
T • The basic funda men tal neuro n
E model , i.e. the McCu lloch- Pitts
neuro n mode l.
R • Excita tory and inhibit ory con nection
paths in the McCu lloch-P itts neuron.
• Proce ss of learni ng in a neural
netwo rk.
• Types of learni ng algorit hms and
3 their classification based on the way
in which the adjust ment to a synaptic
weigh t of a neuro n is formu lated.
3.2.1 Architectu re
The architecture of the McCulloch Pius neuron is shown in Fig. 3. r.
~
X1
,,,,____,
X2
Xn w ~
...___ \
-p
x,,'T , ----~~
-p --------- ---·
~
Xn- 2
-p
~ Wffl
·y· is the ~kCuUoch- Pllh neuron, tt ~a.n ~~l\e ilgnal fu..)m :m~ nmn~-r f\f t.)d\M n~'U:-t't~ TI1~
connection \\ ~i g.hb from -'" ,· ... "( 0 are ~M-·irnh>f'). JenoK•d b~ ·w· an'"i th.:> .__,)nnn·ci(•n \\ ,·i~hts fu"m
f11 rn>d11l'fio11 r,> N, 1111·" / N<'tworks
,
32 • • . ., , p' The McCulloch- PillS neuron Y has the
activati Ctir~
1
I
,. • . , ,,,,., urc 111h1h1tor)' dcnotcu 1,y - . ·
on fun
I if Y- in ~0
f(Yi1) == { o if y- 111 < 0
where 0 is the threshold and y_,,, is the total net input signal
receiv ed by neuro n Y.
The
th reshold 0 should sati sfy the relatio n.
e > nw - p
This is the condition for absolute inhibi tion.
. ll1e McCulloch-Pitts neuron will fire if it receiv
es k or more excitatory inputs and no inhibiton
mputs, where
·
kw 2'. 0 > (k - I) w.
Solve d Examples
Y = l(y
. in) ::: {' if Y 111 >- 2
0 if
Y in <: 2
No" prc~cnt the input"
I if y > t}
Y= O '.f - t}~Yi n ~i,
{
- I if Yin< - 1<t -----
tht weight updat ion is given by
if y -.t- L, then
WJI CW = w old + tx