0% found this document useful (0 votes)
24 views

Learning Performance of Neuron Model Based On Quantum Superposit

1. The document describes a qubit-like neuron model based on quantum mechanics and a Quantum Back Propagation (QBP) learning rule for the model. 2. The model represents neuron states as quantum superpositions and uses complex numbers and rotation gates from quantum information theory. 3. The paper improves the previous QBP neural network model by removing a "dummy input" and tests the improved model on problems like parity check, function identification, and pattern recognition, finding it more efficient than conventional neural networks.
Copyright
© © All Rights Reserved
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
24 views

Learning Performance of Neuron Model Based On Quantum Superposit

1. The document describes a qubit-like neuron model based on quantum mechanics and a Quantum Back Propagation (QBP) learning rule for the model. 2. The model represents neuron states as quantum superpositions and uses complex numbers and rotation gates from quantum information theory. 3. The paper improves the previous QBP neural network model by removing a "dummy input" and tests the improved model on problems like parity check, function identification, and pattern recognition, finding it more efficient than conventional neural networks.
Copyright
© © All Rights Reserved
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 6

Proceedings of t h e 2000 IEEE International Workshop

on Robot and Human Interactive Communicatiori


Osaka. Japan - September 27-29 2000

Learning Performance of Neuron Model based on


Quantum Superposition

Noriaki Koudat , Nobuyuki Matsuit, Haruhiko Nishimuratt

iDept. of Computer Engineering, Faculty of TtStudies of Information Science,


Engineering, Himeji Institute of Technology Hyogo University of Education
2167 Shosha, Himeji, Hyogo, 671-2201, Japan 942- 1 Shimokume, Yashiro, Hyogo, 673-1494,
Japan

Abstract has been explored [21, and Shor found out that the
I n recent years, some researchers have been quantum algorithm factorizes large integers in
exploring quantum computer in view of the neural polynomial time [31. Perus and Kak discussed that
network to realize a distributed and strongly quantum physics makes it possible to describe the
connectionist system t h a t achieves parallel and fast processing state of the neural network 141[51.
information processing. We have proposed a We have already proposed the qubit-like
qubit-like neuron model based on quantum neuron model. In our model, a neuron state is
mechanics and constructed the Quantum Back described by quantum superposition. We have
Propagation learning rule(QBP1. I n this paper, we constructed the Quantum Back Propagation
show our improved QBP neural network model and learning rule(QBP). The QBP learning rule is the
discuss its performance on solving the 4 bit parity BP learning rule for the qubit-like neuron model
check problem, the function and the gray-scale corresponding to conventional BP one(C13P) [SI. We ’

pattern identification problem. Then, we find our have also reported the information processing
model is more excellent than the conventional one efficiency of the model based on QBP in a
in information processing efficiency. multi-layered neural network, and shown that its
performance is more excellent than the BP neural
networks [7],[8l. I n our previous simulation with
simple problems such a s basic logic operations and
1 Introduction the 4 bit parity check problem, we used one
I n recent years, we have to process a n enormous ‘dummy-input’ in input layer whose role was not
amount of information, for example, searches on clear. In t h s paper, to establish a more
online database, visual and sound cognition etc, sophisticated quantum neural computing method,
using computer systems. However, we know we reconstruct our QBP neural network by
standard computer systems have a limitation on removing the dummy input from it, and discuss our
their performance. Therefore, the demand for a improved QBP neural network model and its
more efficient computation than Von Neuman’s has performance on solving the 4 bit parity check
increased. Neural network computation system problem and the arbitrary function and the
modeled on brain information processing[ll has gray-scale pattern identification problem. Then, we
possibihties of more efficient computation power conclude our model is more excellent than the
than the conventional one because of parallel conventional one in information processing
processing. However, in a simulation of the neural efficiency.
computing on the conventional system, it loses the In the next section 2 we review short description
computational velocity of the system, and does not of the quantum information theory and recapitulate
fully achieve the parallel processing. briefly the qubit-like neuron model and the QBP
On the other hand, since the study of Deutsch, the neural network in sections 3 and 4 respectively. In
quantum computation based on quantum physics the following section 5, we report numerical

IEEE
O-780~-6~73-X/00/$10.00@2000 - 112 -
simulation results concerning our model. part and that of I1> to the imaginary part. We have
a representation of the qubit state that uses
complex numbers as follows:
2 Quantum Neuron Theory
f(8)= e“ = cos e + i sin e . (3)
2.1 Qubit
In the field of quantum computers the ‘qubit’ has Here, the i is the imaginary unit f i . The equation
been introduced as the counterpart of the ‘bit’in the (3) that represents qubit states provides the
conventional computers to describe the state of following representations of the rotation gate and
circuit of quantum computation. two-bit controlled NOT gate.
In quantum computer systems the two quantum a) The rotation gate operation
physical states labeled as 1 O> and I 1> express 1 bit The rotation gate is a phase shifting gate that
information. I O> corresponds to the bit 0 of classical transforms the phase of qubit states. As the qubit
computers, while I 1> to the bit 1. state is represented by the equation (3), the gate is
‘Qubit’ state I @ > maintains a coherent realized as the product of the following
superposition of states I O> and I 1> representation:

19) = al0) + PI 1) ( 1) f(e,+ e, = qe, 1. qe, ) . (4)

where a and L3 are complex numbers called b) The two-bit controlled NOT operation
probability amplitude. That is, the qubit state I @ > This operation is realized by making the
collapses into either I O> state with probability la[’, controlled input parameter y as follows:
or I 1> state with probability lp12 , and 7T sin8 + icose ( y = 1)
f(-r-e)
2
=
cos@-isin8 ( y = 0 )
, (5)
l
a12 + = 1. (2)
where y = 1 corresponds to reversal rotation, and
y =O to not rotation. I n the case of y =0, the phase
2.2 Quantum Gate and Its Representation of the probability amplitude of quantum state I1>
The quantum logic gate is constructed by is reversed. However, its observed probability is
connecting to the single bit rotation gate shown invariant so that we are able to regard this case as
Fig.1, and the two-bit controlled NOT gate shown no-rotation.
Fig.2. This means that these two gates are
fundamental to construct a quantum logic gate. The
state of one qubit is rotated through the bit rotation
gate to 6 angle in Fig.1. The two-bit controlled 3 Qubit-like Neuron Model
NOT gate performs XOR operation, as in Fig.2. I t We make the connection between the neuron
follows that if qubit ‘a’is I O>, output is correspond states and the quantum states. That is, we assume
to qubit ‘b’, while if qubit ‘a’ is ID,output is as follows: firing neuron state is defined as qubit
inversion of qubit ‘b’(through NOT operation). state 11>, non-firing neuron state is defined as
qubit state IO>, and the arbitrary neuron state is
the coherent superposition of the two. From the
above Section 2, our qubit-like neuron model is
defined as follows.
FIG. 1 SINGLE BIT ROTATION The k-th neuron state xk that receives inputs from
L other neurons is given by

ab Zaeb
FIG.2 TWO-BIT CONTROLLED NOT.
uk
L
=xf(e/,k)’x,-f(&)
I
(6)

Here in order to express the qubit states, we (7)


connect the probability amplitude of I O> to the real

-113 -
4.2 QBP Learning
In order to estimate our network we consider the
Here, f(x) is the same function as Eq.(3) and g(x) is squared error function
the sigmoid function
1
(9)
g(x) = 1+ exp(-x)
that is used in the CBP. Here K is the number of
In our qubit-like neuron model, we have three learning patterns and tp is the teaching signal for
important parameters, i.e., phase parameter 8 , k the p-th pattern. We, then, use the steepest descent
and reversal parameter 6 . The parameter 8 as a learning rule. That is,
relates to synaptic connection between neurons and
the parameter 1 relates to threshold parameter a s (11)
conventional neuron model. Reversal parameter 6
exists in each neuron in order to increase the
degrees of freedom of the neuron.

4 QBP Neural Network ( y w =a;" -77- ~Etotal


(13)
In order to investigate information processing 36,
efficiency of our model, we construct a network where 71 is the leaning coefficient and within the
based on the model, with the quantum back range 0.0-1.0, in usual.
propagation learning rule(QBP) that is extended
from the conventional back propagation learning 4.3 Simulation Method
rule(CBP1 and then train our network to learn We train the qubit-like neuron network to learn
problems such a s the 4 bit parity check, the the two problem, identification of logic, the
arbitrary function and the gray scale pattern arbitrary function and the gray-scale pattern. As
identification problems a s the benchmark. convergence condition, we define it has "learned, If
squared error is less than "Elower'l as threshold of
4.1 Network Structure convergence until learning iteration reach a t ''Lupper"
Our network structure is shown in Fig.3. This a s upper limit of iteration. And it has "not learned''
structure is the three layer feed-forward network. if square error is not less than Eloweruntil "lupperl'.
Then we define the probability of the observed qubit Moreover, one iteration is defined that all teaching
state I1> a s the network output. If we give input patterns are inputted to network.
data into the network, then the input layer converts 1) Learning coefficient dependence of thle QBP
input values '0'and '1' into the phase 0 and x i n First of all, we investigate the learning coefficient
quantum states, respectively. dependence of the QBP learning. For this
!"d. investigation, the learning coefficient is changed
per 0.1 and 100 trials are executed a t each learning
coefficient. Our network is constructed from 4
Input-1 Output-l input-layer neurons, 7 hidden-layer neurons, and 1
lwut-2 output-layer neuron (denoted by 4-7-1). The number
Output-2
of parameters is 51 in this 4-7-1 network. We use
the 4 bit parity check problem a s a learning issue
Output-N and clarlfjr the relation between the learning
coefficient, the convergence rate, and the average of
learning iteration. Next, we compare the results of
FIG.3 3-LAYERED FEED FORWARD NETWORK. the QBP with the those of the CBP constructed
from the 4-9-1 network with the number of
parameters, 55. I n this simulation, we set the

-114 -
convergence condition is E l o w e r z 0.001 and Lupper
=10000.
2) Identification of the arbitrary function
We train our network to solve the arbitrary
function identification problem. In this report, we
use the following function,p(x), as a n arbitrary
function whose output is normalized to the range
from 0.0 to 1.0,
20
sin x + sin 2 x + 2 ;
Pb)= 4
(0 5 x I 2 n ) . (14) 0 -:”
0.0 0.5 1.o 1.5 2.0
Learning coefficient
In this simulation, we set the convergence condition
is Elowerx 0.001 and Lupper =10000. FIG.4 LEARNING COEFFICIENT DEPENDENCE OF
3) Identification of the gray scale pattern CONVERGENCE RATE.
Next, we train the network to identify the gray
scale pattern as problem applied large structure
network for identifying pattern. This problem is
2500
that we input the gray scale pattern to network,
and train to get output pattern that is same 5 2000
C

.$
inputted one for fixed iteration number. In this t 1500
m
simulation, we use 256*256 pixel 256 degree
intensity gray scale pattern as test pattern. Test
E 1oM)
P
pattern is converted 256 degree intensity of pixel 500

into 0.0-1.0 degree when test pattern is inputted to 0


network. We apply the QBP network whose 0.0 0.5 1.o 1.5 2 .o
Learningcoefficient
construction is 64-32-64, and train for just 5000
times as learning iteration. Same as other FIG.5 LEARNING COEFFICIENT DEPENDENCE OF
simulation, we also train CBP network which is ITERATION
constructed from 64-34-64 to identify that problem
for comparison. Next, the results of the simulation 2) are shown in
Fig6 and Tablel. We set the learning coefficient on
0.4,which is the hghest efficiency from the results
5 Simulation Results of simulation 1) in QBP. Fig.6 and Table1 represent
Figs.4 and 5 show the result of the simulation 1). the QBP converges the given function within 651
From Fig.4, the QBP network converges with 100% iterations, while the CBP is not able to converge
in the learning-coefficient from 0.1 to 1.0, while the within 10000 iterations. In this case, we see the
CBPs is not able to converge at all in the range efficiency of the QBP is more excellent than that of
from 0.1 to 0.4. From the average learning iteration the CBP again.
in Fig.5, it can be seen that the value of the optimal
learning-coefficient of the QBP network is 0.4 . TABLE. 1 RESULTS OF FUNCTION IDENTIFICATION
I QBPmodel I CBPmodel

squared Frrm 0.00099 0.017

-115-
\.
0.2

I 2 3 4 5 6 7 8 9 IO I I 12 13 14 I5 16 17 18
thenumber ofdata

FIG.6 FUNCTION IDENTIFICATION . FIG.8 RESULT OF LEARNING BY QBP NETWORK


PSNR:25.34dB
The result of simulation 31, are shown in Fig.7, 8, 9.
we set the learning coefficient of QBP on 0.4 and
also set the one of CBP on 1.3, which is the most
efficient value of each model in simulation 1).
Between output pattern of QBP and input pattern,
there is the 25.34dB difference in PSNR. It is better
than result of CBP, 20.14dB. From this result, for
the large network, QBP neural network is more
efficient ability of learning than CBP one. However,
this result is obtained when input value is FIG.9 RESULT OF LEARNING BY CBP NETWORK
converted ‘0’ and ‘1’into ‘0’ and ‘ ~ / 2 ’Though
. the PSNRZ20.14dB
first we converted the input value ‘0’ and ‘1’into ‘0’
and ‘E’, the result of simulation is that PSNR is
about 18dB which is worse than the former
conversion. We need to investigate why input value
6 Conclusion
conversion depends on learning problem or network In this paper, we have presented the feed-forward
structure. neural network with qubit-like neuron model that
achieves a parallel computation. I n this model, the
state of quantum superposition is effectuated. We
simulated three identification problems on the
network of QBP model, and found that it has a
highly efficient learning ability. However, we also
found that we need to investigate why the input
value conversion depends on learning.
As our future works, first, we try to solve more
complicate problems like pattern recognition and
others. Second we clarify why the QBP model is
more efficient than the CBP model.
FIG.7 INPUT PATTERN FOR THE SIMULATION3)
Acknowledgment
This study is financially supported by Grant-in-Aid
for Scientific Research (C-10680391) of the Ministry
of Education, Science, Sports.

References
[11 D.E.Rumelhart, G.E.Hinton, R.J.Wdliams, ”Learning
internal representations by error propagation”

- 116 -
Parallel Distributed Processing‘ Explorations in the
Microstructures of Cognition, Vol.1, D.E.Rumelhart,
J.L.McClelland(Eds.), Cambridge, hWMIT Press,
pp.318-362
[Z] D.Deutsch and R.Jozsa, “Rapid solution of problems
by quantum computation”, Proc. of Royal Society of
London series A,439, pp.553-558, 1992
131 P.W.Shor, ”Algorithm for quantum computation :
discrete logarithms and factoring” Proc, the 35th
Annual IEEE Symposium on Foundations of
Computer Science, 1994, pp. 124-134.
141 Mitja Perus, “Neuro-Quantum Parallelism in
BrainaMind and Computers”, Informatica, 20,
pp173-183, 1996
151 S.C.Kak, ”On quantum neural computing”,
Information Sciences, vo1.83, pp.143-163, March
1995.
[SI Matsui, Takai, Nishimura, “A network model based
on qubit-like neuron corresponding to quantum
circuit”, IEICE, Vol.J81-A, No.12, 1998,
pp.1687-1692.
[71 Takai, Matsui, Nishimura, “A neural network based
on quantum information theory.”, Proc. SICE Kansai
branch, Annual Symposium, Vol.J81-A, No.12,
pp.154-157, Oct 1998,.
[SI Nobuyuki Matsui, Masato Takai, Haruhiko
Nishimura, “A Learning Network Based On
Qubit-Like Neuron Model”, Proc. The Seventeenth
IASTED International Conference on . Applied
Informatics, 1999.

-117-

You might also like