2010 Book ComputationalScienceAndItsAppl
2010 Book ComputationalScienceAndItsAppl
Editorial Board
David Hutchison
Lancaster University, UK
Takeo Kanade
Carnegie Mellon University, Pittsburgh, PA, USA
Josef Kittler
University of Surrey, Guildford, UK
Jon M. Kleinberg
Cornell University, Ithaca, NY, USA
Alfred Kobsa
University of California, Irvine, CA, USA
Friedemann Mattern
ETH Zurich, Switzerland
John C. Mitchell
Stanford University, CA, USA
Moni Naor
Weizmann Institute of Science, Rehovot, Israel
Oscar Nierstrasz
University of Bern, Switzerland
C. Pandu Rangan
Indian Institute of Technology, Madras, India
Bernhard Steffen
TU Dortmund University, Germany
Madhu Sudan
Microsoft Research, Cambridge, MA, USA
Demetri Terzopoulos
University of California, Los Angeles, CA, USA
Doug Tygar
University of California, Berkeley, CA, USA
Gerhard Weikum
Max-Planck Institute of Computer Science, Saarbruecken, Germany
David Taniar Osvaldo Gervasi
Beniamino Murgante Eric Pardede
Bernady O. Apduhan (Eds.)
Computational Science
and Its Applications –
ICCSA 2010
International Conference
Fukuoka, Japan, March 23-26, 2010
Proceedings, Part III
13
Volume Editors
David Taniar
Monash University, Clayton, VIC 3800, Australia
E-mail: [email protected]
Osvaldo Gervasi
University of Perugia, 06123 Perugia, Italy
E-mail: [email protected]
Beniamino Murgante
University of Basilicata, L.I.S.U.T. - D.A.P.I.T., 85100 Potenza, Italy
E-mail: [email protected]
Eric Pardede
La Trobe University, Bundoora, VIC 3083, Australia
E-mail: [email protected]
Bernady O. Apduhan
Kyushu Sangyo University, Fukuoka 813-8503, Japan
E-mail: [email protected]
ISSN 0302-9743
ISBN-10 3-642-12178-0 Springer Berlin Heidelberg New York
ISBN-13 978-3-642-12178-4 Springer Berlin Heidelberg New York
This work is subject to copyright. All rights are reserved, whether the whole or part of the material is
concerned, specifically the rights of translation, reprinting, re-use of illustrations, recitation, broadcasting,
reproduction on microfilms or in any other way, and storage in data banks. Duplication of this publication
or parts thereof is permitted only under the provisions of the German Copyright Law of September 9, 1965,
in its current version, and permission for use must always be obtained from Springer. Violations are liable
to prosecution under the German Copyright Law.
springer.com
© Springer-Verlag Berlin Heidelberg 2010
Printed in Germany
Typesetting: Camera-ready by author, data conversion by Scientific Publishing Services, Chennai, India
Printed on acid-free paper 06/3180
Preface
These multiple volumes (LNCS volumes 6016, 6017, 6018 and 6019) consist of
the peer-reviewed papers from the 2010 International Conference on Computa-
tional Science and Its Applications (ICCSA2010) held in Fukuoka, Japan during
March 23–26, 2010. ICCSA 2010 was a successful event in the International Con-
ferences on Computational Science and Its Applications (ICCSA) conference se-
ries, previously held in Suwon, South Korea (2009), Perugia, Italy (2008), Kuala
Lumpur, Malaysia (2007), Glasgow, UK (2006), Singapore (2005), Assisi, Italy
(2004), Montreal, Canada (2003), and (as ICCS) Amsterdam, The Netherlands
(2002) and San Francisco, USA (2001).
Computational science is a main pillar of most of the present research, in-
dustrial and commercial activities and plays a unique role in exploiting ICT in-
novative technologies. The ICCSA conference series has been providing a venue
to researchers and industry practitioners to discuss new ideas, to share complex
problems and their solutions, and to shape new trends in computational science.
ICCSA 2010 was celebrated at the host university, Kyushu Sangyo Univer-
sity, Fukuoka, Japan, as part of the university’s 50th anniversary. We would like
to thank Kyushu Sangyo University for hosting ICCSA this year, and for in-
cluding this international event in their celebrations. Also for the first time this
year, ICCSA organized poster sessions that present on-going projects on various
aspects of computational sciences.
Apart from the general track, ICCSA 2010 also included 30 special sessions
and workshops in various areas of computational sciences, ranging from compu-
tational science technologies, to specific areas of computational sciences, such as
computer graphics and virtual reality. We would like to show our appreciation
to the workshops and special sessions Chairs and Co-chairs.
The success of the ICCSA conference series, in general, and ICCSA 2010, in
particular, was due to the support of many people: authors, presenters, partic-
ipants, keynote speakers, session Chairs, Organizing Committee members, stu-
dent volunteers, Program Committee members, Steering Committee members,
and people in other various roles. We would like to thank them all. We would
also like to thank Springer for their continuous support in publishing ICCSA
conference proceedings.
ICCSA 2010 was organized by the University of Perugia (Italy), Monash Univer-
sity (Australia), La Trobe University (Australia), University of Basilicata (Italy),
and Kyushu Sangyo University (Japan)
General Chairs
Bernady O. Apduhan Kyushu Sangyo University, Japan
Osvaldo Gervasi University of Perugia, Italy
Advisory Committee
Marina L. Gavrilova University of Calgary, Canada
Andrès Iglesias University of Cantabria, Spain
Tai-Hoon Kim Hannam University, Korea
Antonio Laganà University of Perugia, Italy
Katsuya Matsunaga Kyushu Sangyo University, Japan
Beniamino Murgante University of Basilicata, Italy
Kazuo Ushijima Kyushu Sangyo University, Japan (ret.)
Publicity Chairs
Jemal Abawajy Deakin University, Australia
Koji Okamura Kyushu Sangyo University, Japan
Yao Feng-Hui Tennessee State University, USA
Andrew Flahive DSTO, Australia
VIII Organization
Tutorial Chair
Andrès Iglesias University of Cantabria, Spain
Awards Chairs
Akiyo Miyazaki Kyushu Sangyo University, Japan
Wenny Rahayu LaTrobe University, Australia
Workshop Organizers
Application of ICT in Healthcare (AICTH 2010)
Salim Zabir France Telecom /Orange Labs Japan
Jemal Abawajy Deakin University, Australia
Program Committee
Kenneth Adamson Ulster University, UK
Margarita Albertı́ Wirsing Universitat de Barcelona, Spain
Richard Barrett Oak Ridge National Laboratory, USA
Stefania Bertazzon University of Calgary, Canada
Michela Bertolotto University College Dublin, Ireland
Sandro Bimonte CEMAGREF, TSCF, France
Rod Blais University of Calgary, Canada
Ivan Blecic University of Sassari, Italy
Giuseppe Borruso Università degli Studi di Trieste, Italy
Martin Buecker Aachen University, Germany
Alfredo Buttari CNRS-IRIT, France
Carlo Cattani University of Salerno, Italy
Alexander Chemeris National Technical University of Ukraine
“KPI”, Ukraine
Chen-Mou Cheng National Taiwan University, Taiwan
Min Young Chung Sungkyunkwan University, Korea
Rosa Coluzzi National Research Council, Italy
Stefano Cozzini National Research Council, Italy
José A. Cardoso e Cunha Univ. Nova de Lisboa, Portugal
Gianluca Cuomo University of Basilicata, Italy
Alfredo Cuzzocrea University of Calabria, Italy
Ovidiu Daescu University of Texas at Dallas, USA
Maria Danese University of Basilicata, Italy
Pravesh Debba CSIR, South Africa
Oscar Delgado-Mohatar University Carlos III of Madrid, Spain
Roberto De Lotto University of Pavia, Italy
XII Organization
Sponsoring Organizations
ICCSA 2010 would not have been possible without the tremendous support of
many organizations and institutions, for which all organizers and participants of
ICCSA 2010 express their sincere gratitude:
University of Perugia, Italy
Kyushu Sangyo University, Japan
Monash University, Australia
La Trobe University, Australia
University of Basilicata, Italia
Information Processing Society of Japan (IPSJ) - Kyushu Chapter
and with IPSJ SIG-DPS
Table of Contents – Part III
The Time Series Image Analysis of the HeLa Cell Using Viscous Fluid
Registration . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 189
Soichiro Tokuhisa and Kunihiko Kaneko
Abstract. This paper designs a loop reduction scheme and measures its
performance for the wireless process control application running on the
grid-style traffic light network, aiming at improving the response time of
the control system. Based on the WirelessHART standard, which assigns
a slot to each (sender, receiver) pair according to the routing schedule,
the allocation scheme puts all the pairs having no interference to a single
slot, reducing the loop length. For further reduction, the classic Dijkstra’s
shortest path algorithm is modified such that the number of end-to-end
paths starting from the horizontal links and the vertical link, respec-
tively, is almost same. The transmission of the controller, which initiates
all message delivery and is the main bottleneck point, will not be blocked.
The simulation result demonstrates that the proposed scheme can sig-
nificantly reduce the control loop length, and the modified path finding
algorithm further achieves about 9.8 % improvement, just sacrificing the
3.2 % of transmission success ratio.
1 Introduction
Vehicular telematics networks can keep the vehicle connected to a network even
while it is on the move. This network exploits matured wireless communication
technologies and essentially includes static elements which provide the access
point to moving vehicles. Correspondingly, each component communicates in
two-level hierarchy in vehicular networks [1]. Level 1 corresponds to connection
between static nodes and level 2 between a static node and moving vehicles
trying to access the network. Level 1 communication creates a kind of wireless
mesh networks which have been intensively researched and widely deployed into
diverse field areas, exploiting commonly available protocols such as IEEE 802.11,
Zigbee, and the like [2]. Based on this network, many vehicular applications can
be developed and serviced.
This research was supported by the MKE, Korea, under the ITRC support program
supervised by the NIPA. (NIPA-2009-C1090-0902-0040)).
Corresponding author.
D. Taniar et al. (Eds.): ICCSA 2010, Part III, LNCS 6018, pp. 1–10, 2010.
c Springer-Verlag Berlin Heidelberg 2010
2 J. Lee et al.
Traffic lights are found in every street, and they can desirably install a wire-
less communication interface as they have sufficient power provision and their
locations are highly secure. On such a traffic light network, it is possible to im-
plement a monitor-and-control application when the network includes sensors
and actuators [3]. Practically, a lot of traffic-related or environmental devices
such as speed detectors, pollution meters, and traffic signal controllers, can be
covered by the vehicular network. Moreover, vehicles can also carry a sensor
device and report the collected sensor data to a static node when connected.
The message of process control applications is necessarily time critical and how
to schedule messages is the main concern. It depends on the node distribution,
namely, topology, and the communication protocol. First, traffic lights are placed
in each intersection of Manhattan-style road networks, so the traffic light network
has grid topology. Additionally, the network protocol must provide predictable
network access.
The WirelessHART standard provides a robust wireless protocol for various
process control applications [4]. First of all, the protocol standard exploits the
slot-based access scheme to guarantee a predictable message delivery, and each
slot is assigned to the appropriate (sender, receiver) pair. In addition, for the sake
of overcoming the transient instability of wireless channels, a special emphasis is
put on reliability by mesh networking, channel hopping, and time-synchronized
messaging. This protocol has been implemented and is about to be released to
the market [5]. It can accommodate a split-merge operation to mask channel
errors [6] as well as an efficient routing scheme to find the path that is most
likely to successfully deliver the message [7].
For the timely control action, the network must deliver the sensor and control
message not only reliably but also timely, or as fast as possible. Thus, every
message transmission schedule, or slot assignment, is decided in priori. Moreover,
to speed up the response time of a control process, the length of the slot schedule
must be minimized. In the mesh network, more than one transmission can be
accomplished simultaneously as long as transmitters are far way enough not
to interfere each other. If the node is equipped with an directional antenna,
more transmissions can be overlapped. However, the slot assignment permitting
multiple transmissions is a NP hard problem, which has no polynomial time
solution. In this regard, this paper is to design and evaluate the performance of
a slot allocation scheme to the the sensor and control messages for the grid-style
traffic light control network.
The paper is organized as follows: After defining the problem in Section 1,
Section 2 introduces the background of this paper focusing on the WirelessHART
protocol and slot allocation schemes. Section 3 designs control loop reduction
schemes on the target network. The simulation result is discussed in Section 4,
and finally Section 5 summarizes and concludes this paper.
band [8]. The link layer provides deterministic slot-based access on top of the
time synchronization primitives carried out continuously during the whole net-
work operation time. According to the specification, the size of a single time
slot is 10 ms, and a central controller node coordinates routing and communi-
cation schedules to meet the robustness requirement of industrial applications.
For more reliable communication, CCA (Clear Channel Assessment) [9] before
each transmission and channel blacklisting is exploited to avoid specific area of
interference and also to minimize interference to others. In some WirelessHART
implementation, the network manger traverses a communication graph or a grid
by the breadth-first search and allocates slots according to this order [8].
In each slot, a sender can try another channel if the CCA result of the channel
on the primary schedule is not clear, even though such an operation is not yet
defined in the current standard. For each destination node, a controller may
reserve two alternative paths having sufficient number of common nodes. Two
paths split at some nodes and meet again at other nodes. When two paths
split, a node can select the path according to the CCA result in a single slot
by switching to the channel associated with the secondary route. When two
paths merge, the node can receive from two possible senders by a timed switch
operation. Besides, the WirelessHART protocol can be reinforced by many useful
performance enhancement schemes such as the virtual-link routing scheme that
combines split-merge links and estimates the corresponding error rate to apply
the shortest path algorithm.
Wireless mesh networks are cost-effective solutions for ubiquitous high-speed
services, and its performance depends on the routing strategy. Routing has been
extensively studied in wireless mesh networks, and most routing schemes are
based on the shortest path algorithm [10]. The path cost is different according
to the main goal of each routing scheme, including actual distance, transmis-
sion rate, and error characteristics. In most wireless process control systems, the
central routing mechanism is generally taken, as not only the traffic character-
istics are accurately known, but also each network access is predictable. Hence,
it is possible to jointly allocate channels, route a message set, and schedule each
transmission [11]. In addition, the routing procedure is executed repeatedly ac-
cording to the link condition change and the schedule is distributed to each
node [12].
A control loop consists of three phases, namely, reading state variables from
sensors, deciding the control action, and sending the value of control variables
to the actuators. The first and third steps involve the network and have their
own communication schedules.
in the original algorithm. In each node scan, the node having the highest success
probability will be selected. The path from the controller, T0,0 , to each node is
shown in Figure 2(a). The number in the parenthesis means the hop length of
the path. Each column is a time slot and each slot has a (sender, receiver) pair.
As each node has just one transmitter and one receiver module, neither the same
sender nor the same receiver can appear twice in the same column. So, the initial
allocation should be modified to avoid such a conflict.
To this end, the scheduler checks if two entries interfere. They interfere if 1)
two senders are the same, or 2) two receivers are the same, or 3) the sender of
one slot and the receiver of the other slot are the same. This criteria is valid
just for the directional antenna which can increase spatial reuse and reduce
interference by directing radio beam towards a desired direction [15]. When
the node transmits using an omnidirectional antenna, which radiates or receives
electromagnetic energy in all directions, the probability of interference increases,
as the two transmissions will collide when the two senders are less than two hops
away from each other. If two transmissions are not compatible in the same slot,
the one with a shorter path will be moved one slot back, as the longer path
is more critical to the whole control loop length. Such a move-back procedure
is repeated until there is no change. Figure 2 (b) show the final result. In this
figure, 0-3 means T0,0 → T1,0 .
Slot 1 2 3 4 Slot 1 2 3 4 5 6 7 8 9 10 11
8 (4) 0−3 3−6 6−7 7−8 8 (4) 0−3 3−6 6−7 7−8
5 (3) 0−3 3−4 4−5 5 (3) 0−3 3−4 4−5
7 (3) 0−3 3−6 6−7 7 (3) 0−3 3−6 6−7
2 (2) 0−1 1−2 2 (2) 0−1 1−2
4 (2) 0−3 3−4 4 (2) 0−3 3−4
6 (2) 0−3 3−6 6 (2) 0−3 3−6
1 (1) 0−1 1 (1) 0−1
3 (1) 0−3 3 (1) 0−3
As T0,1 appears twice in slot t + 1, one as a sender in the first row, and
the other as a receiver in the second row. On the contrary, if the first link for
Dest 2 is T0,0 → T1,0 as in Table 1(b), and two transmissions can be done
simultaneously on slot t + 1, reducing the control loop length. In case of Figure
2(a), 6 destinations takes T0,0 → T1,0 while just 2 destinations takes T0,0 →
T0,1 for the first step. So, the length reduction in not so significant. To make
the number of two first links as close as possible, the network can sacrifice the
optimality of the shortest path algorithm.
We classify grid nodes into 3 groups. The first group, GD , has the nodes on
the diagonal from top left to bottom right. The second group, DL , has the nodes
on the lower triangle of the grid, while the third group, DU , has the node on the
upper triangle. Accordingly, each group can be specified as follows:
GD : { Ti,j | i = j }
GL : { Ti,j | i > j }
GU : { Ti,j | i < j }
For nodes in GL , we make every route start from the link T0,0 → T1,0 , and for
those in GU , from the link T0,0 → T0,1 . GD nodes can take the route according to
the normal Dijkstra’s algorithm. For GL nodes, after setting the slot error rate
of T0,0 → T0,1 to 1.0, the scheduler runs the shortest path algorithm, and vice
versa for the GU nodes. Likewise, the number of appearances for T0,0 → T0,1 and
A Control Loop Reduction Scheme for Wireless Process Control 7
T0,0 → T1,0 becomes almost even. While this allocation cannot find the optimal
path to each node, it can maximize the simultaneous transmissions in each slot
and reduce the control loop length.
4 Performance Measurement
This section measures the performance of the proposed loop reduction scheme
via simulation using SMPL which provides abundant functions and libraries
for discrete event scheduling [16]. Only the downlink graph was considered for
simplicity, as uplink and downlink communications are symmetric and calculated
completely in the same way. For each parameter setting, 500 sets of link error
rates are generated and the success ratio and the loop length ratio are averaged.
The first experiment measures the performance of our loop reduction scheme
and compares with the upper and lower bounds. This experiment is conducted
for the classic Dijkstra’s algorithm. First, the lower bound accounts for the case
that each end-to-end transmission can start one slot after another without be-
ing blocked from the largest hop-length destination. In this case, the control loop
length is ideally smallest and thus the lower bound can be calculated as in Eq. (1).
where n × n − 1 is the number of noncontroller nodes in the grid, and for the
last message initiated by the controller, it takes as many slots as the hop counts
to the destination.
Next, the upper bound of the control loop length is estimated by assuming
that there is no overlapped transmission. Hence, it corresponds to the sum of
total hops to each node. For the n × n grid, the number of total transmissions
for a control round is calculated as in Eq. (2).
2(n−1)
(n − |n − i − 1|) · i, (2)
i=1
where each iteration indexed by i corresponds to the hop count from the con-
troller. Nodes having the same hop counts need the same number of slots. The
number of nodes having i hops increases by one until i reaches (n-1), and then
decreases also by one until i reaches 2(n-1).
Figure 3 shows the comparison results according to the grid dimension and
the slot error rate. First, in Figure 3(a), the average slot rate is 0.1, distributing
exponentially, while the grid dimension ranges from 3 to 15. The curves plots
the ratio of each loop length to the upper bound. Hence, the curve marked
as UpperBound always remain 1.0. The ratio of the packet length to the upper
bound is just 5.7 % larger at maximum, compared with the lower bound to upper
bound ratio. When the dimension is 3, both ratios are around 0.58, as nodes are
not sufficiently apart from each other and not so many transmissions can be
overlapped. The gap between 2 curves reaches 5.7 % at the grid dimension of 10,
8 J. Lee et al.
but decreases afterwards. It indicates that our scheme can achieve almost ideal
loop reduction without complex and time-consuming space search iterations. In
addition, Figure 3(b) shows the effect of the slot error rate, which can lead to
the variation in the length of each end-to-end path. Here, the grid dimension
is set to 5. Even though the gap between the lower bound and our packed
scheme gets smaller according to the increase of the slot error rate, its effect is
insignificant.
1 1
"UpperBound" "UpperBound"
0.9 "LowerBound" "LowerBound"
0.8 "Packed" 0.8 "Packed"
Loop length ratio
1 1
"Grid" "Grid"
"Proposed" "Proposed"
0.9 0.8
Loop length ratio
Success ratio
0.8 0.6
0.7 0.4
0.6 0.2
0.5 0
4 6 8 10 12 14 4 6 8 10 12 14
Square grid dimension Square grid dimension
Finally, Figure 5 shows the effect of the slot error rate in the proposed routing
scheme. In this experiment, the grid dimension is also set to 5, while the slot
error rate changes from 0 to 0.5. The success ratio gap increases along with
the error rate, reaching 4.4 % at maximum. However, the proposed scheme can
achieve consistently about 8 % loop reduction in all error ranges.
1 1
"Grid" "Grid"
"Proposed" "Proposed"
0.8 0.8
0.6 0.6
0.4 0.4
0.2 0.2
0 0
0 0.1 0.2 0.3 0.4 0.5 0 0.1 0.2 0.3 0.4 0.5
Slot error rate Slot error rate
Fig. 5. Loop reduction vs. success ratio according to slot error rate
5 Conclusion
This paper has designed a loop reduction scheme and measured its performance
for the wireless process control application running on the grid-style traffic light
network, aiming at improving the response time of the control system. Based on
the WirelessHART standard, which assigns a slot to each (sender, receiver) pair
according to the routing schedule, the allocation scheme puts all the pairs having
no interference to a single slot, reducing the loop length. For further reduction,
the classic Dijkstra’s shortest path algorithm is modified such that the number
of end-to-end paths starting from the horizontal links and the vertical link,
respectively, is almost same. The transmission of the controller, which initiates
all message delivery and is the main bottleneck point, will not be blocked. The
simulation result demonstrates that the proposed scheme can significantly reduce
the control loop length, and the modified path finding algorithm further achieves
about 9.8 % improvement, just sacrificing the 3.2 % of transmission success ratio.
As for future work, a fault-tolerant flooding scheme is expected to be very
useful in wireless process control [17]. So, we will design a slot assignment scheme
combining the split-merge operation to overcome node or link failure for the
control message broadcast.
References
1. Bucciol, P., Li, F.Y., Fragoulis, N., Vandoni, L.: ADHOCSYS: Robust and service-
oriented wireless mesh networks to bridge the digital divide. In: IEEE Globecom
Workshops, pp. 1–5 (2007)
2. Gislason, D.: ZIGBEE Wireless Networking. Newnes (2008)
10 J. Lee et al.
1 Introduction
According to the ongoing deployment of in-vehicle telematics devices and ve-
hicular networks, wireless vehicular communications have become an important
priority for car manufactures[1]. The vehicular network can be built in many
different ways according to the variety of available wireless communication tech-
nologies, for example, infrastructure-based cellular networks, VANET (Vehicular
Ad hoc NETwork), and roadside networks. Moreover, not just the vehicle but
also diverse communication entities are participating in this network, includ-
ing traffic lights, roadside units, sensors, gas stations, and many other facilities.
Correspondingly, abundant application scenarios are now possible. In addition
to the standard safety-related applications such as traffic accident propagation
This research was supported by the MKE, Korea, under the ITRC support program
supervised by the NIPA. (NIPA-2009-C1090-0902-0040)).
Corresponding author.
D. Taniar et al. (Eds.): ICCSA 2010, Part III, LNCS 6018, pp. 11–20, 2010.
c Springer-Verlag Berlin Heidelberg 2010
12 J. Lee et al.
and emergency warning[2], information retrieval is very useful for drivers and
passengers[3]. Specifically, a driver may want to know the current traffic con-
dition of specific road segments lying on the way to his/her destination, query
several shops to decide where to go, and check parking lot availability[4].
In the information retrieval service, queries must be delivered to the destina-
tions and the result is sent back to the query issuer. A message proceeds to its
destination according to a routing protocol each vehicle cooperatively runs. As
each vehicle can move only along the road segment and static nodes such as gas
stations and traffic lights are generally placed on the roadside[5], the message de-
livery path must trail the actual road layout[6]. Just as in the vehicle’s path, in-
tersection areas having many vehicles are important for the delay and reliability
of message transmissions, because the large number of vehicles makes the routing
decision very complex and raises intervehicle interference. Moreover, the carry and
forward strategy is preferred to cope with the disconnection problem in the sparse
vehicular network[7]. As will be described in the next section, this behavior further
increases the complexity of message routing near the intersection area.
Definitely, there exists an intersection which has much more traffic than others
and this hot area may be a bottleneck for both vehicle and data traffic. As an end-
to-end path is highly likely to involve one or more such intersection areas, how
to manage the communication in this area is very critical to the communication
performance such as the transmission delay and the delivery ratio. In this regard,
this paper is to design and measure the performance of a reliable and high-speed
message switching scheme around the hot intersection area for the vehicular
network, taking advantage of multiple channels. It is possible to create multiple
channels in a cell in many available wireless protocols such as IEEE 802.11
series[2], Zigbee, and WirelessHart[8]. Existing researches have also pointed out
that multiple network interfaces does not cost too much[9]. Moreover, the slot-
based collision-free access scheme can be employed for predictable channel access,
short delay, and reliability, while the probing-based access can efficiently reduce
the effect of the channel error inherent in the wireless frequency channel.
This paper is organized as follows: After issuing the problem in Section 1,
Section 2 describes the background of this paper and related work. Section 3
designs a wireless switch for intersection areas and performance measurement
results are demonstrated and analyzed in Section 4. Finally, Section 5 summa-
rizes and concludes this paper with a brief introduction of future work.
time
A−D1 RTS/CTS
Ch1 Data
B C
Ch2 Data A−D2 RTS/CTS
Many vehicles passing by the intersection area may cause channel interference,
so an efficient method is necessary to cope with this problem. This problem
can be solved by dual channels. The slots of two channels are synchronized,
and two slots of the same time instance must be assigned to the same node.
At the beginning of each slot, the transmitter node gets two messages having
different destinations from its message queue. Our previous work has proposed a
channel probing mechanism from two different nodes[14]. On the contrary, this
paper makes a single node be assgined time slots on each channel and select the
appropriate one. A channel probing result its corresponding action is similar in
both schemes.
Let these two destinations be D1 and D2 , respectively. The sender probes, for
example, with RTS/CTS methods, D1 and D2 on the two channels as shown in
Fig 1 (b). The transmission on the bad channel is meaningless not solely for the
power consumption aspect. It may extend the instable energy and its duration.
Table 1 shows the probing result and corresponding actions. As shown in row
1, D1 on channel 1 and also D2 on channel 2 are both in good state, D1 and
D2 send as scheduled. In row 2, every channel status is good except the one for
D2 on channel 2. If we switch < D1 , D2 > to < D2 , D1 >, both nodes can
successfully send their messages. Otherwise, only A can succeed. In this case,
the channel switch can save one transmission loss. Row 8 describes the situation
that D1 is good only on channel 2 while D2 also only on channel 1. Switching
two channels saves 2 transmissions that might fail on the regular schedule.
Each switch node maintains a message queue. The buffer space can be assumed
to have no limitation, but it is desirable to discard a message which stayed in
the buffer too much so as to obviate undue delay for the subsequent messages.
After all, we assume that a message is automatically deleted from its queue
when its waiting time exceeds the given discard interval. In addition, the rout-
ing decision outside the intersection area can be exploited by other work and
16 J. Lee et al.
there are many available schemes especially targeting at the urban area[15]. For
example, VADD can cope with the situation that a vehicle moves after issuing
a retrieval request. Finally, how to assign slots is not our concern, and we just
consider the round-robin style allocation and assume that each slot is assigned
to a switch node. There also exist some channel allocation schemes for the give
traffic requirement[16].
4 Performance Measurement
We evaluate the performance of our scheme, focusing on the effect of channel
switches, via simulation using SMPL, which provides a simple and robust library
for discrete event trace similar to ns-2 event scheduler[17]. The main performance
metrics are the delivery ratio and the access delay. The delivery ratio measures
the ratio of the number of successfully delivered messages through the intersec-
tion area to the number of total messages arrived at the 4 switch nodes. The
access delay is the time interval from the instant a message arrives at a switch
node to the instant it gets out of the intersection switch. The average of access
delays just takes the messages that have been successfully transmitted, so it is
likely to increase when the number of successful transmissions increases. After
all, this section shows the effect of the channel error probability, offered load,
and the discard interval, one by one. In the following experiments, the message is
assumed to arrive at each switch node according to the exponential distribution.
In addition, for simplicity, each message fits a single time slot, making simple to
estimate the traffic load. Moreover, every time is aligned to a slot length.
The vehicle distribution is taken from the Jeju taxi telematics system which
keeps track of each member taxi[1]. Fig. 2 shows a snapshot of vehicle distri-
bution on the real road network. This history is very useful for the vehicular
6 1
"Normal" "Normal"
"Proposed" "Proposed"
Access delay (slot time)
5.8 0.8
Delivery ratio
5.6 0.6
5.4 0.4
5.2 0.2
5 0
0 0.05 0.1 0.15 0.2 0.25 0.3 0.35 0.4 0.45 0.5 0 0.05 0.1 0.15 0.2 0.25 0.3 0.35 0.4 0.45 0.5
Channel error probability Channel error probability
Fig. 3. Access delay at low load Fig. 4. Delivery ratio at low load
network design and the traffic pattern analysis. In this figure, the intersection is
represented by a rectangle and the road segment is by a line, while the location
of taxis during the specific time interval is marked with a small dot. The intersec-
tion area has more dots, which appear just along the road segment, even though
the number of dots is different for each segment which meets at an intersection.
Fig. 3 and Fig. 4 show the effect of the channel error probability when the
offered load is relatively low, namely, 0.5. The curve marked as proposed is the
performance of the case of channel switching. The channel error probability is a
probability that a channel is clean when a switch node is to send its message. This
experiment assumes that a message is discarded if it is not served until 16 time
slots from its arrival. Fig. 3 reveals that a message can exit the intersection area
around in 5.6 time slots, that is, less than 1.5 round. In Fig. 4 the delivery ratio
is almost 1.0 when the channel error rate is close to 0. At the higher error rate,
we can achieve better improvement, as there are more cases channel switching
is possible. When the channel error probability is greater than 0.4, the proposed
scheme can improve the delivery ratio by up to 18.5 %.
Fig. 5 and Fig. 6 plot the effect of the channel error probability when the
offered load is 1.0. On high load, a switch node can almost always find two
12 1
"Normal" "Normal"
"Proposed" "Proposed"
Access delay (slot time)
11.8 0.8
Delivery ratio
11.6 0.6
11.4 0.4
11.2 0.2
11 0
0 0.05 0.1 0.15 0.2 0.25 0.3 0.35 0.4 0.45 0.5 0 0.05 0.1 0.15 0.2 0.25 0.3 0.35 0.4 0.45 0.5
Channel error probability Channel error probability
Fig. 5. Access delay at high load Fig. 6. Delivery ratio at high load
18 J. Lee et al.
12 1
11
Access delay (slot time)
0.8
10
Delivery ratio
9 0.6
5 0
0.5 0.6 0.7 0.8 0.9 1 0.5 0.6 0.7 0.8 0.9 1
Load Load
Fig. 7. Access delay vs. load Fig. 8. Delivery ratio vs. load
messages having different destinations. As can be found in Fig. 5, the access delay
for both cases stays at 11.2 time slots, namely, less than 3 rounds. This indicates
that the discard interval can control the access delay quite well regardless of
channel errors. When the channel error probability gets larger than 0.4, the
proposed scheme can improve the delivery ratio by up to 13.0 %. Fig. 6 shows
that both the delivery ratio and its performance gap are less than those in Fig.
4, indicating that the message discarded in the queue outnumbers that saved by
the channel switch.
Fig. 7 and Fig. 8 demonstrate the effect of message load to the access delay and
the delivery ratio, respectively. In this experiment, the channel error probability
is set to 0.1 and the discard interval is fixed to 16 time slots. Fig. 7 shows that
the access delays for two cases are almost same and the degree of inclination
gets smaller after the load is 0.8, as more messages are discarded for the range
of 0.8 through 1.0. In Fig. 8, the two curves show the constant gap for all load
range, while the gap is 8.1 % at maximum, indicating that the early discard
more affects the delivery ratio on higher load.
Fig. 9 and Fig. 10 show the effect of discard intervals. It can be simply expected
that the shorter the discard interval, the shorter the access delay. Fig. 9 exhibits
7 1
0.9
Access delay (slot time)
0.8
Delivery ratio
5
0.7
4 "Normal"
"Proposed" 0.6 "Normal"
"Proposed"
3
0.5
2 0.4
0 5 10 15 20 25 30 35 0 5 10 15 20 25 30 35
Discard interval Discard interval
the largely linear increase of access delays according to the discard interval. Two
curves also have no conceivable difference as in the other cases. However, Fig. 10
reveals that the effect of discard intervals is not so significant to the delivery
ratio, less than 18 % over the interval range of 5 through 35. Two curves show
a similar pattern, but the gap gets a little bit larger according to the increase of
discard intervals. For all experiments, it can be pointed out by our experiment
that the improved delivery ratio doesn’t affect the transit time.
5 Conclusion
This paper has proposed and measured the performance of a message switch
scheme for a hot intersection area in vehicular telematics network. Installed at
each corner of an intersection, switch nodes cooperatively exchange messages ac-
cording to the current channel status, speeding up the intersection transit time
and improving the delivery ratio. The simulation result shows that the proposed
scheme can improve the delivery ratio by up to 18.5 % for the given range of
channel error probability, and 8.1 % for the given range of network load. This
result comes from the fact that our scheme can avoid the complex routing and
message interference in the intersection of intensive traffic. Next, dual channels
can efficiently cope with channel errors. Even though such employment of ad-
ditional equipments brings extra cost and loses the advantage of autonomous
operations, access predictability, enhanced delivery speed, and better delivery
ratio can compensate for the cost. In short, the wireless switch can be installed
at the intersections, which have a lot of traffic especially in the urban area, to
enhance the communication quality and reliability.
As future work, we are planning to assess the performance of our scheme with
the real-life location history data obtained by the Jeju taxi telematics system[1]
and analyze the end-to-end performance characteristics to revise our algorithm
and test the diverse message scheduling policy.
References
1. Lee, J., Park, G., Kim, H., Yang, Y., Kim, P., Kim, S.: A telematics service system
based on the Linux cluster. In: Shi, Y., van Albada, G.D., Dongarra, J., Sloot,
P.M.A. (eds.) ICCS 2007. LNCS, vol. 4490, pp. 660–667. Springer, Heidelberg
(2007)
2. Society of Automotive Engineers: Dedicated short range communication message
set dictionary. Technical Report Standard J2735 (2006)
3. Dikaiakos, M., Iqbal, S., Nadeem, T., Iftode, L.: VITP: An information transfer
protocol for vehicular computing. In: ACM VANET, pp. 30–39 (2005)
4. Zhao, J., Arnold, T., Zhang, Y., Cao, G.: Extending drive-thru data access by
vehicle-to-vehicle relay. In: ACM VANET, pp. 30–39 (2008)
5. Lochert, C., Scheuermann, B., Wewetzer, C., Luebke, A., Mauve, M.: Data aggre-
gation and roadside unit placement for a VANET traffic information system. In:
ACM VANET, pp. 58–65 (2008)
20 J. Lee et al.
6. Korkmaz, G., Ekici, E., Ozguner, F., Ozguner, U.: Urban multihop broadcast pro-
tocol for inter-vehicle communication systems. In: ACM VANET, pp. 76–85 (2004)
7. Davis, J., Fagg, A., Levine, B.: Wearable computers as packet transport mech-
anism in highly-partitioned ad-hoc networks. In: Int’l Symposium on Wearable
Computing (2001)
8. Song, S., Han, S., Mok, A., Chen, D., Nixon, M., Lucas, M., Pratt, W.: Wireless
HART: Applying Wireless Technology in Real-Time Industrial Process Control.
In: The 14th IEEE Real-Time and Embedded Technology and Applications Sym-
posium, pp. 377–386 (2008)
9. Naumov, V., Gross, T.: Connectivity-aware routing (CAR) in vehicular ad hoc
networks. In: IEEE Infocom, pp. 1919–1927 (2007)
10. US Department of Transportation: Integrated Vehicle-Based Safety Systems
(IVBSS) Phase I Interim Report. DOT HS 810 952 (2008),
http://www.itsdocs.fhwa.dot.gov/JPODOCS//REPTS_TE/14434.htm
11. Kodialam, M., Nandagopal, T.: Characterizing the capacity region in multi-radio
multi-channel wireless mesh networks. In: ACM MobiCom, pp. 73–87 (2005)
12. Zhao, J., Cao, G.: VADD: Vehicle-assisted data delivery in vehicular ad hoc net-
works. In: IEEE INFOCOM (2006)
13. Lee, J., Kang, M., Park, G., Shin, S.: Design of a reliable real-time scheduling
policy for dual-channel networks. Journal of Information Science and Engineering,
1407–1419 (2007)
14. Lee, J., Park, G., Shin, S., Kang, M.: Design of intersection switches for the vehic-
ular network. In: Hong, C.S., Tonouchi, T., Ma, Y., Chao, C.-S. (eds.) APNOMS
2009. LNCS, vol. 5787, pp. 523–526. Springer, Heidelberg (2009)
15. Jaap, S., Bechler, M., Wolf, L.: Evaluation of Routing Protocols for Vehicular Ad
Hoc Networks in City Traffic Scenarios. In: Proceedings of the 5th International
Conference on Intelligent Transportation Systems Telecommunications (2005)
16. Li, H., Shenoy, P., Ramamritham, K.: Scheduling communication in real-time sen-
sor applications. In: Proc.10th IEEE Real-time Embedded Technology and Appli-
cations Symposium, pp. 2065–2080 (2004)
17. MacDougall, M.: Simulating Computer Systems: Techniques and Tools. MIT Press,
Cambridge (1987)
A Rapid Code Acquisition Scheme for Optical CDMA
Systems
Abstract. In this paper, we propose a novel code acquisition scheme called mod-
ified multiple-shift (MMS) for optical code division multiple access (CDMA)
systems. By using the modified local code whose sign is positive or negative,
the proposed MMS scheme provides a shorter mean acquisition time (MAT) than
the conventional multiple-shift (MS) scheme. The simulation results demonstrate
that the MAT of the MMS scheme is shorter than that of the MS scheme in both
single-user and multi-user environments.
1 Introduction
In code division multiple access (CDMA)-based systems, the data demodulation is pos-
sible only after a code synchronization is completed. Therefore, the code synchroniza-
tion is one of the most important tasks in CDMA-based systems [1]. Generally, the
code synchronization consists of two stages: code acquisition and tracking. Achieving
the code synchronization is called code acquisition and maintaining the code synchro-
nization is called tracking [2], of which the former is dealt with in this paper. In code
acquisition process, the most significant performance measure is the mean acquisition
time (MAT), which is a mean time that elapses prior to acquisition.
An optical CDMA system uses a spreading code called optical orthogonal code
(OOC) proposed by Salehi [3]. Due to its good auto-correlation and cross-correlation
properties, the OOC has been widely used for various CDMA-based systems including
optical CDMA systems [4], [5]. Keshavarzian and Salehi introduced the serial-search
(SS) [4] scheme using the OOC, which is simple; however, its MAT increases as the
code length becomes longer. Thus, the SS scheme is not suitable for rapid acquisition
of a long code used in multi-user environments. In order to overcome this drawback, in
[5], the same authors proposed the multiple-shift (MS) scheme using the OOC, which
consists of two stages and offers a shorter MAT compared with that of the SS scheme.
Corresponding author.
D. Taniar et al. (Eds.): ICCSA 2010, Part III, LNCS 6018, pp. 21–30, 2010.
c Springer-Verlag Berlin Heidelberg 2010
22 Y. Lee et al.
In this paper, we propose a novel code acquisition scheme called modified multiple-
shift (MMS). The MMS scheme comprises two stages similarly to the MS scheme,
however, by using the modified local code differently from the MS scheme, it provides
a shorter MAT than that of the MS scheme.
The remainder of this paper is organized as follows. Section 2 describes the system
model. In Section 3, we present the conventional MS and proposed MMS schemes. Sec-
tion 4 analyzes the MAT performance of the MMS scheme. In Section 5, the simulation
results show the MATs of the MS and MMS schemes in single-user and multi-user en-
vironments. Section 6 concludes this paper.
2 System Model
In an optical CDMA channel, there exist various kinds of impairments such as noise,
multipath signals, and multiple access interference (MAI). The influences of noise and
multipath signals can be almost completely mitigated by using fiber-optic medium;
however, that of MAI should be alleviated in the receiver [6], [7]. In this paper, thus,
we consider a multi-user environment without the influences of noise and multipath
signals. Then, the received signal r(t) can be written as
N
r(t) = s(n) (t − τ (n) ), (1)
n=1
where s(n) (t) is the transmitted signal of the n-th user; τ (n) ∈ [0, Tb ) denotes the
time delay of the n-th user with the bit duration Tb ; and N is the number of users. We
consider the on-off-keying (OOK) modulation and assume that the bit rate is the same
for all users. Thus, the transmitted signal s(n) (t) can be expressed as
∞
(n)
s (n)
(t) = bi c(n) (t − iTb ), (2)
i=−∞
(n) F −1 (n)
where bi is the i-th binary data bit of the n-th user and c(n) (t) = j=0 aj p(t−jTc )
(n)
is the OOC of the n-th user with the chip duration Tc and sequence aj ∈ {0, 1} of
(n)
length F and weight K (the total number of ‘1’s in aj ), where the rectangular pulse
p(t) over [0, Tc ] is defined as
1, 0 ≤ t < Tc ,
p(t) = (3)
0, otherwise.
The OOC has four parameters F , K, λa , and λc , where λa and λc are auto-correlation
and cross-correlation constraints, respectively [3]. In order to maintain the strict orthog-
onality of the OOC, both λa and λc have to be zero; however, since an OOC consists
of 0 and 1, the ideal strict orthogonality cannot be satisfied. Thus, in this paper, both λa
and λc are set to 1.
A Rapid Code Acquisition Scheme for Optical CDMA Systems 23
F chips
(b) The second stage local code of the MS scheme (original OOC)
Fig. 1. The local codes of the MS scheme when OOC of (32,4,1,1) is used
3.1 MS Scheme
In the MS scheme, total F cells in the search space are divided into Q groups, each of
which contains M cells. The relation of Q and M is given by
F
Q= , (4)
M
where · denotes the ceiling operation.
The MS scheme consists of two stages. In the first stage, the received signal r(t) is
correlated with the first stage local code shown in Fig. 1. The correlation is repeated
on a group-by-group basis. If the correlation value corresponding to a certain group
exceeds a given threshold T HMS,f irst , the group is declared to be the correct group
having the time delay τ (n) and the process is transferred to the second stage. In the
second stage, the correlation-based search is performed again with the second stage
local code (original OOC) on a cell-by-cell basis over M cells in the correct group. As
in the first stage, when the correlation value corresponding to a certain cell exceeds a
given threshold T HMS,second , which is different from T HMS,f irst , the cell is declared
to be an estimate of the time delay τ (n) .
Using the definition in [5], the MAT TMS of the MS scheme is given by
Q+1 M +1
TMS = + . (5)
2 2
√
From
√ (4) and (5), we can find that the minimum value of TMS equals to F when
M = F.
24 Y. Lee et al.
F chips
(a) The first stage local code of the MMS scheme when M = 3
(b) The second stage local code of the MMS scheme (original OOC)
Fig. 2. The local codes of the MMS scheme when OOC of (32,4,1,1) is used
F chips
tmin
(c) The first stage local code of the MMS scheme when M = 4
Fig. 3. The local codes of the MS and MMS schemes when M > tmin
4 Performance Analysis
In this section, we derive the MAT expression of the MMS scheme by modeling the
acquisition process as a discrete-time Markov process [1] with the circular flow graph
diagram shown in Figs. 4 and 5. In Fig. 4, ‘ACQ’ and ‘F A’ represent the acquisition
and false alarm states, respectively, and PD (pd ) and PF A (pf a ) denote the detection and
false alarm probabilities in the first and second stages, respectively. Fig. 5 represents the
transition process diagrams of Hdet (z) in Fig. 4.
26 Y. Lee et al.
ACQ
FA
zL
s2
pfaz h(z)
s3
Hdet(z)
s1 (1-pfa)z
H(z) Hmiss(z)
PFAz
QQ s(M/2)+1
H(z) QQ-1 FA pfaz PFAz Q1 (1-PFA)z
sM/2
L
z (1-pfa)z Q2
QQ-2 h(z)
s(M/2)+2
h(z) h(z)
H(z)
sM
s(M/2)+3
Q3
Qi+1
Qi
H(z)
(a) ACQ
pdz
s1 sM/2
PDz h(z)
QQ (1-PD)z Q1
(b) ACQ
pdz
s(M/2)+1 sM
PDz h(z)
QQ (1-PD)z Q1
Let us assume that the search is started at Qi , then the transfer function between Qi
and ACQ nodes can be written as
(v)
(v) H Q−i (z)Hdet (z)
Ui (z) = . (11)
1 − Hmiss (z)H Q−1 (z)
1 Q−i
Q
Hdet (z)
U (z) = H (z), (12)
1 − Hmiss (z)H Q−1 (z) Q i=1
28 Y. Lee et al.
2 M/2 (v)
where Hdet (z) = M v=1 Hdet (z). Using the moment generating function, we can
obtain the following relationship between U (Z) and TMMS .
where E(·) denotes the statistical expectation operation. In (13), E(TMMS ) can be
computed by its moment generating function as
dU (z)
E(TMMS ) = = U (1), (14)
dz z=1
and thus, can be obtained as
Hdet (1) + Hmiss (1) 2 − pd PD
E(TMMS ) = + (Q − 1)H (1) , (15)
pd PD 2pd PD
where H (1), Hdet
(1), and Hmiss (1) can be expressed as
M
H (1) = 1 + PF A (1 + Lpf a ), (16)
2
(M/2) − 1
Hdet (1) = 2pd PD + pd PD (1 + Lpf a ), (17)
2
and
Hmiss (1) = (1 − PD ) + PD (1 − pd )[(M/2) + 1 + {(M/2) − 1}Lpf a], (18)
respectively.
When pd = PD = 1 and pf a = PF A = 0 (i.e., single-user case), (15) is re-written
as
Q+1 M +2
E(TMMS ) = + . (19)
2 4
√
Differentiating
(19), we finally obtain the optimum M and minimum TMMS as 2F
and F
2 + 1, respectively, when F 1.
5 Simulation Results
In this section, we compare the MAT of the conventional MS scheme with that of the
proposed MMS scheme in single-user and multi-user environments. Simulation pa-
rameters are as follows: F = 200; K = 5; λa = λc = 1; N = 1, 2, 3, and 4;
T HMS,f irst = T HMS,second = T HMMS,f irst = T HMMS,second = K; and L = 10.
We assume that each user transmits the data 1 or 0 with equal probability and chip
synchronization is perfect.
Fig. 6 shows the MAT performance of the MS and MMS schemes as a function
of M in a single-user environment. In Fig. 6, the dotted and solid lines represent the
simulation results of the MS and MMS schemes, respectively, and, the markers ∇ and
A Rapid Code Acquisition Scheme for Optical CDMA Systems 29
55
MS, Simulation
50 MMS, Simulation
MS, Theory
45 MMS, Theory
40
MAT 35
30
Minimum MAT
25
20
15
10
0 5 10 15 20 25
M
Fig. 6. The MATs of the MS and MMS schemes as a function of M in a single-user environment
55
MS, 2 users
50 MMS, 2 users
MS, 3 users
45 MMS, 3 users
MS, 4 users
40 MMS, 4 users
MAT
35
30
25
20
15
10
0 5 10 15 20 25
M
Fig. 7. The MATs of the MS and MMS schemes as a function of M in a multi-user environment
O represent the theoretical results of the MS and MMS schemes, respectively. From the
figure, we can observe that the MAT of the MMS scheme is shorter than that of the MS
scheme, and√ confirm that the MS and√MMS schemes provide the minimum MAT when
M = 14 ( F
14) and M = 20 ( 2F = 20), respectively. The difference between
the MATs of the MS and MMS schemes increases as M increases. In Fig. 7, we can see
the MAT performance of the MS and MMS schemes as a function of M in a multi-user
environment. From Fig. 7, we can confirm that the MAT of the MMS scheme is shorter
than that of the MS scheme, and the difference between the MATs of the MS and MMS
schemes increases as M increases as in the single-user environment.
30 Y. Lee et al.
6 Conclusion
In this paper, we have proposed a novel code acquisition scheme called MMS for optical
CDMA systems. Exploiting the modified local code, the proposed MMS scheme can
provide a shorter MAT compared with that of the MS scheme. The performance of the
MMS scheme has been analyzed using the circular flow graph diagram. The simulation
results have confirmed that the MMS scheme offers a shorter MAT compared with that
of the MS scheme in both single-user and multi-user environments.
Acknowledgments. This research was supported by the MKE (The Ministry of Knowl-
edge Economy), Korea, under the ITRC (Information Technology Research Center)
support program supervised by the NIPA (National IT Industry Promotion Agency)
(NIPA-2009-(C1090-0902-0005)).
References
1. Polydoros, A., Weber, C.L.: A unified approach to serial-search spread spectrum code
acquisition-Part I: General theory. IEEE Trans. Commun. 32(5), 542–549 (1984)
2. Chong, D., Lee, B., Kim, S., Joung, Y.-B., Song, I., Yoon, S.: Phase-shift-network-based dif-
ferential sequential estimation for code acquisition in CDMA systems. Journal of Korea In-
form. and Commun. Society 32(3), 281–289 (2007)
3. Salehi, J.A.: Code division multiple-access techniques in optical fiber networks - Part I: Fun-
damental principles. IEEE Trans. Commun. 37(8), 824–833 (1989)
4. Keshavarzian, A., Salehi, J.A.: Optical orthogonal code acquisition in fiber-optic CDMA sys-
tems via the simple serial-search method. IEEE Trans. Commun. 50(3), 473–483 (2002)
5. Keshavarzian, A., Salehi, J.A.: Multiple-shift code acquisition of optical orthogonal codes in
optical CDMA systems. IEEE Trans. Commun. 53(3), 687–697 (2005)
6. Salehi, J.A., Brackett, C.A.: Code division multiple-access techniques in optical fiber networks
- Part II: Systems performance analysis. IEEE Trans. Commun. 37(8), 834–842 (1989)
7. Stok, A., Sargent, E.H.: Lighting the local area: Optical code-division multiple access and
quality of service provisioning. IEEE Network 14(6), 42–46 (2000)
Optimal and Suboptimal Synchronization Schemes for
Ultra-Wideband Systems
1 Introduction
Ultra-wideband (UWB) is a technology that has distinct features characterized by its
wide bandwidth, and has been adopted as a standard for wireless personal area networks
(WPANs) such as IEEE 802.15.3a and IEEE 802.15.4a, due to its high data rate and
low-power transmission [1]-[4].
In UWB systems, the received signal must be first synchronized with a locally gener-
ated template prior to data demodulation, and thus, a rapid synchronization is one of the
most important technical issues in UWB systems [5], [6]. For the fast synchronization
in UWB systems, several schemes have been investigated, where timing information
corresponding to any of the received multipath components is regarded to be correct
[7]-[10].
In [7], Homier and Scholts presented synchronization schemes with various search
methods, and analytically derived their mean synchronization time in a noise-free en-
vironment, and in [8], Vijayakumaran and Wong introduced a fast synchronization
scheme with the permutation search method. Ramachandran and Roy [9] analyzed the
schemes of [7] with the tapped-delay-line channel and IEEE 802.15.3a channel models,
and Arias-de-Reyna and Acha-Catalina [10] proposed a new serial search synchroniza-
tion scheme using specific search spacing based on Fibonacci sequence. However, these
Corresponding author.
D. Taniar et al. (Eds.): ICCSA 2010, Part III, LNCS 6018, pp. 31–41, 2010.
c Springer-Verlag Berlin Heidelberg 2010
32 D. Chong et al.
schemes could cause poor performance in receiver operations such as demodulation af-
ter the synchronization, if the locally generated template would be synchronized with
a low-power multipath component. Generally, the first one among multipath compo-
nents has the largest average power. Thus, the synchronization with the first multipath
component is expected to offer a good performance in receiver operations after the syn-
chronization.
In this paper, we propose new synchronization schemes for UWB systems, allowing
us to obtain the timing information of the first multipath component efficiently. First, an
optimal scheme is proposed based on a maximum-likelihood (ML) criterion. However,
the optimal scheme requires the channel information, and thus, is difficult to implement
practically. Thus, a simpler suboptimal scheme is developed. The simulation results
show that the proposed schemes provide a substantial performance improvement over
the conventional scheme.
This paper is organized as follows. Section 2 describes the system model. In Section
3, the optimal and suboptimal schemes are proposed. Section 4 presents the simula-
tion results of the conventional and proposed schemes over several multipath channel
models. Section 5 concludes this paper with a brief summary.
2 System Model
In this paper, we consider a direct sequence (DS)-UWB system [9]. During the acquisi-
tion process, we assume that an unmodulated signal (acquisition preamble) is transmit-
ted. Then, the transmitted DS-UWB signal s(t) can be expressed as
N−1
s(t) = Ec ci p(t − iTc ), (1)
i=0
where Ec is the signal energy; ci ∈ {1, −1} is the i-th chip of a pseudo noise (PN)
sequence with a period of N chips; Tc is the chip duration; and p(t) is the second
derivative Gaussian pulse [7], [9] with a duration of Tc . At the receiver, the received
signal r(t) can be expressed as
p −1
−1 L
N
r(t) = Ec αj ci p(t − iTc − jTc − τ Tc ) + w(t), (2)
i=0 j=0
where Lp denotes the number of multipaths; αj is the channel coefficient of the j-th
1−e−μ −(j−1)μ
multipath with the average power 1−e −μLp e , where μ is a decay factor; τ is the
time delay; and w(t) is an additive white Gaussian noise (AWGN) process with mean
zero and two-sided power spectral density N20 .
In this paper, we consider a parallel receiver for UWB synchronization shown in
Fig. 1. The receiver first yields the m-th correlator output ym by correlating r(t) with
N −1
the locally generated template signal gm (t) = n=0 cn p(t − (n + m)Tc ), where
m = 0, 1, · · · , N − 1, over a correlation time N Tc , and then, takes the absolute value
of ym to remove the influence of the signal conversion due to the channel reflections.
Optimal and Suboptimal Synchronization Schemes for Ultra-Wideband Systems 33
r (t )
NTc y0 y0
³ dt
0
g 0 (t )
NTc y1 y1
³ dt
0
Synchronization
Demodulation
g1(t ) scheme
#
#
NTc yN 1 yN 1
³ dt
0
g N 1(t )
Subsequently, the synchronization process produces an estimate of the time delay τ em-
ploying the absolute correlator outputs, and finally, the demodulation process is started.
In the conventional scheme, an estimate τ̂c of the time delay is obtained as
τ̂c = arg max |ym |. (3)
0≤m≤N −1
3 Proposed Schemes
In UWB channels, the first multipath component generally has the largest power [11],
[12]. Based on which, we propose novel schemes based on the ML approach for select-
ing the first multipath component efficiently.
34 D. Chong et al.
We first calculate the mean and variance of ym to obtain the distribution of the cor-
relator output ym expressed as
N Tc
ym = r(t)gm (t)dt
0
N Tc N −1 N
p −1
−1 L
= Ec αj ci cn
0 i=0 n=0 j=0
Since τ is distributed uniformly over [0, N − 1], the optimal scheme based on the ML
approach can be expressed as
τ̂o = arg max f (y|m). (11)
0≤m≤N −1
In (11), we can remove the term which is independent from m, which results in
Lp −1
fj (ym+j )
τ̂o = arg max . (12)
0≤m≤N −1
j=0
fw (ym+j )
In (13), we can remove the terms which are independent from m and constant. Then,
the optimal scheme can be reduced as
Lp −1
τ̂o = arg max ym+j αj , (14)
0≤m≤N −1
j=0
which is optimal to select the first multipath component in UWB channel. However,
Lp −1
the optimal scheme requires information on channel coefficients {αj }j=0 , and thus,
is difficult to implement. Hence, we develop a simpler suboptimal scheme.
Lp −1
When m = τ = 0, j=0 ym+j αj of (14) is computed as
Lp −1 Lp −1
yj αj = N Ec α2j
j=0 j=0
Lp −1
= |N Ec αj ||αj |
j=0
Lp −1
= |yj ||αj | (15)
j=0
L −1
which requires the absolute values of channel coefficients. By removing {|αj |}j=0
p
Q−1
τ̂s = arg max |ym+j | (17)
0≤m≤N −1
j=0
36 D. Chong et al.
for UWB synchronization, where Q is the combining size of the correlator outputs.
Combining the correlator outputs corresponding to the paths with low-power may lead
to an increase in the noise variance rather than in the signal strength. Thus, Q is set to
the average number of paths with power attenuated by at most 10 dB from the power of
the first path, and can be predetermined if a channel model is given.
4 Simulation Results
In this section, we compare the performance of the optimal, suboptimal, and conven-
tional schemes in terms of the probability of false synchronization defined as the proba-
bility that the estimate of the scheme does not correspond to a timing information of the
first multipath component. We define the signal to noise ratio (SNR) as Ec /N0 and use a
PN sequence with a period of 255 chips and a chip duration of 0.5 ns. We simulated the
synchronization process both in the tapped-delay line channel model with a tap spac-
ing of 0.5 ns and in the IEEE 802.15.3a channel model [11], [12], where four different
environments (CM1, CM2, CM3, and CM4) are defined. As the channel environment
changes from CM1 to CM4, the multipaths spread more widely.
Figs. 2-5 show the probabilities of false synchronization of the schemes in the
tapped-delay line channel model. As shown in figures, the proposed schemes have a
better performance compared with that of the conventional scheme. Since the optimal
scheme uses the channel information, it has the best performance. On the other hand,
0
10
Probability of false synchronization
-1
10
-2
10
-3
10
Conventional
conventional1
Optimal
data2
Suboptimal
data3
-4
10
0 5 10 15 20 25 30 35 40
Ec / N 0 (dB)
0
10
-1
10
-2
10
-3
10
Conventional
conventional1
Optimal
data2
data3
Suboptimal
-4
10
0 5 10 15 20 25 30 35 40
Ec / N 0 (dB)
0
10
Probability of false synchronization
-1
10
-2
10
-3
10
Conventional
conventional1
Optimal
data2
data3
Suboptimal
-4
10
0 5 10 15 20 25 30 35 40
Ec / N 0 (dB)
0
10
-1
10
-2
10
-3
10
Conventional
conventional1
Optimal
data2
Suboptimal
data3
-4
10
0 5 10 15 20 25 30 35 40
Ec / N 0 (dB)
0
10
Probability of false synchronization
-1
10
-2
10
-3
10
Conventional
conventional1
Optimal
data2
Suboptimal
data3
-4
10
0 5 10 15 20 25 30 35 40
Ec / N 0 (dB)
0
10
-1
10
-2
10
-3
10
Conventional
conventional1
Optimal
data2
Suboptimal
data3
-4
10
0 5 10 15 20 25 30 35 40
Ec / N 0 (dB)
0
10
Probability of false synchronization
-1
10
-2
10
-3
10
conventional1
Conventional
data2
Optimal
data3
Suboptimal
-4
10
0 5 10 15 20 25 30 35 40
Ec / N 0 (dB)
0
10
-1
10
-2
10
-3
10
conventional1
Conventional
data2
Optimal
data3
Suboptimal
-4
10
0 5 10 15 20 25 30 35 40
Ec / N 0 (dB)
since the suboptimal scheme does not use the channel information, it exhibits some
performance degradation compared with the optimal scheme.
Figs. 6-9 show the probabilities of false synchronization of the schemes in the IEEE
802.15.3a channel model. As shown in figures, the proposed schemes have a better per-
formance compared with that of the conventional scheme. Another important observation
is that the performance of the suboptimal scheme is saturated. This is because each corre-
lator output in the IEEE 802.15.3a channel model includes the interference occurred by
the other correlator outputs differently from that in the tapped-delay line channel model.
However, the optimal scheme utilizes all channel information, and thus, provides the best
synchronization performance regardless of the influence of the interference.
In this paper, we have proposed novel synchronization schemes for UWB systems. We
first derive an optimal scheme based on the ML criterion and then develop a simpler
suboptimal scheme. Simulation results have shown that both proposed schemes have a
better synchronization performance than the conventional scheme. The optimal scheme
has the best performance; however, it requires the channel information. On the other
hand, since the suboptimal scheme does not require the channel information, it has
some performance degradation compared with the optimal scheme.
Optimal and Suboptimal Synchronization Schemes for Ultra-Wideband Systems 41
In the practical channel environment (i.e., IEEE 802.15.3a channel model), the sub-
optimal scheme cannot overcome the influence of the interference, and thus, its perfor-
mance is saturated. In the future, we will deal with the synchronization scheme robust
to the influence of the interference.
Acknowledgments. This research was supported by the MKE (The Ministry of Knowl-
edge Economy), Korea, under the ITRC (Information Technology Research Center)
support program supervised by the NIPA (National IT Industry Promotion Agency)
(NIPA-2009-(C1090-0902-0005)).
References
1. Win, M.Z., Scholtz, R.A.: Impulse radio: how it works. IEEE Commun. Lett. 2(2), 36–38
(1998)
2. Porcino, D., Hirt, W.: Ultra-wideband radio technology: potential and challenges ahead.
IEEE Commun. Mag. 41(7), 64–74 (2003)
3. Yang, L., Giannakis, G.B.: Ultra-wideband communications: an idea whose time has come.
IEEE Sig. Process. Mag. 21(6), 26–54 (2004)
4. Gezici, S., Kobayashi, H., Poor, H.V., Molisch, A.F.: Performance evaluation of impulse ra-
dio UWB systems with pulse-based polarity randomization. IEEE Trans. Sig. Process. 53(7),
2537–2549 (2005)
5. Aedudodla, S.R., Vijayakumaran, S., Wong, T.F.: Timing acquisition in ultra-wideband com-
munication systems. IEEE Trans. Veh. Technol. 54(5), 1570–1583 (2005)
6. Wu, L., Wu, X., Tian, Z.: Asymptotically optimal receivers with noisy template: design and
comparison with RAKE. IEEE J. Sel. Areas Commun. 24(4), 808–814 (2006)
7. Homier, E.A., Scholts, R.A.: Rapid acquisition of ultra-wideband signals in the dense multi-
path channel. In: Proc. IEEE UWBST, Baltimore, MD, pp. 105–109 (2002)
8. Vijayakumaran, S., Wong, T.F.: A search strategy for ultra-wideband signal acquisition. IEEE
Trans. Commun. 53(12), 2015–2019 (2005)
9. Ramachandran, I., Roy, S.: On acquisition of wideband direct-sequence spread spectrum
signals. IEEE Trans. Commun. 5(6), 1537–1546 (2006)
10. Arias-de-Reyna, E., Acha-Catalina, J.J.: Blind and efficient serial search strategy for ultraw-
ideband signal initial acquisition. IEEE Trans. Veh. Technol. 58(6), 3053–3057 (2009)
11. Molisch, A.F., Foerster, J.R.: Channel models for ultrawideband personal area networks.
IEEE Wireless Commun. 10(6), 14–21 (2003)
12. Molisch, A.F.: Ultrawideband propagation channels - theory, measurement, and modeling.
IEEE Trans. Veh. Technol. 54(5), 1528–1545 (2005)
Partial Information Relaying with Multi-Layered
Superposition Coding
1 Introduction
D. Taniar et al. (Eds.): ICCSA 2010, Part III, LNCS 6018, pp. 42–51, 2010.
c Springer-Verlag Berlin Heidelberg 2010
Partial Information Relaying with MLSC 43
available at source, relay and desination nodes. It is reported that partial decode-
&-forward (DF) relaying with superposition coding can increase the capacity [1],
[3], [6], but these contributions considered the partial DF under single antenna
scenario. If multiple antennas are used at source, relay and destination nodes,
it increases the degree-of-freedom to form the partial information by the MLSC
more efficiently so as to increase the overall data rate. Here, the MLSC trans-
lates source data into two types of linearly combined data, such as basic layer(s)
and superposed layer(s), where relay forwards only the superposed layer(s). The
superposed layer is decoded at destination node after relay decodes and for-
wards it. Since the basic layer is not relayed, it is decoded at destination node
after first decoding the superposed layer and successfully cancelling it out. If all
data streams (layers) are superposed ones, it is same as full information relaying
(FIR) via two-hop transmission. In the other case, if all of those streams are
basic ones, it is same as DT. Hence the proposed MLSC can be viewed as the
hybrid of FIR and DT.
If the amount of partial information to be relayed is changed in a controlled
manner, the second-hop time can be adjusted adaptively [3]. As the basic layer
is dominant than the superposed layer, the second-hop time can be made shorter
than usual and the spectral efficiency can be improved. However, if (S-D) link
condition is not sufficient enough to increase the amount of basic layer, the
amount of superposed layer should be increased accordingly. Therefore, how to
split the total power between the two layers depends on individual link gains, and
a low-rate channel state information (CSI), in terms of the signal-to-interference-
plus-noise ratio (SINR) is required at source node to perform the proposed MLSC
for partial information relaying.
The rest of the paper is organized as follows. The MIMO cooperative system
model with MLSC is described in Section 2. Section 3 explains how to set the key
design parameters of MLSC in order to maximize the overall capacity. Simulation
results and discussions are given in Section 4, and then concluding remark in
Section 5.
2 System Model
As shown in Fig.1, there are single source, relay and destination nodes. Each
node has NS , NR , ND antennas, respectively. H0 , H1 and H2 are the channel
matrices of size ND × NS , NR × NS , ND × NR associated with (S-D), (S-R) and
(R-D) links, respectively. Because of half-duplexing mode assumed here, relay
can not transmit and receive the data streams at the same time.
In the first hop, source node broadcasts the signal x to relay and destination
nodes. x is composed of xs and xb where xs are the superposed streams being
broadcast from L ≤ NS source antennas, and xb are the basic streams from the
remaining (NS − L) source antennas. Destination node stores the received signal
in the buffer. Relay node decodes only the superposed streams xs .
In the second hop, relay node forwards xr which is the same as xs if decoding
is correct. Then, destination node first decodes the received signal from relay
44 J. Kim and D.I. Kim
node for obtaining xs , and cancels out from the received signal in the first hop.
After the successive interference cancellation (SIC), destination node is able to
decode the received signal from source node for obtaining xb .
Mathematically, the above transmission and reception can be modeled as fol-
lows: P0 and P1 represent the diagonal power matrices at source and destination
nodes, respectively. n0 , n1 , n2 are additive white Gaussian noise (AWGN) vec-
tors with zero mean and σ02 , σ12 , σ22 variances.
xs
x=
xb
yr = H1 P0 x + n1
1st hop :
yd,1 = H0 P0 x + n0
2nd hop : yd,2 = H2 P1 xr + n2
Partial Information Relaying with MLSC 45
Fig. 2 shows how to forward partial information of the superposed data stream
(L = 1) on top of the basic one when NS = NR = ND = 2.
2.1 MMSE-SIC
S −L
N
NS
Rb = Rb,m = log2 1 + p0,m hH
0,m
m=1 m=L+1
−1
× H0 (m + 1)P0 (m + 1)HH
0 (m + 1) + σ02 IN h0,m (1)
L
L
Rs = Rs,m = log2 1 + p1,m hH
1,m
m=1 m=1
−1
× H1 (m + 1)P1 (m + 1)HH 2
1 (m + 1) + σ1 IN h1,m (2)
NR
R2 = R2,n
n=1
= log2 det IND + H2 P1 HH
2 (3)
In the above hi,m denotes the mth column of matrix Hi (i = 0, 1) with N rows
and M columns, Pi (m) = diag{pi,m , pi,m+1 , . . . , pi,M−1 , pi,M }, and Hi (m) is the
N × (M − m + 1) matrix [hi,m , hm+1 , . . . , hi,M−1 , hi,M ]. The superscript (·)H
denotes the hermitian transpose operation, and IN is the identity matrix of size
N × N . Note that Rb is evaluated under the assumption of perfect SIC while Rs
in the presence of the inter-stream interferences from basic layers.
The source data is divided into a number of superposed layers and the rest of
basic layers with a total power constraint of PT . If we define the power division
factor to allocate the total power to a set of superposed layers and the rest of
basic layers, they are allocated the powers α × PT and (1 − α) × PT , respectively.
The MLSC should share a total of source antennas NS among the superposed
and basic layers, and if L source antennas are used for the superposed layers
where 0 ≤ L ≤ NS , the remaining (NS − L) antennas are assigned to the basic
layers. Assuming that the SINR estimates are available at source and relay nodes,
46 J. Kim and D.I. Kim
Once Rb , Rs and R2 are evaluated, we can derive the overall capacity Rsc for
the partial information relaying with MLSC. A total number of bits that source
transmits to destination is N × (Rb + Rs ) where N denotes the duration of the
first hop. Since relay forwards only N × Rs , the duration of the second hop can
be N2 = NRR2s so that the overall capacity can be derived as [3]
where R0 is the capacity of (S-D) link, R1 the capacity of (S-R) link, and R02
the capacity for maximal ratio combining (MRC) of (S-D) and (R-D) links in
the second hop.
Partial Information Relaying with MLSC 47
3.1 Precoding
In (2) we notice that relay node decodes xs in the presence of the inter-stream
interferences from xb . Given that relay node does not need to decode xb , a
precoding matrix to decorrelate the interferences may be inserted to increase
the capacity Rsc . For this the singular value decomposition is applied to H1 as
H1 = U1 Σ1 V1H (13)
Ĥi = Hi V1 (14)
L Ns
m=1 Rs,m (α) + m=L+1 Rb,m (1 − α) R2
max Rsc (α, L) = L (15)
α,L
m=1 Rs,m (α) + R2
α=0→L=0
48 J. Kim and D.I. Kim
4 Simulation Results
For the simulation we consider the multiple antennas cooperation system ( Ns
= Nr = Nd = 4 antennas ). SNR is defined by the ratio of total power PT and
noise variance σ 2 . Fig. 3 shows the simulation result when SNRs of (S-R) and
(R-D) are all 20dB. The result of MLSC is compared with results of FIR and
DT in this graph. It is seen that MLSC shows much better performance than
other schemes.
Especially when SNR of (S-D) is around 10dB, performance gap of MLSC
and the others is significant. To explain the reason, we illustrate Fig. 6 showing
a trend of optimized α value. It is seen that the value of α gets smaller as the
SNR of (S-D) approaches high value. From this result, MLSC operates similary
as DT at the low SNR of (S-D) and FIR at high. Because small alpha indicates
that MLSC operates like DT, however large alpha does like FIR. Around the
middle of SNR of (S-D) axis, the value of alpha is close to 0.5. This result means
that when channel gain of (S-D) is some normal value (in this case 6-12dB
region), signal power of the superposed and basic layers are similar. Hence the
duration of second hop time is adaptively modulated and that causes significant
performance difference between MLSC and the other schemes.
For the asymmetric link, we illustrate Fig. 4 and Fig. 5. In Fig. 4, SNRs of
(S-R) and (R-D) are 15 and 20dB respectively and the capacity gain of MLSC at
low SNR region shows more prominent compared with Fig. 3 and Fig. 5. From
this result, MLSC can be more useful when SNRs of (S-D) and (S-R) are not so
different.
Partial Information Relaying with MLSC 49
20
18
16
Average Throughput 14
12
10
6
MLSC with Precoding
MLSC without Precoding)
4
Full Information Relaying
Direct Transmission
2
0 2 4 6 8 10 12 14 16 18 20
SNR (S−D)
Fig. 3. Average capacity when both SNRs of (S-R) and (R-D) are 20dB
16
14
Average Throughput
12
10
6
MLSC with precoding
4
MLSC without precoding
Full Information Relaying
Direct Transmission
2
0 5 10 15
SNR (S−D) [dB]
Fig. 4. Average capacity when SNRs (S-R) of and (R-D) are 15 and 20dB respectively
14
12
Average Throughput
10
Fig. 5. Average capacity when SNRs (S-R) of and (R-D) are 20 and 15dB respectively
0.8
0.7
Average alpha vaule
0.6
0.5
0.4
0.3
0.2
0.1
0
0 5 10 15 20
SNR (S−D) [dB]
Fig. 6. Average α value versus SNR of (S-D) when both SNRs of (S-R) and (R-D) are
20dB
5 Conclusion
We have proposed the partial information relaying based on MLSC where relay
forwards only a part of source data. By controlling the amount of information to
be relayed, considering the asymmetric link conditions, the second-hop time at
relay can be adjusted adaptively. Although finding an optimized power division
factor causes some overhead to the proposed relaying scheme, we can achieve
much improved overall capacity. Moreover, it was shown that the precoding
method can be effective to decorrelate the inter-stream interference and yield bet-
ter capacity. Because we have only considered a single relay node, multiple-relay
Partial Information Relaying with MLSC 51
cooperative system based on the MLSC can be a future work. Here, which and
how many links are allocated to the basic and superposed layers would be an
interesting issue to be further addressed.
References
1. Yuksel, M., Erkip, E.: Broadcast Strategies for the Fading Relay Channel. In: IEEE
Proc. MILCOM 2004, October 2004, vol. 2, pp. 1060–1065 (2004)
2. Laneman, J.N., Tse, D.N.C.: Cooperative diversity in wireless networks: efficient
protocols and outage behavior. IEEE Trans. Inform. Theory 50, 3062–3080 (2004)
3. Popovski, P., de Carvalho, E.: Improving the rates in wireless relay systems through
superposition coding. IEEE Trans. Wireless Commun. 7, 4831–4836 (2008)
4. Wolniansky, P.W., Foscini, G.J., Golden, G.D., Valenzuelar, R.A.: V-BLAST: an
architecture for realizing very high data rates over the rich-scattering wireless chan-
nel. In: Proc. of URSI International Symposium on Signals, Systems and Electronics
(ISSSE 1998), Pisa, Italy, September 1998, pp. 295–300 (1998)
5. Chung, S.T., Lozano, A., Huang, H.C., Sutivong, A., Cioffi, J.M.: Approacing the
MIMO Capacity with a Low-Rate Feedback Channel in V-BLAST. EURASIP Jour-
nal on Applied Signal Processing 5, 762–771 (2004)
6. Goparaju, A.K., Wei, S., Liu, Y.: On superposition coding based cooperative diver-
sity schemes. In: Proc. 65th IEEE Vehicular Technology Conference (VTC Spring
2007), Dublin, Ireland, April 2007, pp. 1046–1050 (2007)
7. Andrea Goldsmith: Wireless Communication, Cambridge (2005)
8. Boyd, S., Vandenberghe, L.: Convex Optimization, Cambridge (2003)
Performance Measurement of the Hybrid
Prefetch Scheme on Vehicular Telematics
Networks
Junghoon Lee1 , Gyung-Leen Park1, , Youngshin Hong1 ,
In-Hye Shin1 , and Sang Joon Lee2
1
Dept. of Computer Science and Statistics, 2 Dept. of Computer Engineering
Jeju National University, 690-756, Jeju Do, Republic of Korea
{jhlee,glpark,yshong,ihshin76,sjlee}@jejunu.ac.kr
1 Introduction
D. Taniar et al. (Eds.): ICCSA 2010, Part III, LNCS 6018, pp. 52–62, 2010.
c Springer-Verlag Berlin Heidelberg 2010
Performance Measurement of the Hybrid Prefetch Scheme 53
gateway and valid, the gateway can send it immediately. Otherwise, the gateway
must refresh its local copy, either proactively or reactively.
Prefetching is a technique in which a data consumer proactively fetches from
the server the data item that is predicted to be needed in the near future [3]. The
main concern of the prefetch scheme is necessarily how to decide the prefetch
order. Not to mention, the data item many vehicles commonly need must be
prefetched first. However, the size of such information is too large to refresh the
whole set. For example, the amount of traffic information for a city or country
is too huge for a gateway to reload all data items for each update period. In the
location-dependent information service, drivers around a specific gateway are
generally interested in such information as train schedule at nearby terminals
and up-to-date traffic conditions of the route to their destinations [4]. If multiple
vehicles access the same item, the prefetch policy for the gateway must take into
account the spatial and temporal locality between the requests from different
vehicles.
Prefetch schemes are basically limited by both prediction accuracy and the
penalty for misprediction [5]. The accuracy depends on the efficiency of a prefetch
policy and the size of cache. However, considering the cost and size of installable
memory these days, the cache size is not a restriction any more. As contrast, as
the amount of data records for a city is very large, it is impossible for each gate-
way to download from the server immediately, while all records are updated at
the server almost simultaneously. Hence, the refresh order most impacts the data
validity in the gateway. The prefetch order is based on the prediction which item
is likely to be requested most and soon. In addition, the penalty of misprediction
is definitely the waste of network bandwidth, or sometimes communication cost,
and the increase in response time.
There have been a lot of cache/prefetch strategies for single mobile host, and
their concern mainly lie in mobility prediction and personalization. As contrast,
prefetch policies for the gateway must be based on the common access pattern
from the vehicles in the vicinity [6]. So, prefetch in gateways must consider
the temporal and spatial locality of data access, possibly exploiting classical
prefetch policies such as LRU (Least Recently Used) and FAR (Furthest Away
Replacement) [7]. Data prefetching and caching have many features in common,
and each policy in cache management has its counterpart in the prefetch policy.
First, in cache management, LRU evicts the item that is least recently used.
This name can be sustained in the prefetch domain if we change first to last.
That is, LRU in the prefetch policy means that the data item least recently used
will be updated last. Similarly, FAR in cache management replaces the item
which is farthest away from the reference point. In the prefetch, the item that
is closest to a reference point will be fetched first. After all, LRU and FAR can
be interchangeably used in cache and prefetch.
Combining LRU and FAR, our previous paper has proposed a hybrid prefetch
scheme which groups the data records into two sets according to whether a record
was referenced during the given previous interval or not [8]. In each set, records
are ordered by the Euclidean distance from the gateway. The hybrid scheme
54 J. Lee et al.
can fetch the closer item first in each group. Even though the record-level hit
ratio is less than the LRU scheme, the hybrid scheme can enhance the request-
level instant reply ratio. However, user requirement has a variety, not just the
basic hit or reply ratio. In this regard, this paper is to measure and analyze the
performance of the hybrid prefetch scheme in terms of diverse user satisfiability
criteria. For example, a user satisfies if at least 90 % of information is instantly
available or a user can tolerate up to 10 second delay.
This paper is organized as follows: After defining the problem in Section 1,
Section 2 describes the background of this paper and related work. Section 3
explains the experimental setup such as the target network model and the
data access pattern. After the performance measurement results are analyzed in
Section 4, Section 5 summarizes and concludes this paper with a brief introduc-
tion of future work.
city level scenario, for a gateway on which most traffic is concentrated, the links
taken by the vehicle starting from a specific gateway location for a whole day
are below 30 %. In each set, links are ordered by the Euclidean distance to
take into account spatial locality. The first set is able to gather the links that
many drivers commonly take and will be prefetched prior to the second set. The
link far away from the gateway will be fetched later within this group. In basic
LRU, when two links are not referenced yet, it is not possible to have an item
closer to the gateway fetched first. The hybrid scheme can first fetch the closer
item to account for the higher probability to be referenced within the second
group.
For a car to reach a place far away, it must start from a link near to it. In
the vehicular telematics network, the movement of a vehicle is restricted to the
road network. So, the distance between the two points can be represented either
by the Euclidean distance or by the network distance. The Euclidean distance is
the length of the straight line connecting the two end points, while the network
distance is the sum of each road segment distance along the path. In the road
network, even if a location is very close in the Euclidean distance, its network
length can be very large when there is no direct path from the start position.
FAR-E denotes the policy which prefetches first the item closest to the given
point in the Euclidean distance, while FAR-N uses the network distance instead
of the Euclidean distance. Spatial locality is well reflected by FAR schemes.
3 Experiment Setup
Global network
Apparently, one of the most frequently used services for the telematics network
is path finding. In the advanced telematics service scenario, besides the simple
route selection, a driver may want to know detailed live information along the
path, including traffic condition update, seat availability at the roadside restau-
rant, and an instant discount coupon of shopping stores that a driver may visit
during his travel. With the data, he can make a detailed plan or even try to
another route. Those data are stored in the central information server, period-
ically updated in general, and retrieved by the vehicle by way of gateways. In
the meantime, all information stored in the information server is associated and
indexed by the link ID, so the link ID is the search key to the local cache of
gateways as well as server database. Link ID, which is analogous to the street
name, is a reasonable key index, as most POI (Point Of Interest) information is
associated with a street.
In this service scenario, a vehicle first decides a path plan by the start and the
destination using a path finding algorithm such as A* embedded in the telematics
device [14]. Then, an entire route consist of link IDs is submitted to the gateway,
which searches its local cache for each link ID and waits for missing items to
be valid. The gateway cannot reply until every requested link record is ready.
Hence, the more the data miss, the poorer the response time, as more interaction
with the remote server is needed. The vehicle can request the query only when
it can reach a gateway and receive the reply when it remains connected until the
gateway returns the reply. In depends on the driver’s decision whether he will
keep waiting when the response time gets too long. However, the driver at least
knows the unavailability of information before his start.
The realistic data access pattern is very important for the performance assess-
ment of data prefetch schemes. This paper employs the location history data
collected by the the Jeju taxi telematics system which constantly tracks the lo-
cation of vehicles [15]. Each vehicle reports its GPS reading every minute to the
central server via the cellular network. During the 2.3 months’ test period which
has the largest number of enrollments, up to 1.3 M records have been collected.
The road network of Jeju area, being represented as a graph data structure, con-
sists of about 17,000 nodes, or intersections and about 25,000 links, or the road
that connects two nodes. Our analysis system processed the commercially used
digital map to make a graph structure and detailed road shape easily accessible.
We can estimate the route taken by a vehicle and assume that this vehicle re-
trieves the information on the links belonging to the route before it has started
its trip.
The experiment is based on the trajectory of each travel obtained from the
location history data. The path, the series of links, taken by a vehicle is submitted
to the data retrieval service. The snapshot locations of a vehicle do not exactly
represent the path actually taken, however, it is possible to find start and end
points of a trip by investigating the change of taxi status connected to the
tachometer. By applying the well-known A* algorithm [14] with the start and
58 J. Lee et al.
destination of a trip, the whole set of links involved in the path taken by the car
can be generated. Next, in deciding the location of a gateway, the start positions
of all trips are first examined to find the area which has the maximum number
of starts. Actually, the city hall area is such a spot. The experiment focuses on
this single gateway, as gateways run independently in general. The simulation
takes 2,190 trips (94,546 records) that start on one day from the location within
300 km radius from the gateway location, considering the transmission range of
wireless communication interface.
4 Performance Measurement
This section measures and assesses the performance of the hybrid prefetch scheme
via simulation using SMPL [16] which provides a lot of functions and libraries for
discrete event scheduling, easily combined with the commonly used compilers. The
experiment first measures the hit ratio and the instant response ratio for the pro-
posed scheme, comparing with LRU, FAR-N, FAR-E, and random prefetch de-
scribed in Section 2. Then, the reply ratio according to the satisfiability and the
time tolerance will be measured. Finally, the instant reply ratio is further analized
according to the time flow and the hop count, or the length of each request.
To focus on the prefetch strategy, the experiment also assumes that there
is just one server and its contents are updated every ten minutes. It can be
necessarily much shorter. After all, request arrivals, periodic updates at the
server, and record refreshes at the gateway are integrated into a discrete event
scheduler. In addition, the refresh time of each record distributes exponentially,
considering the server-side load as well as the network delay between the gateway
and an information server. To parameterize the refresh time, we define update
load. If this value is 1.0, it takes one update period to refresh all link records,
that is, it takes 10 minutes to refresh 25,000 records in the Jeju city scenario. If
it is 5.0, only 20 % of entire link information can be downloaded to a gateway.
1 1
"Random"
0.8 0.8 "FAR-N"
"FAR-E"
Instant reply ratio
"LRU"
"Hybrid"
0.6 0.6
Hit ratio
0.4 0.4
"Random"
"FAR-N"
0.2 "FAR-E" 0.2
"LRU"
"Hybrid"
0 0
0 0.5 1 1.5 2 2.5 3 3.5 4 4.5 5 0 0.5 1 1.5 2 2.5 3 3.5 4 4.5 5
Update load Update load
Figure 2 plots the record level hit ratio for each policy. For all schemes, the hit
ratio decreases according to the increase of update load. LRU shows the highest
hit ratio and decreases slowly according to the load increase, as it can best reflect
temporal locality and refreshes first the link referenced at least once. Even in
the road network, there exists a link the drivers are most likely to take. The hit
ratio of the random scheme gets down to zero when the update load reaches 3.0,
namely, only 1/3 of link data can be refreshed on average. The hybrid scheme is
the next to LRU, as it can partially exploit the temporal locality. The hit ratio
of the hybrid scheme is better than that of FAR-E by up to 23.1 % and worse
than that of LRU by up to 20 % when the update load is 5.0.
Figure 3 shows the instant reply ratio for the respective strategies. The re-
sponsiveness of a gateway depends on how many link data are already refreshed
when a request arrives. The gateway can reply instantly when all the requested
link data are valid. This rate can be considered as the request-level hit ratio.
Even though LRU has the highest record-level hit ratio, the hybrid scheme shows
a better instant reply ratio by up to 20.3 % when the update load is 1.5. The gap
gets smaller by up to 1.9 % along with the increase of the update load. Other
spatial-locality-based schemes such as FAR-E and FAR-N also show better per-
formance than LRU until the update load becomes 1.5 and 3.5, respectively. We
can recognize that geographical affinity can better cover the whole link set in a
trip than temporal locality. The gap between FAR-E and the hybrid scheme is
13.1 % at maximum.
Figure 4 plots the reply ratio according to the satisfiability, which denotes
the bound a user satisfies. For the satisfiability of 0.9, a user is satisfied with
the result if 90 % of link information is returned instantly. This scenario is more
practical to the real-life service model. The experiment changes the satisfiability
from 0.6 to 1.0, and the update load is set to 0.5. The hybrid scheme shows the
best performance when the satisfiability lies between 0.8 and 1.0. However, LRU,
having the highest link-level hit ratio, gets better when the satisfiability goes
below 0.8. The hybrid scheme outperforms other spatial-locality-based schemes
for all ranges. All schemes except LRU show a linear increase in the reply ratio.
1 1
0.9 0.9
0.8 0.8
Reply ratio
Reply ratio
We think that the satisfiability is generally set to be above 0.9 for a user to get
sufficient information for his/her route.
Figure 5 plots the reply ratio according to the time tolerance. If the tolerance
is 2 seconds, a user will be satisfied with the result if he/she receives whole data
records within 2 seconds. Here, the update load is set to 0.5 again. Interestingly,
Figure 5 shows that the time tolerance has no significant influence to the instant
reply ratio. This result indicates that refreshing the core set is most important
to the performance. Anyway, the hybrid scheme shows the best instant reply
ratio for all ranges, particularly outperforming FAR-E by up to 1.2% in this
parameter setup.
180 1
"Requests"
160
140 0.9
Number of requests
Figure 6 and Figure 7 show the hourly instant reply ratio. The data access
pattern has an important effect on the hit ratio and also the instant reply ratio.
In our information service model, a vehicle requests a data record according
to the destination of a trip. When trips have a common destination or more
common route pieces, the instant reply ratio will be better. For example, during
the commute, many vehicles head to an office area in the downtown and the
residential area. Figure 6 shows the hourly change in the number of requests or
trips. We can find some peaks on two commute times, namely just before and
after the office hour around 9 and around 18, respectively, in addition to the
shopping hours around 14. Around 5 o’clock, there are just 20 trips, and the
performance in this period is a little bit meaningless.
Figure 7 traces the change of the hourly instant reply ratio. The initial reply
ratio is relatively low as just a few previous access history data is available.
So, the hit ratios of FAR-E and hybrid schemes are almost same, and so are
those of random and LRU schemes. As the time passes by, each curve shows
different behavior, but he hybrid scheme seems to be less affected by the change
of access pattern. For 10 o’clock, just after the morning commute, the instant
reply ratio decrease a little bit because the destination of each trip is quite
different. Anyway, each scheme shows a slight drop in the reply ratio when
the traffic pattern changes, for example, the morning and evening commute,
shopping time, and personal time after office hour. The hybrid scheme can reach
Performance Measurement of the Hybrid Prefetch Scheme 61
600 1
"Requests"
500
0.9
0.8
300
0.7
200 "Random"
"FAR-N"
0.6 "FAR-E"
100 "LRU"
"Hybrid"
0 0.5
0 20 40 60 80 100 0 20 40 60 80 100
Hops (Number of records) Hops (Number of records)
the stable status after initialization. While LRU shows the 22.6 % of difference
in the instant reply ratio, the hybrid scheme shows just 13.2 % difference.
Figure 8 and Figure 9 plot the effect of hop counts, which is analogous to the
number of records in a request. The instant reply ratio is expected to decrease
when a request has more links, as the gateway can reply instantly only if all of
link records are valid. Among 2,190 requests, more than 500 requests contains 30
- 40 records, and the number of records in a request is concentrated in the range
of 20 through 50. 50 requests has less than 10 links, while 100 requests consists of
more than 100 hops, which means a vehicle must have taken a long trip. Figure
9 shows instant reply ratio according to the hop counts. Here, the update load
is set to 0.5. We can recognize that random and LRU schemes are more affected
by the number of records. On the contrary, the instant reply ratios of hybrid,
FAR-N, FAR-E schemes do not significantly change until the hop count reaches
70. Hence, the number of records in a request is not a critical factor in the spatial
locality-based prefetch schemes.
5 Conclusion
This paper has measured and analyzed the performance of the hybrid prefetch
scheme for gateways on the vehicular telematics network via simulation using
SMPL. The target service indexes every information by the link ID, analogous to
the street ID or name, and updates at reguarl intervals. To take advantage of both
LRU and FAR techniques, the hybrid scheme groups links according to whether
they are referenced during the given time interval or not, orders by the Euclidean
distance in each group, and fetches the referenced set first. The experiment
results first have revealed that the hybrid scheme can improve the request-level
instant reply ratio by up to 20.3 %, overcoming 20 % loss of record-level hit ratio,
compared with LRU. Second, this scheme outperforms other spatial-locality-
based schemes at all satisfiability ranges and LRU when the satisfiability is
above 0.8, also showing stable behavior for the practical time tolerance range.
Third and finally, it quickly stablizes after the prefetch memory initialization
and is less affected by the data access pattern change, demonstrating steady
instant reply ratio for the number of records in each request.
62 J. Lee et al.
References
1. Said, E., Omar, E., Robert, L.: Data prefetching algorithm in mobile environments.
European Journal of Scientific Research 28, 478–491 (2009)
2. Yu, B., Gong, J., Xu, C.: Data aggregation and roadside unit placement for a vanet
traffic information system. In: ACM VANET, pp. 49–57 (2008)
3. Denko, M.: Cooperative data caching and prefetching in wireless ad hoc networks.
International Journal of Business Data Communications and Networking 3, 1–15
(2007)
4. Liang, B., Drew, S., Wang, D.: Performance of multiuser network-aware prefetching
in heterogeneous wireless systems. Wireless Networks 15, 99–110 (2009)
5. Sato, K., Koita, T., Fukuta, A.: Broadcasted location-aware data cache for vehic-
ular applications. EURASIP Journal on Embedded Systems (2007)
6. Lee, J., Shin, I., Park, G., Kim, I., Rhee, Y.: Design of a cache management scheme
for gateways on the vehicular telematics network. In: Gervasi, O., Taniar, D., Mur-
gante, B., Laganà, A., Mun, Y., Gavrilova, M.L. (eds.) Computational Science and
Its Applications – ICCSA 2009. LNCS, vol. 5592, pp. 831–840. Springer, Heidelberg
(2009)
7. Zheng, B., Xu, J., Lee, D.: Cache invalidation and replacement strategies for
location-dependent data in mobile environments. IEEE Transactions on Comput-
ers 51, 1141–1153 (2002)
8. Lee, J., Park, G., Kim, S., Kim, H., Shin, S.: A hybrid prefetch policy for the retrieval
of line-associated information on vehicular networks. In: ACM SAC (2010) (accepted)
9. Gitzenis, S., Bambos, N.: Efficient data prefetching for power-controlled wireless
packet networks. In: International Conference on Mobile and Ubiquitous Systems:
Networking and Services, pp. 64–73 (2004)
10. Deshpande, P., Kashyap, A., Sung, C., Das, S.: Predictive methods for improved
vehicular WiFi access. In: ACM SIGMOBILE, pp. 263–276 (2009)
11. Balasubramanian, A., Mahajan, R., Venkataramani, A., Zahorjan, J.: Interactive
WiFi connectivity for moving vehicles. In: ACM SIGCOMM, pp. 427–438 (2009)
12. Zheng, B., Lee, W., Lee, D.: On semantic caching and query scheduling for mobile
nearest-neighbor search. Wireless Networks 10, 653–664 (2004)
13. Society of Automotive Engineers: Dedicated short range communication message
set dictionary. Tech. Rep. Standard J2735, SAE (2006)
14. Goldberg, A., Kaplan, H., Werneck, R.: Reach for A*: Efficient point-to-point
shortest path algorithms. MSR-TR-2005-132. Microsoft (2005)
15. Lee, J., Park, G., Kim, H., Yang, Y., Kim, P., Kim, S.: A telematics service system
based on the Linux cluster. In: Shi, Y., van Albada, G.D., Dongarra, J., Sloot, P.M.A.
(eds.) ICCS 2007. LNCS, vol. 4490, pp. 660–667. Springer, Heidelberg (2007)
16. MacDougall, M.: Simulating Computer Systems: Techniques and Tools. MIT Press,
Cambridge (1987)
17. Ting, Y., Chang, Y.: A novel cooperative caching scheme for wireless ad hoc net-
works: Group caching. In: IEEE Int’l Conference in Networking, Architecture, and
Storage, pp. 62–68 (2007)
Power Control for Soft Fractional Frequency
Reuse in OFDMA System
Young Min Kwon, Ok Kyung Lee, Ju Yong Lee, and Min Young Chung
1 Introduction
The next generation mobile broadband wireless communication systems, such
as IEEE 802.16e Mobile WiMAX and 3GPP LTE (Long Term Evolution), are
based on Orthogonal Frequency Division Multiple Access (OFDMA) to support
the high data rate service [1][2]. Mobile WiMAX and 3GPP LTE also adopt the
frequency reuse factor of one, in which each cell serves users with whole system
bandwidth. In OFDMA, however, users with the same channel simultaneously
suffer Co-Channel Interference (CCI) of neighbor cells, which may cause severe
degradation of system performance, especially in cell edge region [3]. The CCI
can be reduced with the frequency reuse factor of more than one. However, it
may reduce the system capacity [4].
Several frequency reuse schemes have been studied to mitigate the interference
of the cell edge region and to increase the system capacity [5]−[7]. The Fractional
This work was supported by National Research Foundation of Korea Grant funded by
the Korean Government (2009-0074466) and by the MKE (The Ministry of Knowl-
edge Economy), Korea, under the ITRC (Information Technology Research Center)
support program supervised by the NIPA(National IT Industry Promotion Agency
(NIPA-2009-(C1090-0902-0005)).
Corresponding author.
D. Taniar et al. (Eds.): ICCSA 2010, Part III, LNCS 6018, pp. 63–71, 2010.
c Springer-Verlag Berlin Heidelberg 2010
64 Y.M. Kwon et al.
Frequency Reuse (FFR) [8] is proposed for mobile WiMAX in IEEE 802.16e
working group, and Soft Frequency Reuse (SFR) [9] is proposed in the 3GPP
working group. FFR scheme and SFR scheme divide the available spectrum into
two reserved parts, subbands for the inner region and subbands for the outer
region. The subband for inner User Equipment (UE) is common in each cell,
and the subband for outer UEs is different among adjacent cells. Thus, UEs in
outer region do not suffer the CCI from neighbor cells within first-tier, so that
the spectral efficiency of outer region is increased. However, since FFR does
not use whole available frequency bandwidth, overall cell throughput in a cell
is lower than reuse one. SFR has two types of subband, the subband with high
transmission power level and with low transmission power level. It assigns the
subband with high power to outer users and the subband with low power to
inner users. Since SFR can use whole system band in a cell, overall cell capacity
in a cell is higher than FFR. However, overall system capacity of SFR may be
lower than that of reuse one environment.
Soft FFR scheme has been proposed to improve overall cell throughput of
FFR [10]. While FFR do not use the subbands allocated for outer region in the
adjacent cells, soft FFR uses these subbands for the inner UEs with low transmit
power. As a result, the soft FFR also use the subband with high transmit power
level and with low transmit power level like SFR. Unlike SFR, soft FFR uses the
common subband, which can guarantee the throughput of inner users. In this
paper, we investigate the performance of soft FFR in the environment, where
subbands have several transmit power levels. An outline of the rest of this paper
is as follows. In Section 2, we propose a power control in soft FFR scheme.
The result and analysis are then presented in Section 3. Finally, conclusions are
drawn in Section 4.
When frequency reuse factor is equal to one, UEs in the edge of cell may suffer
high outage probability due to the CCI from adjacent cells. When frequency
reuse factor is more than one, adjacent cells can operate on mutually orthogonal
frequency, which yields small CCI from adjacent cells. However, since a cell can
serve UEs with a part of whole system bandwidth, the spectral efficiency is low.
To compromise these approaches, frequency reuse schemes such as FFR or soft
FFR have been proposed. Both FFR and soft FFR statically partition the cell
into two distinct geographical regions: the inner region and the outer region, as
shown in Fig. 1.
FFR divides the system bandwidth into the subband for UEs in inner region
and the subband for UEs in outer region. In Fig. 1, the system bandwidth is
divided into four subbands, F0, F1, F2, and F3. In FFR, Base Station (BS)
allocates the subband with different reuse factor according to the user’s location
in a cell. The reuse factor of F0 is equal to one and F0 is allocated to users in
the cell’s inner region. UEs in the inner region are close to the serving BS and
Power Control for Soft Fractional Frequency Reuse in OFDMA System 65
far from interfering BSs, and they are served with the common subband. Also
one of three subbands, F1, F2, and F3 is allocated to UEs in the outer region
and the other subbands are used for adjacent cells. Thus, there is no co-channel
interference among frequency bands in the first-tier outer region. Since each cell
cannot use the subband for the other region in adjacent cells, overall capacity of
the system may be reduced. There is a trade-off between the spectral efficiency
of users in the outer region and overall capacity.
Inner region
Outer region
Outer
Cell 1
Inner
Cell 1
F0 F0 F1 F2 F3
Cell 2
Outer
Inner
F1
F0 F1 F2 F3
Cell 2
Cell 3
F0 F2 F3 F0
Cell 3
Outer
Inner
F0 F1 F2 F3
In FFR, since a cell uses partial subbands among overall system frequency
bands, the overall capacity may be decreased. To improve the overall capacity,
soft FFR scheme has been proposed, which can serve UEs with the subband for
the outer region of adjacent cell. As in Fig. 2, it divides the system bandwidth
into four subbands, F0, F1, F2, and F3, like FFR. While FFR does not use the
subband for outer region in the adjacent cells, soft FFR uses these subbands with
transmit power level, Pin , for the inner UEs. In soft FFR, UEs in the inner region
are served with additional subband with Pin as well as subband with Pcomm .
UEs in the outer region of soft FFR are served in the subband with transmission
power level, Pout . Since soft FFR can use all subbands in each cell, the overall
capacity of soft FFR may be larger than that of FFR. Due to the interference
by the subband with Pin , soft FFR may have lower spectral efficiency of outer
UEs as compared with FFR.
In soft FFR, we control the spectral efficiency of outer UEs by the transmission
power ratio (ρ), which is the ratio of the inner transmission power level to outer
transmission power level, that is, ρ = Pin /Pout . Soft FFR has common subband
with transmission power level, Pcomm . Since the common subband maintains
specific transmission power and is only used by inner UEs, it can guarantee a
certain amount of capacity for inner region.
66 Y.M. Kwon et al.
Fig. 2 explains the control of the transmission power in soft FFR. In soft
FFR, subbands with Pin are assigned to the inner UEs. On the other hand,
subbands with Pout , F1, F2, or F3, are assigned to the UEs in outer region,
which do not overlap with the subband of the adjacent cells. The SINR of outer
UEs is controlled by Pout , which can be dynamically adjusted according to the
population of outer UEs and the required data rate in outer region. In the next
section, we investigate the effect of power control where Pcomm has a fixed value
and Pout is adjusted according to the requirements of outer UEs.
Fixed
Inner region
Outer region
P comm.
Cell 1
power
P out
P in P in
Cell 1
F0 F1 F2 F3
F0 + F2 + F3
Fixed
Cell 2
P comm.
F1
power
P out
P in P in
Cell 2
Cell 3
F0 F1 F2 F3
F0 + F1 + F3 F2 F3 F0 + F1 + F2
Fixed
Cell 3
P comm.
power
P out
P in P in
F0 F1 F2 F3
3 Simulation Results
The effect of power control in soft FFR is evaluated through system level sim-
ulation. We consider the two-tier hexagonal cellular system with nineteen cells,
where the inter-site distance is 500 meters. Each cell is divided into two regions:
the inner region and the outer region. We assume that the ratio of the inner
region area to the outer region area is two. The 48 UEs are located with a uni-
form distribution in a cell. Two thirds of whole UEs are in the inner region, and
the rest are in the outer region. A BS allocates subband to UE by estimating
the signaling from UEs. We assume that the BS knows a serving UE’s location
exactly according to signaling. We use simulation parameters in Table 1 [2].
In the established system environment, we evaluate the average cell through-
put of soft FFR in inner region and outer region. At a TTI of simulation, the
BS in a cell gathers the location information of UEs and divides UEs into two
groups, the inner group and the outer group, and it allocates subband resources
to each UE using the proportional fair scheduling [11]. In soft FFR, the subband
with Pcomm and Pin are assigned to UEs in inner group, whereas the subband
with Pout is allocated to UEs in the outer group. The throughput of an UE is
obtained based on the Signal to Interference and Noise Ratio (SINR) of UE in
Power Control for Soft Fractional Frequency Reuse in OFDMA System 67
Parameter Value
Channel bandwidth 10 MHz
Carrier frequency 2 GHz
FFT size 1024
Number of subcarriers 601
Subcarrier spacing 15kHz
Cellular layout Hexagonal grid, 19 sites
Inter-site distance 500 m
Bandwidth 10 MHz
Log-normal shadowing 8dB
Penetration loss 20 dB
Propagation loss 128.1 + 37.6 log10 (R(km))
BS antenna gain 15 dBi
UE antenna gain 0 dBi
White noise power density -174dBm/Hz
Scheduling Proportional Fair
TTI 1 ms
M
Tcell = Tm (2)
m=1
cells of the second tier use the same subband with the target cell, the interference
of the second tier is increased. As a result, cell throughput is affected more by
the interference from the second tier than the first tier. Therefore, as ρ decreases
from 0.1 to 0.001, the outer cell throughput of soft FFR is not increased any
more although the inner cell throughput is decreased.
Fig. 3. Cell throughput of FFR and soft FFR according to power ratio, ρ = Pin /Pout
(Pcomm = 10W , total transmission power = 40W )
Now we consider the environment that soft FFR has the fixed power ratio,
ρ = 0.1. When the transmission power of common subband is fixed to 5W ,
10W , and 20W , we increase the power of subband for outer UEs, Pout , and
estimate the cell throughput in each region as Pout changes. Since the power
ratio ρ is constant, the transmission power level of subband for inner UEs, Pin ,
is also increased as Pout increases. Fig. 4 (a) shows the cell throughput of each
region in this environment. In the inner region, although Pcomm is increased, the
cell throughput is not increased significantly. Since the common subband has
frequency reuse factor with one, the signal power of serving BS and interfering
BSs are simultaneously increased as Pcomm increases. As a result, the SINR of
UEs using the common subband is less sensitive to the change of transmission
power. In the outer region, since the power ratio (ρ) is fixed, the cell throughput
does not vary although the outer transmission power level is increased. The
overall cell throughput is less affected by the transmission power level of common
subband, Pcomm , with constant power ratio (ρ).
Fig. 4 (b) shows that total transmission power increases as Pcomm increases,
when ρ is fixed to 0.1. However, the cell throughput remains almost the same.
Therefore, while we can achieve similar throughput performance, we can save
the transmission power by reducing Pcomm .
To investigate the power control by adjusting Pout , we assign the fixed trans-
mit power to Pcomm and Pin , which are 10W and 5W , respectively. In this
Power Control for Soft Fractional Frequency Reuse in OFDMA System 69
Fig. 4. Cell throughput and total transmission power of soft FFR according to the
outer transmission power level, Pout ( ρ = Pin /Pout = 0.1, Pcomm = 5W, 10W, 20W )
environment, we estimate the cell throughput in each region and the total trans-
mit power according to outer transmit power level, Pout . In Fig. 5 (a), when
Pout is increased, the cell throughput in outer region is increased due to the
increased signal power. On the other hand, the cell throughput in inner region
is decreased due to the increased interference. However, the total transmission
power increases to improve the cell throughput of outer UEs in Fig. 5 (b). Thus,
Pout can be controlled to the throughput requirement of outer UEs in soft FFR.
If a few UEs exist in outer region, we reduce the transmission power of the sub-
band used in outer region, which yields high throughput for inner region due to
70 Y.M. Kwon et al.
Fig. 5. Cell throughput and total transmission power according to Pout in soft FFR
(Pcomm = 10W , Pin = 5W )
the reduced interfering signal power. As a result, the overall throughput can be
increased and the soft FFR scheme can save the total transmission power.
4 Conclusion
In this paper, we investigated the effect when the power of subband for each
region changes in soft FFR. In the simulation, we observe that the transmis-
sion power level of common subband, Pcomm , does not influence the overall cell
throughput significantly. Thus, we save the total transmission power by reducing
Power Control for Soft Fractional Frequency Reuse in OFDMA System 71
Pcomm within the required cell throughput. In addition, we also save the total
transmission power by using the relation between Pin and Pout . If the power in
the outer region increases, the throughput in the outer region increases and the
throughput in the inner region decreases. Otherwise, the throughput in the outer
region decreases and the throughput in the inner region increases. As the power
in the outer region decreases, the overall capacity increases, since the spectral
efficiency in the inner region is higher than that in the outer region. According
to the throughput requirements in the outer region, we can reduce the power
consumption for the outer region, while maintaining high system throughput.
References
1. IEEE 802.16e-2005: IEEE Standard for Local and Metropolitan Area Networks,
Part 16: Air Interface for Fixed and Mobile BroadBand Wireless Systems (October
2005)
2. 3GPP TR 25.814 ver 7.1.0, Physical layer aspects for evolved Universal Terrestrial
Radio Access (UTRA) (Release 7) (September 2006)
3. Giuliano, R., Monti, C., Loreti, P.: Wireless technologies advances for emergency
and rural communications - WiMAX fractional frequency reuse for rural environ-
ments. IEEE Wireless Communications 15(3), 60–65 (2008)
4. Wang, Z., Stirling-Gallacher, R.A.: Frequency reuse scheme for cellular OFDM
systems. Electronics Letters 38(8), 387–388 (2002)
5. IEEE 802.20 MBWA C802.20 − 05 − 69, Air Interface Spec Final Fixed (November
2005)
6. 3GPP R1-050599, Interference Mitigation Considerations and Results on Fre-
quency Reuse, Siemens (June 2005)
7. 3GPP R1-050833, Uplink Interference Mitigation via Power Control, LG Eletrinics
(May 2006)
8. Mobile WiMAX Part I: A Technical Overview and Performance Evaluation,
WiMAX Forum (February 2006)
9. 3GPP R1-050507, Soft frequency reuse scheme for UTRAN LTE, Huawei, Tech.
Rep. (May 2005)
10. IEEE 802.16 Broadband Wireless Access Working Group, IEEE C802.16m-08/782,
Fractional Frequency Reuse in Uplink, LG Electronics (August 2008)
11. Andrews, M.: Instability of the proportional fair scheduling algorithm for HDR.
IEEE Transactions on Wireless Communications 3(5), 1422–1426 (2004)
Authentication – Based Medium Access Control
to Prevent Protocol Jamming: A-MAC
1 Introduction
Wireless Local Area Network (WLAN) provides users with mobility within a broad
coverage area with connection to networks. In addition, it is cost-efficient, allowing
ease of integration with other networks and network components. For these reasons,
WLAN is used by enterprises, government, and the military, as well as small office
and home offices. However, it has inherent security weaknesses due to wireless char-
acteristics. Furthermore, it uses the Industrial Scientific Medical (ISM) band to be
highly susceptible to interference. Security-sensitive groups are still unwilling to use
WLAN due to these weaknesses. If we use open source scanning software, such as
*
This research was supported by the MKE (The Ministry of Knowledge Economy), Korea, un-
der the ITRC (Information Technology Research Center) support program supervised by the
NIPA (National IT Industry Promotion Agency) (NIPA-2009-C1090-0902-0003).
D. Taniar et al. (Eds.): ICCSA 2010, Part III, LNCS 6018, pp. 72–84, 2010.
© Springer-Verlag Berlin Heidelberg 2010
Authentication – Based Medium Access Control to Prevent Protocol Jamming 73
KISMET, we can easily analyze the packets [1]. Many researchers have been study-
ing these issues, since these weaknesses are well known. Thus, WEP, WPA, TKIP,
802.11i standards were formed. However, the standards cannot resolve Protocol
Jamming Attacks. Furthermore, Protocol Jamming Attacks consume little energy, and
can be easily implemented. An adversary can decrease the wireless networks
throughput for a long time, especially, when jammers are hidden.
Virtual Carrier Sense (VCS) jamming and De-authentication / Disassociation jam-
ming are typical Protocol Jamming Attacks. VCS jamming attacks exploit a weakness
in the Carrier Sense Multiple Access / Collision Avoidance (CSMA/CA). CSMA/CA
systems use VCS to avoid collision. An adversary exploits the VCS process to de-
crease the wireless networks throughput. De-authentication / Disassociation jamming
attacks exploit the protocol concerned with connection in the mobile station and AP.
All stations have to go through the process of authentication / association before send-
ing data. If an AP sends a De-authentication / Disassociation frame to a station, the
stations have to terminate the connections. Unfortunately, based on the 802.11 stan-
dards, the station cannot reject the notification of these frames. Therefore, these at-
tacks are more effective and powerful than VCS attacks [2].
We consider that the cause of Protocol Jamming Attacks is authentication prob-
lems. That is, APs and mobile stations cannot authenticate RTS / CTS / De-
authentication / Disassociation frames. Even if there is an authentication process, it
can be attacked by replay-attacks. Replay attacks also decrease network throughput.
Thus, we need authentication, integrity and anti-replay-attacks features to prevent
Protocol Jamming Attacks. In this paper, the proposed A-MAC substitutes digested
message for Cyclic Redundancy Check (CRC). The digested message is created when
the Universal Hashing Message Authentication Codes (UMAC-32) are XORed with a
Hidden Sequence Number. UMAC-32 ensures the authenticity and the integrity of
transmitted messages. In addition, it can check the transmission error. So we can sub-
stitute the digested message for CRC. The Hidden Sequence Number for anti-replay
attacks is not seen in the transmitted frames. It is stored in a Sequence Table of each
correspondent. In this approach, we can achieve integrity, authentication and anti-
replay attacks security features, and prevent Protocol Jamming Attacks that degrade
network throughput, as well as checking for transmission errors, without frame size
overhead. We also introduce Key Shift & Sequence Number Select Notification
(KSSN) that notify key and SN change by shift row and shift column processes, so
that we compensate for the short 32bits digested message. Using the KSSN, each cor-
respondent can change the key and SN, if necessary. That is, the KSSN is verified by
the A-MAC, each correspondent changes the key and SN. This method is suitable for
control and management frames that are short frame and occur frequently, because the
KSSN can use a one-way key exchange mechanism, instead of four-way hand-
shaking. Thus, we can prevent VCS and De-authentication / Disassociation jamming
attacks using the A-MAC. Our simulation shows its performance and effectiveness.
The remainder of this paper is organized as follows. Section 2 describes the back-
ground of the proposed scheme and related work. We introduce adversary models, A-
MAC and KSSN process in Section 3. Security analysis and performance analysis are
presented in Section 4. Finally we conclude the paper and outline directions for future
work in Section 5.
74 J. Jeung, S. Jeong, and J. Lim
2.1 Background
802.11 wireless networks use Carrier Sense Multiple Access with Collision Avoid-
ance (CSMA/CA). The Distributed Coordination Function (DCF) is the basis of the
standard CSMA/CA access mechanism to avoid collisions. It first checks to see that
the radio link is clear before transmitting. Stations use a random back-off after each
frame, with the first sender seizing the channel, to avoid collisions. The DCF may use
the Request To Send (RTS) / Clear To Send (CTS) clearing technique to further re-
duce the possibility of collisions.
Carrier sensing is used to determine if the medium is available. Two types of car-
rier sensing functions in 802.11 manage this process: Physical Carrier Sensing and
Virtual Carrier Sensing. If either carrier sensing function indicates that the medium is
busy, the MAC reports this to higher layers [3]. Unlike wired communications, Physi-
cal Carrier Sensing cannot provide all the necessary information, due to the hidden
node problem. In Fig. 1, A is about to send frames to B. However, D cannot recognize
the situations, because A’s physical signal cannot reach D. Since A and D cannot
hear each other, they may sense the media is free at the same time and both try to
transmit simultaneously; this causes collision in the network. This is the hidden node
problem. To reduce this problem, 802.11 standards adopt RTS, CTS, and VCS
mechanisms. Fig. 1 shows these mechanisms. Activity on the medium by stations is
represented by the shaded bars, and each bar is labeled with the frame type. Inter-
frame spacing is depicted by the lack of any activity. DIFS is Distributed Inter-Frame
Space and SIFS is Short Inter-Frame Space. If A wants to send frames to B, A sends
the RTS frame to B, and then B replies with the CTS frame. Neighbor stations, C and
D can overhear RTS and CTS frames, and then defer access to the medium until the
State 3
Authenticated
& Associated
A D
C
Successful
Disassociation
association
DIFS
RTS Data
Sender A State 2
SIFS
Authenticated Deauthentication
SIFS SIFS & Unassociated
CTS ACK
Receiver B
Successful
NAV (RTS) Deauthentication
authentication
C
DIFS
NAV (CTS) State 1
D Unauthenticated
& Unassociated
Defer access
Network Allocation Vector (NAV) elapses. VCS is provided by the NAV. The NAV
is a timer that indicates the amount of time the medium will be reserved, in microsec-
onds. It is usually presented by 2 bytes. If a station sets the timer for which they ex-
pect to use the medium, other stations count down from the NAV to 0. When the
NAV reaches 0, the VCS indicates that the medium is idle. These RTS, CTS, and
VCS mechanisms will prevent collisions.
Stations have to go through three states for connection to Access Point (AP). Fig. 2
shows the state diagram of 802.11. Stations are either authenticated or unauthenti-
cated and can be associated or unassociated. These situations can be combined into
three states. State 1 is not authenticated and not associated, State 2 is authenticated
but not associated, and State 3 is authenticated and associated. Each state is a succes-
sively higher point in the development of an 802.11 connection. All stations start in
State 1, and data can be transmitted only in State 3. That is, any stations cannot send
data in Sate 1 and State 2. When a station transfers from one state to another state, it
would be vulnerable.
Any current or upcoming 802.11 standards would not help mitigate the risk of Proto-
col Jamming [4]. Researchers have tried to resolve these problems recently. Zhou
proposed a packet-by-packet authentication method [5]. The author encrypted all
frame content by a secret key, and then attached the encrypted content to the end of
the original frame. For example; A → B: {RTS, E(sID, dID, TS, SN)k}, B → A:
{CTS, E(sID, dID, TS, SN+1)k}. User A decrypts the encrypted attachment, verifies
sID, dID, TS, and SN+1. If the frame is correctly decrypted, A can commence data
transmission. However, this authentication scheme increases transmission overhead.
Assume TS and SN are 4 bytes each, the total encrypted attachment is 20 bytes (6
bytes each for sID and dID). Considering RTS / CTS frame are just 20 bytes / 16
bytes, the encrypted attachment is a considerable overhead. Karlof proposed the Tiny-
Auth packet format for wireless sensor networks [6]. However, it has a weakness in a
replay attack, because it does not use any kind of sequence number or time-stamp in
the TinySec-Auth packet. So, TinySec is not suitable in WLAN. Bellardo has a dif-
ferent view point [7]. In contrast to the authentication scheme, this scheme places two
different limits on the duration values accepted by stations. The low limit has a value
equal to the amount of time required to send an ACK frame, plus media access back-
off for that frame. The high limit has a value equal to the amount of time required to
send the largest data frame, plus the media access back-off for that frame. However,
this scheme still has a weakness in RTS/CTS flooding. Since each station cannot ver-
ify the adversary that frequently send false RTS/CTS frames, the stations defer their
transmission, and consequently the wireless networks throughput is drastically low-
ered. Some research prevents Protocol Jamming Attacks using the creation of a series
of protocol extension and replacements (e.g., WEP, WPA, 802.11i, 802.11w) [2].
However, these schemes need complicated key management, powerful computational
process and frame size overhead. They are not easy to implement.
76 J. Jeung, S. Jeong, and J. Lim
Malicious adversary models are VCS jamming attacks and De-authentication / Disas-
sociation jamming attacks. VCS jamming attacks exploit the weakness of the virtual
carrier sensing mechanism, especially RTS/CTS process. De-authentication / Disas-
sociation jamming attacks exploit the weakness of the state diagram in Fig. 2. When a
station transfers from one state to another state, a vulnerable point occurs, especially
in the De-authentication, Disassociation frames.
Fig. 3(a) shows RTS/CTS adversary and NAV adversary in VCS jamming attacks.
If a malicious adversary frequently sends a RTS frame to A, A sets the NAV timer, as
much as the duration field of RTS frame indicates. According to the VCS mechanism,
A has to defer access to the medium until the NAV elapses, since the VCS mecha-
nism indicates that the medium is busy. If a malicious adversary repeatedly sends a
RTS frame to A, station A cannot have the opportunity to access the medium. This
scenario will decrease the total wireless network throughput. We define this adversary
as the RTS/CTS Adversary. In the case of Station B, a malicious adversary sets the
NAV timer for a longer period. According to VCS mechanism, B simply thinks that
the other station is about to send a large data set. As a result, B has to defer access to
the medium during the time of the duration field. The duration field has 16 bits. An
adversary can set the NAV duration for a 216 time slot. Assume that one NAV time
slot is 1 μs; B has to wait approximately 0.06 second. This delay is considerable. If
there are many stations around the malicious adversary, the total wireless network
throughput is drastically lowered. We define this adversary as the NAV adversary.
These two adversary models apply to the CTS transmission process too.
Sender A Receiver B
M K
RTS Remove CRC RTS
UMAC
CTS CTS
De-auth SN De-auth
K
Dis-ass Dis-ass
UMAC UMAC(K,M) SN SN
the SNs agree with each other, the sender A is proved to be a legitimate station. If the
SN does not agree with each other, there are either transmission errors or Protocol
Jamming Attacks. Transmission errors are closely related to Signal to Noise Ratio
(SNR). We can easily check the SNR in 802.11 systems by checking the Received
Signal Strength Indication (RSSI). Therefore, although the RSSI is good, the SN does
not agree with each other for several times. The probability of Protocol Jamming At-
tacks becomes high.
We can apply the A-MAC to an infrastructure WLAN. For example, an adversary
sends de-authentication and disassociation frames to a certain station, the station can
verify the frames using the A-MAC mechanism. If those frames are verified, they
follow the instruction of each frame, otherwise they discard the frames. In case of
sending lots of false RTS / CTS frames to a certain station, counter measures will be
different. That is, key distribution schemes can make the method different. For exam-
ple, an AP and stations share the same secret key, the stations can derive the SN from
the frames using A-MAC. If the SNs of successive frames are incremented by one,
the station will be legitimate, otherwise the stations will be suspected as an adversary.
In contrast to the prior method, if the AP and each station share the 1:1 matching se-
cret key, neighbor stations cannot derive the SN due to a different key. But the AP
can derive the SN, and verify the frames. In that case, suppose the AP recognizes the
adversary, it can notify the protocol jamming situation to the stations. The stations re-
ceived the notification can discard the frames and cope with the situation.
A-MAC uses UMAC-32 for the digested message, instead of CRC, since UMAC-
32 ensures the authenticity and the integrity of transmitted messages. Furthermore, it
is the fastest message authentication code reported in cryptographic literature [8]. We
can derive Tag from the UMAC-32 process. Tag’s size is 32 bits, 64 bits, 96 bits, 128
bits. In this paper, A-MAC use a 32 bits Tag so that the digested message is XORed
with a 32 bits Hidden SN. Tag = H_K1 (M) ⊕ F_K1 (nonce), M is an input message,
H is a secret hash function, F is a pseudo random function, and K1 is secret random
key shared by sender and receiver. Nonce is a value that changes with each digested
message. In this paper, M is the content of the CRC removed frame, nonce is the Hid-
den SN. The proposed scheme can achieve authentication, integrity and anti-replay
Authentication – Based Medium Access Control to Prevent Protocol Jamming 79
attack without overhead. Therefore, A-MAC can overcome the weakness of Karlof’s
scheme [6] and Bellardo’s scheme [7]. Especially, it does not expand the original
frame size. This characteristic also overcomes the problem of Zhou’s scheme [5].
As using a short 32 bits digested message, A-MAC may be attacked by brute-force at-
tack that is a technique to defeat the authentication mechanism by trying successively
all the words in an exhaustive list. We propose two methods that complicate the at-
tached 32 bits message of A-MAC to mitigate this vulnerability.
At first, A-MAC does not count the initial SN from zero. If the initial SN starts
from zero, a malicious adversary can calculate {UMAC (K, M)}. Thus, the probabil-
ity of detecting the key becomes high. A-MAC derives the initial SN from the 128
bits key to mitigate this probability. Table 1. is an example of Sequence Table. We
make a 4 × 4 byte matrix using the 128 bits secret key. Each row and each column is
32 bits. The first row becomes the initial SN of the RTS frame, the second row be-
comes the initial SN of the CTS frame, the third row becomes the initial SN of the de-
authentication frame, the fourth row becomes the initial SN of the disassociation
frame. After being set, the initial SN is incremented by one every frame exchange.
In this scheme, it is not an easy for a malicious adversary to estimate the key and the
digested message.
Second, each correspondent (station or AP) can change the secret key and SN
if necessary. We define this process as Key Shift & SN Select Notification (KSSN).
Fig. 5 shows that each correspondent changes the secret key and SN simultaneously.
Sequence Table
Node De- Dis-
Key RTS CTS
ID auth ass
ABCD
EFGH
A ABCD EFGH I JK L MNOP
I JK L
MNOP
BCDE
FGH I
B BCDE FGH I JKLM NOPA
JKLM
NOPA
* * * * * *
* * * * * *
In Fig. 5, we assume that sender A wants to change the secret key and SN. A sets
the subtype of Frame Control Header to 0111. In 802.11 standards, management sub-
type 0111 and control subtype 0111 are reserved and not currently used [3]. In this
paper, we use this subtype 0111 to send the KSSN message. After setting the subtype,
sender A transmits {UMAC (K, M) ⊕ SN ⊕ KSSN} to receiver B. The received
{UMAC (K, M) ⊕ SN ⊕ KSSN} message is XORed with {UMAC (K, M) ⊕ SN} by
receiver B. B derives the KSSN from this process. B compares the last 13 bits of
KSSN with the last 13 bits of SN.
80 J. Jeung, S. Jeong, and J. Lim
Sender A Receiver B
M
K
RTS Remove CRC RTS
UMAC
CTS CTS
KSSN
De-auth De-auth
Dis-ass SN
Dis-ass
K SN UMAC(K,M) SN
KSSN KSSN
OK
1 2 3 4
1 row 2 row 3 row 4 row column column columncolumn
Fig. 6 shows the 16 bits Key Shift Notification, 3 bits SN Select Notification, and
the last 13 bits SN KSSN format. We form the 4 × 4 byte matrix using the 128 bits
secret key, as in Fig. 7. The front 8 bits are related to Shift Row. The first 2 bits indi-
cate the first row, 3 ~ 4 bits indicate the second row, 5 ~ 6 bits indicate the third row,
7 ~ 8 bits indicate the fourth row. 00 indicates four bytes left shift, 01 indicates one
byte left shift, 10 indicates two bytes left shift, 11 indicates three bytes left shift. The
9 ~ 16 bits are related to the Shift Column. The allocation method is similar to Shift
Row. The 17 ~ 19 bits indicate which row or column is selected as the new SN. If the
17th bit is 0, it selects the new SN from the row. If it is 1, it selects the new SN from
the column. 00 indicates the first row or column, 01 indicates the second row or col-
umn, 10 indicates the third row or column, 11 indicates the fourth row or column. The
last 13 bits are used for authentication. If the last 13 bits of KSSN and the last 13
bits of SN are verified, the receiver can change the secret key and SN. The sender
transmits the KSSN several times to overcome transmission error. This depends on
channel states.
Fig. 7 shows the Key Shift and SN Select operations. If the receiver derives the
KSSN (KSSN: 00011011 00011011 010 + 13 bits) from the received message, the re-
ceiver verifies the last 13 bits of KSSN comparing them to the last 13 bits of SN.
Authentication – Based Medium Access Control to Prevent Protocol Jamming 81
Shift Row
Old Key
S 3,0 S 3,1 S 3,2 S 3,3 S 3,0 S 3,1 S 3,2 S 3,3 S 3,3 S 3,0 S 3,1 S 3,2
New Key
Shift Column
If they are verified, the receiver changes the secret key and SN. The first 8 bits
(00011011) indicate the first row is shifted by 4 bytes, the second row is shifted by 1
byte, the third row is shifted by 2 bytes, and the fourth row is shifted by 3 bytes. The
next 8 bits (00011011) indicate that the first column is down-shifted by 4, the second
column is down-shifted by 1 byte, the third column is down-shifted by 2 bytes, and
the fourth column is down-shifted by 3bytes. As a result of the operations, the new
key is generated. 17~19 bits (010) indicates that the third row is the new SN of the
RTS frame. Then, other rows are sequentially allocated to other frames (e.g. the
fourth row is the new SN of CTS frame, the first row is the new SN of de-
authentication frame, and the second row is the new SN of association frame).
Hidden SN, a malicious adversary could perform Denial of Service (DoS) attacks.
Assuming that A transmits a RTS control frame to B. A → B: {FC (Frame Control
Header), D (Duration), RA (Receiver Address), TA (Transmitter Address), SN,
CRC}. An adversary can attempt to predict the SN, and then frequently sends a RTS
control frame like this. Adversary → B: {FC, D, RA, TA, SN + (i = i+1), CRC}. B
will frequently send CTS or ACK frame. As a result, the total wireless network
throughput will be lowered. However, the A-MAC can prevent these DoS attacks us-
ing the Hidden SN. Furthermore, is uses the 4 byte initial SN from the 128 bits secret
key, instead of zero. This makes it more difficult to estimate the SN. Second, A-
MAC can change the secret key and SN using the KSSN. This can prevent brute-force
attacks. A malicious adversary can try to break the digested message with the prob-
ability of a 1 over 232. Assuming that an adversary repeatedly sends a forged frame,
he could be accepted by AP after about 231. On 11Mbps 802.11b, an adversary can
send about 27,000 forgery attempts per second (where RTS = 15 μs, SIFS = 10 μs)
without considering transmission delay. Therefore, it would take over 44 hours. How-
ever, the A-MAC can change the secret key and SN frequently in a short time without
a complex key exchange mechanism. In addition, A-MAC does not have any frame
size overhead. These characteristics can overcome the weakness of Zhou’s scheme [5]
and Bellardo’s scheme [7]. Therefore, the proposed A-MAC is useful for government,
the military, and enterprises.
We analyze the performance and effectiveness of the proposed A-MAC. In the simu-
lation environment, we do not consider complex key distribution systems (e.g. Re-
mote Authentication Dial-In User Service: RADIUS), because the key distribution
process does not affect our simulation, the secret keys are presumed to have been de-
livered to the AP and stations. The jammer and AP send RTS, CTS, de-authentication,
and disassociation frames. When stations receive the frames, they process the A-MAC
operation. If those frames are verified, they follow the instruction of each frame, oth-
erwise they discard the frames.
Fig. 8 shows the simulation topology. There are one AP, four stations and one
jammer. Especially, the jammer is hidden and near the AP. All stations can hear the
RTS, CTS, de-authentication, and disassociation frame of the AP and Jammer. The
jammer starts to attack at 6 seconds. Under transmission errors, although the station is
a legitimate user, it cannot access the medium. In the real world, we can check the
RSSI level and adjust the threshold. However, we just set the threshold value to one in
this simulation. That is, we do not consider transmission error. Fig. 9 shows the
throughput under the RTS/CTS Adversary model. The jammer transmits the faked
RTS frame at the rate of 100, 200 Frames Per Second (FPS). As the faked RTS frame
increases, the throughput lowers. Under a rate of 200 FPS, the throughput falls to 5.25
Mbps. However, we can sustain the throughput under the A-MAC. Fig. 10 shows the
throughput under the NAV Adversary model. The Jammer sets the duration field to
the length of 211, 212. This attack is more serious than the previous attack. Under a
length of 212, the total throughput drastically falls to 50%, since the NAV at 212 (212 ×
1μs = 0.004 sec) has a considerable delay in 802.11 systems. However, we can also
sustain the throughput under the A-MAC. Fig. 11 shows the throughput under the
Authentication – Based Medium Access Control to Prevent Protocol Jamming 83
5 Conclusions
In this paper, we introduced the A-MAC to prevent VCS Attacks and De-
authentication / Disassociation Attacks that are typical Protocol Jamming Attacks. A-
MAC can authenticate frames using UMAC-32 and Hidden SN. A-MAC frequently
changes the secret key and SN by using shift row and shift column operations to pre-
vent the weakness of short 32 bits hashing codes. Therefore, A-MAC can achieve
integrity, authentication, and anti-replay-attacks security features, and can prevent
Protocol Jamming Attacks that degrade wireless network throughput. Our simulation
84 J. Jeung, S. Jeong, and J. Lim
shows that A-MAC can sustain the network throughput under Protocol Jamming At-
tacks. In the near future, we will study a physical layer anti-jamming scheme in
802.11 based systems. If we mitigate Protocol Jamming Attacks and physical layer at-
tacks in 802.11 based systems, security-sensitive groups (e.g. government, enter-
prises, and the military) have a preference to use WLAN. We hope that our research
will help that situation come true.
References
1. http://www.kismetwireless.net
2. Liu, C., Yu, J.: Rogue Access Point Based DoS attacks against 802.11 WLANs. In: The
Fourth AICT (2008)
3. Gast, M.S.: 802.11 Wireless Networks: The Definitive Guide, pp. 33–66. O’Reilly
Publisher, Sebastopol (2002)
4. Malekzadeh, M.: Empirical Analysis of Virtual Carrier Sense Flooding Attacks over Wire-
less Local Area Network. Journal of Computer Science (2009)
5. Zhou, Y., Wu, D., Nettles, S.: Analyzing and Preventing MAC-Layer Denial of Service
Attacks for Stock 802.11 Systems. In: Workshop on BWSA, Broadnets (2004)
6. Karlof, C., Sastry, N., Wagner, D.: TinySec: A link Layer Security Architecture for Wire-
less Sensor Networks. In: SenSys (2004)
7. Bellardo, J., Savage, S.: 802.11 Denial-of-Service Attacks: Real Vulnerabilities and Practi-
cal Solutions. In: USENIX Security Symposium (2003)
8. Krovetz, T.: RFC4418-UMAC: Message Authentication Code using Universal Hashing. In:
IEEE Network Working Group (March 2006)
Femtocell Deployment to Minimize Performance
Degradation in Mobile WiMAX Systems
1 Introduction
Mobile WiMAX (IEEE 802.16e) based on IEEE 802.16e standard has been com-
mercialized in 2006 with rapid growth of various multimedia applications [1,2].
Recently, user demands for various multimedia services such as mobile IPTV
are continuously increasing. On the other hand, it is expected that 60 percent
of voice calls and 90 percent of data services will take place in indoors [3], and
in current mobile communication systems, the signal strength transmitted from
base station (BS) can be very low in indoor environments because the signal
This work was supported by National Research Foundation of Korea Grant funded by
the Korean Government (2009-0074466) and by the MKE (The Ministry of Knowl-
edge Economy), Korea, under the ITRC (Information Technology Research Center)
support program supervised by the NIPA (National IT Industry Promotion Agency)
(NIPA-2009-(C1090-0902-0005) and NIPA-2009-(C1090-0902-0046)).
Corresponding author.
D. Taniar et al. (Eds.): ICCSA 2010, Part III, LNCS 6018, pp. 85–95, 2010.
c Springer-Verlag Berlin Heidelberg 2010
86 C.S. Kim et al.
3 System Model
To analyze the performance when femtocells are deployed in mobile WiMAX
systems, we consider a system model with 19 macrocells as shown in Fig. 3. In
our model, femtocells are uniformly deployed in the centered macrocell 1. The
attenuation of transmitted signal, called path loss, can differ according to channel
environment (e.g., indoor or outdoor). As path loss models, we apply Wireless
World Initiative New Radio (WINNER) II model to evaluate the throughput
[13]. Simulation parameters are in Table 1 [14].
To calculate the path loss between macrocell BS and UE, we use non-line of
sight (NLOS) outdoor propagation model. Also, we use NLOS indoor propaga-
tion model for calculating path loss between femtocell BS and UE. Path loss of
the indoor and outdoor signal is given as [13]:
where d is the distance between the transmitter and the receiver in meter. In or-
der to reflect the interference between macrocell and femtocell, indoor to outdoor
and outdoor to indoor path loss model are considered as follows [13].
P Lin to out (dout , din ) = P Lout (dout + din ) + 14 + 15(1 − cos(θ))2 + 0.5din ,
where Ifl emtoBS j is the interference which occurs due to the signal transmited to
l
the macrocell user l in the femtocell BS j and ImacroBS k is the interference which
occurs due to the signal transmited to the macrocell user l in the macrocell BS
k. SIR of user l in the uplink at macrocell BS i is:
i
i PmacroUE l
SIRmacroUE l = M ,
F i i
j=1 If emtoUE j + k=1,k=l ImacroUE k
where Ifi emtoUE j is the interference which occurs due to the signal scheduled
i
to transmit from the femtocell UE j to the macrocell BS i, and ImacroUE k is
the interference which occurs due to the signal scheduled to transmit from the
90 C.S. Kim et al.
Pfi emtoUE l
SIRfi emtoUE l = F M i .
i
j=1,j=l If emtoUE j + k=1 ImacroUE k
The received SIR is used to calculate channel capacity with the Shannon-Hartley
theorem. Channel capacity C is calculated as follows.
C = BW · log2 (1 + SIR),
4 Performance Evaluation
Fig. 4 shows uplink and downlink channel capacity of macrocell according to
the number of femtocells. As the number of femtocell increases, the uplink chan-
nel capacity of macrocell steeply decreases, but downlink channel capacity of
macrocell decreases more slowly than uplink channel capacity. The reason is
that uplink transmit power of macrocell UE is lower than downlink transmit
power of macrocell BS and strength of uplink interference is similar to strength
of downlink interference. In order to maintain uplink and downlink channel ca-
pacity of macrocell high, the maximum transmit power of femtocell BS and UE
is controlled. We set the transmitted power of femtocell BS and UE to 0.1 W,
0.01 W, and 0.001 W in the simulation. When the transmitted power of femtocell
UE uses 0.001 W, the decreasing slope of uplink channel capacity of macrocell is
reduced because co-channel interference from femtocell UE is decreased. On the
other hand, macrocell UE in the downlink can get higher SIR than uplink of
macrocell due to a high signal strength of macrocell BS. However, the downlink
channel capacity of macrocell has smaller improvement than uplink of macrocell
even if the transmitted power of femtocell BS uses 0.001 W.
Compared with the channel capacity of macrocell as shown in Fig. 4, uplink
and downlink channel capacity of femtocell is high as shown in Fig. 5. Femtocell
has a high received SIR because the distance between femtocell BS and UE
is shorter than the distance between macrocell BS and UE. Since transmitted
power of macrocell UE is low compared with macrocell BS, co-channel inter-
ference from macrocell UE is relatively low. Although the transmitted power
of femtocell’s UE is weakened, the uplink channel capacity of femtocell main-
tains high. Accordingly, we can increase the uplink channel capacity of macro-
cell. The downlink channel capacity of femtocell is constant because interference
Femtocell Deployment to Minimize Performance Degradation 91
(a) Downlink
(b) Uplink
(a) Downlink
(b) Uplink
5 Conclusion
References
1. IEEE 802.16e-2005: Local and Metropolitan Networks-Part 16: Air Interface for
Fixed and Mobile Broadband Wireless Access System, Amendment 2: Physical
and Medium Access Control Layers for Combined Fixed and Mobile Operation in
Licensed Bands and Corrigendum 1 (2006)
2. WiMAX Forum: Mobile WiMAX-Part I: A Technical Overview and Performance
Evaluation (2006)
3. Gordon Mansfield: Femto Cells in the Us Market-Business Drivers and Femto Cells
in Us Market-Business Drivers and Consumer Propositions. FemtoCells Europe
(2008)
4. Kim, R.Y., Kwak, S.K., Etemad, K.: WiMAX Femtocell: Requirements, Chal-
lenges, and Solutions. IEEE Commun. Magazine 47(9), 87–91 (2009)
5. Chandrasekhar, V., Andrews, J.G., Gatherer, A.: Femtocell Networks: A Survey.
IEEE Commun. Magazine 46(9), 59–67 (2008)
Femtocell Deployment to Minimize Performance Degradation 95
6. 3GPP R4-071661, Ericsson: Impact of HNB with controlled output power on macro
HSDPA capacity (2007)
7. 3GPP R4-080409, Qualcomm Europe: Simple Models for Home NodeB Interference
Analysis (2008)
8. Claussen, H.: Performance of macro- and co-channel femtocells in a hierarchical
cell structure. In: Proc. IEEE 18th International Symposium on Personal, Indoor
and Mobile Radio Communications (PIMRC 2007), pp. 3–7 (2007)
9. Yeh, S.-p., Talwa, S., Lee, S.-C., Kim, H.: WiMAX femtocells: A Perspective on
Network Architecture Capacity, and Coverage. IEEE Commun. Magazine 46(10),
58–64 (2008)
10. Sung, Y.-S., Jeon, N.-R., Woon, B.-W., Lee, J.-S., Lee, S.-C., Kim, S.-C.: Femto-
cell/Macrocell Interference Analysis for Mobile WiMAX System. In: IEEE VTS
Asia Pacific Wireless Communications Symposium (2008)
11. López-Pérz, D., Valcarce, A., Roche, G.D.L., Zhang, J.: OFDMA Femtocells:
A Roadmap on Interference Avoidance. IEEE Commun. Magazine 47(9), 41–48
(2009)
12. 3GPP TR25.820: 3G Home NodeB Study Item Technical Report. v8.2.0 (2008)
13. WINNER II WP1: WINNER II Part 1 Channel Models. IST-4-027756, D1.1.2,
V1.1 (2007)
14. WiMAX Forum: WiMAX System Evaluation Methodology. v.2.01 (2007)
A Novel Frequency Planning for Femtocells in
OFDMA-Based Cellular Networks Using Fractional
Frequency Reuse
Chang-Yeong Oh, Min Young Chung, Hyunseung Choo, and Tae-Jin Lee
1 Introduction
The concept of indoor cellular networks using home base stations (BSs), femtocells,
with low transmission power is drawing much interest. Femtocells are one of the emerg-
ing cost-effective technologies for both operators and users to enhance coverage and to
support higher data rate by bridging existing handsets and broadband wired networks.
However, deployment of femtocells in pre-existing cellular networks causes some prob-
lems to be addressed. One of them is that it requires intelligent frequency allocation for
femtocells and traditional macrocells when they operate simultaneously in the same
network [3]. So it is important to efficiently allocate frequency resources for femtocells
considering the effect of co-channel interference. Since frequency resources are usually
limited, some methods to enhance the efficiency are desirable.
Traditionally, there exist frequency allocation mechanisms which allocate frequency
resources according to frequency reuse factors (RFs) in multi-cell environment. One is
This research was supported by the MKE(The Ministry of Knowledge Economy), Korea, un-
der the ITRC(Information Technology Research Center) support program supervised by the
NIPA(National IT Industry Promotion Agency) (NIPA-2009-C1090-0902-0005) and (NIPA-
2009-C1090-0902-0046).
Corresponding author.
D. Taniar et al. (Eds.): ICCSA 2010, Part III, LNCS 6018, pp. 96–106, 2010.
c Springer-Verlag Berlin Heidelberg 2010
A Novel Frequency Planning for Femtocells in OFDMA-Based Cellular Networks 97
the shared frequency allocation mechanism, in which frequency is shared with adjacent
cells, i.e., universal frequency reuse or frequency RF of 1. The other is the orthogo-
nal frequency allocation mechanism which allocates frequency according to the pre-
determined frequency patterns among multi cells with the frequency RF of 1 or more.
The former may be viewed as more efficient because all frequency resource can be uti-
lized in each of cells. However, it can suffer from performance degradation since the
performance of users located at the edge of a cell is degraded due to co-channel inter-
ference (CCI) from neighboring cells. So the frequency RF of 3 or above is generally
employed to mitigate CCI in 2G systems. The reduction of CCI is typically gained at
the cost of efficiency in frequency resource allocation [6].
Fractional frequency reuse (FFR), which is the mixture of the shared and orthogonal
frequency allocation, becomes one of the solutions to improve frequency efficiency
in OFDMA-based cellular systems, e.g., WiMAX and 3GPP Long Term Evaluation
(LTE). In most of FFR schemes, each cell is partitioned into two regions: inner region
and outer region. The inner region around a base station (BS) in a cell can use the
frequency RF of 1 due to low CCI from adjacent cells. The outer region far from a BS
in a cell uses different frequency bands with those in adjacent cells with the frequency
RF of 3 to reduce CCI [11].
FFR can be implemented easily in OFDMA-based cellular systems because the fre-
quency band of a cell is divided into subchannels and can be handled by the unit of
subchannel, a group of orthogonal subcarriers. Han et al. propose an FFR scheme to
enhance the flexibility of frequency assignment and cell performance, in which mobile
stations (MSs) in the inner region can use the subchannels assigned to both the inner
and the outer region of a cell according to the frequency partitioning [4]. Another FFR
scheme is introduced by noting that the inner region and the outer region are differently
served by not only frequency bands but also time slots [5].
While the performance of the users located at the edge of a cell is important in
the macrocell networks with FFR, interference management between macrocells and
femtocells is a main issue in femto / macro co-deployment environment with different
channel operation strategies: orthogonal channel allocation and shared channel alloca-
tion [1]. So the frequency assignment for femtocells is one of the key issues [2]. The
authors in [2] propose that femtocells in the inner region of a macrocell use different
subcahnnels with those for macrocell users to minimize interference, and femtocells in
the outer region of a macrocell use the same subcahnnels as those for macrocell users.
They suppose cell environment without FFR, in which the frequency RF of macrocells
is greater than 1.
In this paper, we propose a frequency planning mechanism with FFR for macrocells
and femtocells, in which co-channel operation is allowed in low CCI, and orthogonal
frequency allocation is adopted in high CCI. To reduce the effect of CCI we utilize
resource allocation in both frequency bands and time slots.
The remaining part of this paper is organized as follows. Section 2 describes the
basic FFR scheme for macrocells and presents the proposed frequency planning
98 C.-Y. Oh et al.
Inner Outer
Frequency
F1
F2 F0
F1
Cell 2
time
Frequency
Cell 1
F0 F2
Cell 3 time
Frequency
F0
F0
F3 F3
time
Fig. 1. Proposed frequency band/time slots allocation with FFR for macrocells
A Novel Frequency Planning for Femtocells in OFDMA-Based Cellular Networks 99
F0 F0 F2 F0
α time ratio
femtocell F3
F1+F3 time time
time
(a) Frequency band allocation for macrocell
F2+F3
Macrocell 2
Frequency Frequency Frequency
F1 F1
+
Macrocell 1 + F2
F0 or F1+F3 F2
+
F3 F3
Macrocell 3 time time time
during their time slots. The MSs in the outer region of macrocells can also be served
simultaneously but they must use the orthogonal subchannels in order to avoid high CCI
from neighbor macro BSs. In summary, the inner region and the outer region are allo-
cated shared or separate time slots and frequency bands to mitigate intra- and inter-cell
interference among the inner and the outer regions as well as among cells.
In this section, we propose a novel frequency planning for femtocells under the macro-
cells with FFR. First of all, femtocells should be deployed harmoniously in order not
to make minor impacts on the legacy macrocells but to maximize their performance. In
our proposal, femtocells are divided into the inner group and the outer group depending
on whether a femtocell is located in the inner region or the outer region of a mcarocell.
A femto BS is assumed to support auto-configuration because it has to be able to setup
a user controlled hot-spot. A femto BS reports the SINR of the reference signal of a
macrocell to the macro BS by pilot sensing, which is similar to the macro MSs when
they do in the setup procedure. Then the macro BS determines whether the femtocell
belongs to the inner or the outer group and notifies the appropriate frequency bands to
the femto BS. We assume that all macro and femto BSs are synchronized.
We now consider three factors for the frequency planning of femtocells: frequency,
time, and spatial state. The femtocells in the inner group, i.e., inner femtocells, are
served during the service time of the macro MSs in the outer region, i.e., outer service
time, and the subchannels orthogonal to the macrocell frequency bands are allocated
100 C.-Y. Oh et al.
in order to avoid intra CCI from the macrocell that the femtocell belongs to. For the
macro MSs in the outer region, only 1/3 of the overall frequency band is utilized to
reduce CCI to adjacent macrocells. So these 2/3 unused subcahnnels can be allocated
to the inner femtocells efficiently (see Fig. 2). The CCI between the inner femtocells
and the neighboring macrocells is limited due to low transmission power of femtocells
and relatively large path loss factors, e.g., high attenuation path-loss by long distance
and wall-loss by the walls in buildings.
For the femtocells in the outer group, i.e., outer femtocells, we note that they use low
transmission power and the CCI between the macrocell and the femtocells is small. So
they are allowed to use the same frequency bands and time slots, i.e., inner service time
as those of the macro MSs in the inner region. That is, the outer femtocells can utilize
all the subchannels in the given time slots. As shown in Fig. 2(c), outer femtocells use
the full frequency band during the inner service time and 2/3 of the band during the
outer service time.
3 Performance Evaluation
We consider 2-tier cellular networks consisting of M (=19) macrocells. The macro BSs
are located at the center of each cell. Femto BSs which have uniform separation with
one another are densely deployed within a macrocell (see Fig. 3). We use the modi-
fied COST-Walfish-Ikegami (WI) urban micro model for the non-line-of-sight (NLOS)
outdoor path-loss model [10].
out
Ploss [dB] = 31.81 + 40.5 log10 (d[m]) + χσout , (1)
where d is the distance between a sender and a receiver and χσout represents the outdoor
shadowing (log-normal fading), which is characterized by the Gaussian distribution
Femtocell
Macro BS
Macrocell 1
8
19 9
18 2 10
7 3
17 1 11
6 4
16 5 12
15 13
14
Fig. 3. 2-tier cellular networks and 234 uniformly separated femtocells in a macrocell
A Novel Frequency Planning for Femtocells in OFDMA-Based Cellular Networks 101
Parameter Value
Inter macro cell distance (ICD) 1000 m
Radius of a femtocell 20 m
FFT size 1024
Total number of data subcarriers in one cell 768
Down-link symbol rate 9.76 k symbols/sec
Total frequency bandwidth in one cell 10 MHz
AWGN power density (N0 ) -174 dBm/Hz
Macro BS power 20 W
Femto BS power 20 mW
out
Outdoor Log-Normal fading (σ ) 10 dB
in
Indoor Log-Normal fading (σ ) 4 dB
with zero mean and standard deviation (σ out ). And the modified COST 231-multi wall
(MW) model one floor building [7] is employed for indoor propagation,
T
in
Ploss [dB] = 37 + 3.2 · 10 log10 (d[m]) + Ltw ntw + χσin , (2)
t=1
where d, T , Ltw , ntw , and χσin are the distance between a sender and a receiver, the
number of wall types, the wall loss according to wall type t, and the number of pen-
etrated type t walls, and the shadowing, which is characterized by the Gaussian dis-
tribution with zero mean and standard deviation (σ in ), respectively. We assume each
of the buildings with femto BSs has 1 outer (heavy) wall at 10m from the femto BS,
L1w =15dB, and 1 inner (light) wall at 5m from the femto BS, L2w =3dB. We have K data
subcarriers and Fi femtocells in macrocell i.
102 C.-Y. Oh et al.
80
70
RF of 1
60 FFR
FFR with Femtocells(proposed)
50 SINR threshold
Avg. SINR(dB)
40
30
20
10
−10
−20
0 100 200 300 400 500 600
Macro MS locations from Macro BS of a given cell(m)
Fig. 4. Average (Avg.) SINR of a macro MS according to the distance from the macro BS in a
given cell
The downlink SINR of user l in macrocell i ∈ {1, .., M } on subcarrier k ∈ {1, .., K}
is
(i)
(i) Pl,k
SIN Rl,k = , (3)
M
Fi
cηk · Ikη + cjk · Ikj + N0 Δf
η=1, η=i j=1
(i)
where Pl,k , Ikη , Ikj , and N0 Δf are the received power of user l on subcarrier k in
macrocell i, CCI from macrocell η ∈ {1, .., M }, CCI from femtocell j ∈ {1, .., Fi } in
macrocell i, and Additive White Gaussian Noise (AWGN) on a subcareier, respectively.
And ck ∈ {0, 1} is defined as the collision coefficient having the value 1 if subcarrier k
is allocated to a specific cell so that it generates CCI to a reference cell, i.e., macrocell
i, and 0 otherwise. The downlik SINR of user m on subcarrier k in femtocell j of
macrocell i can be found as
(i,j)
(i,j) Pm,k
SIN Rm,k = , (4)
M
Fi
cηk · Ikη + cνk · Ikν + N0 Δf
η=1 ν=1, ν=j
(i,j)
where Pm,k and Ikν are the received power of user m on subcarrier k in femtocell j of
macrocell i, CCI from femtocell ν ∈ {1, .., Fi } in macrocell i, respectively. We adopt
A Novel Frequency Planning for Femtocells in OFDMA-Based Cellular Networks 103
6
x 10
18
16
14
0
0 5 10 15 20
SINRth(dB)
Modulation and Coding Scheme (MCS) in Table 1 [8] and use simulation parameters in
Table 2. The simplified proportional fairness allocation algorithm with the same power
for subcarriers is considered in [9] to allocate subcarriers.
We present the average (Avg.) SINR variation of a macro MS in Fig. 4 as it moves on
the x-axis from the center to the border when the macro BS is positioned at (x, y)=(0,0).
We consider the case where all macrocells use the same frequency band without femto-
cells, i.e., frequency RF of 1. In this case, we observe the Avg. SINR of a macro MS at
the cell edge is relatively low as we expect. On the other hand, the basic FFR scheme
without femtocells, shows that the Avg. SINR of a macro MS at the cell edge is greater
than that of the RF of 1 due to CCI avoidance from tier-1 cells. We set the SINR thresh-
old 5dB as the criterion to distinguish the inner and the outer region. We can observe
the Avg. SINR of a macro MS with 234 femtocells is similar to that without femtocells.
In the inner region there does not exist any effect of the deployment of femtocells at all
because the inner femto BSs and the macro BS operate at different time slots. The outer
femto BSs with co-channel operation do not interfere with the macro MSs much due to
their low transmission power, high attenuation by long distance, and wall-losses. In the
outer region, there is no additional CCI because all femto BSs and the macro BS are
served at different frequency bands. We can conclude the performance of macro MSs
with the numerous co-existing femtocells is not sacrificed in our proposed mechanism.
In Fig. 5, we evaluate the downlink throughput of a macrocell according to different
SINR thresholds when there exist 36 active macro MSs in the cell. We compare ours
104 C.-Y. Oh et al.
9
x 10
6
Total cell throughput: Proposed
Total cell throughput: RF=3
5.5
5
Throughput(bps)
4.5
3.5
2.5
0 5 10 15 20
SINRth(dB)
with the scheme in [2] assuming that the RF of 3 is adopted. Note that the total band-
width of FFR is much wider than that of the RF of 3 since the inner region adopt the
RF of 1 and can use the full bandwidth for the macro MSs. So the throughput perfor-
mance of macrocell by using FFR is highly enhanced. In our proposal, as the SINR
threshold increases, Avg. SINR of the macro MSs in the inner region becomes higher.
Thus the spectral efficiency is improved. Moreover, the throughput performance of the
macro MSs in the outer region also increases thanks to the alleviation of CCI. However,
if the SINR threshold is set to a high value, the frequency resource may be insufficient
because many macro MSs may be included in the outer region but the bandwidth for
the outer region is limited.
The scheme in [2] serves femtocells and macrocell simultaneously, especially, femto-
cells in the outer region and the macrocell share the frequency bands but the femtocells
in the inner region and the macrocell use different frequency bands. If a macro MS is
in an outer femtocell, it may experience high interference from the femto BS due to
the co-channel operation. As the SINR threshold increases, the number of outer femto-
cells which may interfere with the macro MSs in the outer region also increases. This
degradation is observed in the inner macrocell: RF=3 in Fig. 5. However, this interfer-
ence scenario can be avoided in our proposed mechanism because macro MSs and outer
femtocells are served by orthogonal frequency bands and time slots.
Fig. 6 shows the total cell throughput for downlink, i.e., the sum of throughput of
all macro MSs and femto MSs in a given cell. Total cell throughput depends on the
A Novel Frequency Planning for Femtocells in OFDMA-Based Cellular Networks 105
throughput of femto MSs because femtocells are densely deployed and the spectral
efficiency of femto MSs are much higher than that of macro MSs. We observe a trade-
off in Fig. 5 as the SINR threshold varies. In our proposed mechanism, a femtocell in
the outer region can use wider frequency band and longer time slots than a femtocell
in the inner region while in the scheme in [2] with the RF of 3 a femtocell in the inner
region use wider frequency bands than a femtocell in the outer region. The distribution
of inner and outer femtocells is determined by the SINR threshold and the total cell
throughput is shown to vary by the degree of distribution of femtocells.
4 Conclusion
In this paper, we have proposed a novel frequency planning for femtocells in OFDMA-
based cellular networks using FFR. FFR scheme is an efficient method not only to
enhance performance of edge users in a given cell but also to increase the spectral effi-
ciency by adopting the mixture of reuse factors according to different region in a cell.
Moreover, the basic rule of FFR, i.e., co-channel operation in low CCI or orthogonal
channel operation in high CCI, is considered as a useful policy for frequency plan-
ning of femtocells when they are deployed in pre-existing macrocells. There may ex-
ist some difficulties and risks in sharing frequency among macrocells and femtocells,
which cause performance degradation of macro MSs resulting from the effect of the
deployment of femtocells. So we utilize unused frequency band in a given macrocell to
inner femtocells for specific time slots in order to harmoniously serve macrocells and
femtocells by maximizing diversity gains.
References
1. Lȯpez-Pėrez, D., Valcarce, A., Roche, G., Zhang, J.: OFDMA Femtocells: A Roadmap on
Interference Avoidance. IEEE Communications Magazine 47(9), 41–48 (2009)
2. Guvenc, I., Jeong, M.-R., Watanabe, F., Inamura, H.: A hybrid frequency assignment for
femtocells and coverage area analysis for co-channel operation. IEEE Communications Let-
ters 12(12), 880–882 (2008)
3. Chandrasekhar, V., Andrews, J.G.: Femtocell Networks: A Survey. IEEE Wireless Commu-
nications 15(3), 1284–1536 (2008)
4. Han, S.S., Park, J., Lee, T.-J., Ahn, H.G., Jang, K.: A New Frequency Partitioning and Al-
location of Subcarriers for Fractional Frequency Reuse in Mobile Communication Systems.
IEICE Tranactions on Communications E91-B(8), 2748–2751 (2008)
5. Giuliano, R., Monti, C., Loreti, P.: WiMAX fractional frequency reuse for rural environ-
ments. IEEE Communications Magazine 46(9), 59–67 (2008)
6. Elayoubi, S.-E., Ben Haddada, O., Fourestie, B.: Performance evaluation of frequency plan-
ning schemes in OFDMA-based networks. IEEE Tranactions on Wireless Communica-
tions 7(5), 1623–1633 (2008)
7. Nihtilä, T.: Increasing Femto Cell Throughput with HSDPA Using Higher Order Modulation.
In: Proc. INCC 2008, Lahore, Pakistan, pp. 49–53 (2008)
8. Samsung Electronics, Experimental WiBro MCS Table (2007)
106 C.-Y. Oh et al.
9. Nguyen, T.-D., Han, Y.: A Proportional Fair Algorithm with QoS Provision in Downlink
OFDMA Systems. IEEE Communications Letters 10(11), 760–762 (2006)
10. Baum, D.S., Hansen, J., Salo, J.: An interim channel model for beyond-3G systems: extend-
ing the 3gpp spatial channel model (scm). In: Proc. VTC 2005-Spring, Stockholm, Sweden,
pp. 3132–3136 (2005)
11. Huawei: Soft frequency reuse scheme for UTRAN LTE, 3GPP, R1-050507 (2005)
12. Damosso, E. (ed.): Digital Mobile Radio Towards Future Generation Systems, COST 231
final report (1998)
Association Analysis of Location Tracking Data for
Various Telematics Services*
1 Introduction
The Taxi Telematics system has operated an efficient taxi dispatch service since 2006
in Jeju island, Korea [1]. Each taxi equipped with a GPS receiver reports its location
to the central server every minute. The server is responsible for keeping such reports
from each taxi and handles the call request from a customer by finding and
dispatching a taxi closest to the call point. The tracking data obtained from the Taxi
Telematics system can be utilized for various analysis on both research and business
areas [2]. We have been doing our research to develop a data processing framework
*
This research was supported by the MKE (The Ministry of Knowledge Economy), Korea,
under the ITRC (Information Technology Research Center) support program supervised by
the IITA (Institute for Information Technology Advancement). (IITA-2009-C1090-0902-
0040).
**
Corresponding author.
D. Taniar et al. (Eds.): ICCSA 2010, Part III, LNCS 6018, pp. 107–117, 2010.
© Springer-Verlag Berlin Heidelberg 2010
108 I.–H. Shin and G.-L. Park
for an efficient analysis [3-9]. It is the exact fact that the empty taxi ratio can be
reduced by means of guiding a taxi to the spot where many passengers are waiting
for. In the project, the empty taxi ratio is around 80 % according to a survey in [6].
Thus, many taxi businesses are interested in taxis’ moving patterns in order to
decrease the consumption of fuels as well as increase the income. Today the fuels
consumption is more important to protect the environments. The real pattern analysis
using data mining is an interesting topic due to importance of the value-added
information, as shown in [10]. Especially, [4] has proposed the framework and
perform the clustering analysis for the location recommendation. The given analysis
framework consists of both of the data processing framework and the analysis
processing framework for the refined clustering. However, it does not provide
association rules effectively so far. In this regard, this paper is to propose the
association-specific analysis framework, perform the detailed analysis for association
rule mining, and provide concrete pattern information. The proposed analysis
processing framework consists of both of the given data processing framework [4]
and a novel analysis processing framework for the refined association analysis aiming
at detecting the frequent or meaningful moving patterns. This framework can help
perform even a sophisticated and a quick analysis. Particularly, the paper is to develop
the refined association analysis by means of taking into account interesting factors
such as the driving type of taxi, boarding hour, pick-up/drop-off time, pick-up/drop-
off area, boarding rate, and so on.
The result of analysis shows that the approach employed in the paper can be
extended to similar areas such as developing public transportation systems,
distribution systems, and so on.
The paper is organized as follows. Section 2 proposes the analysis process
framework for the location tracking data collected from the Taxi Telematics system.
Section 3 exhibits the summary of the results obtained from the refined association
analysis to search the taxis’ moving patterns, based on whole data, short driving data,
and long driving data, respectively. Finally, Section 4 concludes the paper.
2 Proposed Framework
This section proposes both of the data processing framework and the analysis
processing framework, in order to show effectively how to analyze the moving
patterns of taxis using association analysis. The proposed frameworks provide the
flow chart which would take a quick look around both of the data flow and the overall
analysis process. Also, it helps quickly perform the same or similar analysis, while
sparing the temporal and economic costs.
To obtain more useful information, data mining analysis process consists of the data
cleansing process, the modeling process which analyzes the refined data using a proper
method, the assessment process, the utilization process, and feedback process, as shown
in Fig. 1. As shown in Fig. 1, Modeling, fourth step of SEMMA [10], fits a predictive
model; modeling a target variable using a cluster analysis, an association analysis, a
regression model, a decision tree, a neural network, or a user-defined model. The paper
exploits an association method for data modeling. Association searches interesting
association or correlation relationships among a large set of data items.
Fig. 2 shows the analysis processing framework for the refined association rule
miming based on the taxi location records to inform their valuable moving patterns
and moreover recommend the best location and time to the taxi drivers. The
framework shows an overall process of the association analysis.
The processing framework could be expansively adapted to another transport
industry, similar to a taxi business, such as delivery by changing some variables or
generating new variables, and repairing the flow chart partially. As shown in Fig. 2, a
data analyzer should load source records first, restore them as the SAS data type,
modify the SAS dataset through the variable transformation and data filtering for data
cleaning. We can create new data extracting the paired data by a pick-up and drop-off
record using data split/merge. Then we select the interesting or significant variables
form the pared dataset by the result of clustering. We adjust each variable’s category
by the result of frequency method, resulting in TaxiType (driving type of Taxi),
SArea/EArea (pick-up area/drop-off area), SDay/EDay (pick-up day/drop-off day),
STimeR/ETimeR (pick-up time/drop-off time). TaxiUseRatioR (boarding rate in
terms of hour) variable, which means the boarding rate according to a driving hour, is
obtained by dividing the number of pick-up report into the number of total report. The
110 I.–H. Shin and G.-L. Park
analyzer completes the integrated dataset by merging original dataset and extracted
interesting factors. It also needs new dedicated dataset for association analysis
consisting of ID and Target variable. We divide association dataset into two by
TaxiType variable with ShortDriveTaxi (short-driving) and the others (long-driving).
The analyzer can return the feedback to the previous step (extraction of interesting
factors), if not detect strong association rules. There is a recurrence of such a refined
data mining analysis process, adjusting the feedback factors which mean modification
of the given factor category and detection of a new factor until obtaining the good
association rules.
Fig. 2. The analysis processing framework for the association rule mining of taxi records
Table 1 shows the extracted interesting factors, or categorical variables, obtained from
clustering and frequency method using the refined and integrated SAS dataset. Note
that each frequency should be alike within each variable’s category, or not biased. As
shown in Table 1, we omit EArea, EDay, ETimeR variables (related with drop-off),
which correspond to SArea, SDay, and STimeR (related with pick-up), due to space
limitation of the paper as well as similar frequency distribution as pick-ups. Note that
STimeR variable provides additionally the average frequency per hour in parenthesis,
due to non-equaled time range.
Association Analysis of Location Tracking Data 111
Percent
Variable Category Description Frequency
(%)
ShortDriveTaxi ~ 30minutes 77447 95.51
TaxiType LongDriveTaxi 30minutes ~ 2hours 2309 2.85
TourTaxi 2hours ~ 1331 1.64
5MinDriving 1~5minutes 44507 54.89
5-10MinDriving 5~10minutes 22254 27.44
DrivingMinR
10-15MinDriving 10~15minutes 8787 10.84
15~30MinDriving 15~30minutes 5539 6.83
LowUse 10~15% 33082 40.80
TaxiUseRatioR MiddleUse 15~20% 23389 28.84
HighUse 20~25% 24616 30.36
SAirport Ariport 3727 4.60
SOldTown Old town 43644 53.82
SArea
SNewTown New town 22929 28.28
SEtc. The others 10787 13.30
SMonday Monday 9932 12.25
STuesday Tuesday 11405 14.07
SWednesday Wednesday 12335 15.21
SDay SThursday Thursday 11582 14.28
SFriday Friday 12996 16.03
SSaturday Saturday 12033 14.84
SSunday Sunday 10804 13.32
SOfficeStart 07~10o'clock 10469 (3490) 12.91
SOfficeHour 10~18o'clock 30673 (3834) 37.83
SOfficeEnd 18~21o'clock 17434 (5811) 21.50
STimeR
SEvening 21~23o'clock 8626 (4316) 10.64
SNight 23~01o'clock 6906 (3453) 8.52
SDawn 01~07o'clock 6979 (1163) 8.61
are selected according the support level and the confidence level. The table shows that
passengers use mostly taxi at the closing time of the office hours with 15.06% of the
support level and 70.04% of the confidence level while the support level and the
confidence level is 10.64% and 100%, respectively, for the evening hours (21~23
o'clock). Also it shows that only Friday is associated with HighUse of the taxies with
5.42% of the support level and 33.80% of the confidence level.
We depict graphs to compare those strong rules in Fig. 3 and Fig. 4. In Fig. 3, each
polygon’s shape, its color, and its size are regarded as confidence, support, lift,
respectively. Fig. 4 shows the rule comparison in terms of two important factors, the
confidence level and the support level.
Table 3 shows the associations based on the short-driving taxi records. Eight strong
associations are selected. Fig. 5 and Fig. 6 show the comparative graph among those
Fig. 6. Confidence and support of the main association rules for short-driving taxies
strong rules. The main association rules mean that short-driving taxis pick up many
passengers at the close of office hours with support level 14.87% and confidence level
69.86%, respectively. Also mainly passengers move either within the old town or the
new town within 5 minutes by a taxi. Especially, many taxies picking up a passenger
move within the old town at the close of office hours.
the long driving. Table 4 shows the meaningful and interesting associations based on
the long-driving taxi records. Fig. 7 and Fig. 8 depict the main association rules.
Finally, the main association indicates that long-driving taxis mainly move from old
town to old town, especially at the close of office hours.
Fig. 8. Confidence and support of the main association rules for long-driving taxies
4 Conclusions
The paper has developed an analysis processing framework for detecting the hidden
association rules of moving objects. In order to show the effectiveness of the
framework, the framework has been applied to the real-life location history data
accumulated from the Taxi Telematics system developed in Jeju Island. The proposed
framework provides the flow chart which would have a glance around the overall
analysis process as well as help save temporal and economic costs consumed to
quickly address the same or similar mining analysis. The paper has extracted diverse
interesting categorical factors such as taxi’s driving type, driving time, driving area,
boarding rate, and so on. The approach enables us to obtain meaningful and value-
added association rules of moving objects by performing repeatedly the refined
association analysis as follow: 1) obtaining the integrated dataset through the data
cleaning process such as filtering and split-merge, 2) extracting the various
categorical variables from the integrated dataset after performing the primary analysis
such as frequency and clustering, 3) performing the method based on another
dedicated dataset created for association, and 4) deducing the meaningful and value-
added pattern information such moving pattern, or 5) returning the feedback in order
to appropriately adjust the categories of extracted factors, if desired. The approach
can be applied not only to taxi business but also to similar areas such as public
transportation services and distribution industries.
References
[1] Lee, J., Park, G., Kim, H., Yang, Y., Kim, P., Kim, S.: A telematics service system based
on the Linux cluster. In: Shi, Y., van Albada, G.D., Dongarra, J., Sloot, P.M.A. (eds.)
ICCS 2007. LNCS, vol. 4490, pp. 660–667. Springer, Heidelberg (2007)
Association Analysis of Location Tracking Data 117
[2] Hariharan, R., Toyama, K.: Project Lachesis: Parsing and modeling location histories. In:
Egenhofer, M.J., Freksa, C., Miller, H.J. (eds.) GIScience 2004. LNCS, vol. 3234, pp.
106–124. Springer, Heidelberg (2004)
[3] Lee, J., Park, G.: Design and implementation of a movement history analysis framework
for the taxi telematics system. In: Asia-Pacific Conference on Communications, pp. 1–4
(2008)
[4] Shin, I., Park, G., Saha, A., Kwak, Y., Kim, H.: Analysis of Moving Patterns of Moving
Objects with the Proposed Framework. In: Gervasi, O., Taniar, D., Murgante, B., Laganà,
A., Mun, Y., Gavrilova, M.L. (eds.) ICCSA 2009. LNCS, vol. 5593, pp. 443–452.
Springer, Heidelberg (2009)
[5] Lee, J., Hong, J.: Design and implementation of a spatial data processing engine for the
telematics network. Applied Computing and Computational Science (2008)
[6] Lee, J.: Traveling pattern analysis for the design of location-dependent contents based on
the Taxi telematics system. In: International Conference on Multimedia, Information
Technology and its Applications (2008)
[7] Liao, Z.: Real-time taxi dispatching using global positioning systems. Communication of
the ACM, 81–83 (2003)
[8] He, H., Jin, H., Chen, J., McAullay, D., Li, J., Fallon, T.: Analysis of Breast Feeding
Data Mining Methods. In: Proc. Australasian Data Mining Conference, pp. 47–52 (2006)
[9] Madigan, E.A., Curet, O.L., Zrinyi, M.: Workforce analysis using data mining and linear
regression to understand HIV/AIDS prevalence patterns. Human Resource for Health 6
(2008)
[10] Matignon, R.: Data Mining Using SAS Enterprise Miner. Wiley, Chichester (2007)
An Efficient ICI Cancellation Method for
Cooperative STBC-OFDM Systems
1 Introduction
Spatial diversity offers prominent benefits in link reliability and spectral efficiency
through the use of multiple antennas at the transmitter and/or receiver side [1], [2].
Unfortunately, the use of multiple-antenna techniques might not be practical for some
reasons; especially due to the antenna size and power constraints. For this reason,
recently, there has been significant interest in a class of techniques known as coopera-
tive communication, which allows single-antenna mobiles to reap some of the bene-
fits of multiple-antenna systems by cooperation of in-cell users [3], [4].
Since space time block codes (STBC) were developed originally for frequency flat
channels [2], an effective way to use them on frequency selective channels is to use
them along with orthogonal frequency division multiplexing (STBC-OFDM) [5]. By
using space time coded with OFDM, frequency selective channel is converted into
multiple frequency flat channels.
D. Taniar et al. (Eds.): ICCSA 2010, Part III, LNCS 6018, pp. 118–128, 2010.
© Springer-Verlag Berlin Heidelberg 2010
An Efficient ICI Cancellation Method for Cooperative STBC-OFDM Systems 119
2 System Model
The cooperative communication system considered in this paper includes one source
node, one destination node, and one of relay node. The basic premise in this work is
that source and relay have information of their own to send, and would like to cooper-
ate in order to send this information to the receiver according to the cooperative
STBC scheme as shown in Fig. 1. That is, we do not consider signal distortion in the
source-to-relay link. In Fig. 1, two consecutive symbols on subcarrier k, Xk,0 and Xk,1,
are encoded and transmitted from the source and the relay antenna, respectively, to
obtain the diversity order 2.
During even symbol period, Xk,0 and Xk,1 are transmitted from the source antenna
and relay antenna, and during odd symbol period, (-Xk,1)* and (Xk,0)* are transmitted
from the source antenna and the relay antenna, where (·)* denotes the complex conju-
gate operation.
120 K. Won et al.
time 2 s : X k ,1
time 2 s + 1: X k*,0
H k1 = H k1,0 = H k1,1
H k0 = H k0,0 = H k0,1
time 2 s : X k ,0
time 2 s + 1: − X k*,1
Yk ,0 = H k0 ⋅ X k ,0 + H k1 ⋅ X k ,1 + Wk0 ,
(1)
Yk ,1 = H k0 ⋅ ( − X k ,1 ) + H k1 ⋅ ( X k ,0 ) + Wk1 , 0 ≤ k ≤ N − 1
* *
where N is FFT size, H kj,i is channel frequency response (CFR) of received signal on
the j-th link, and Wk j is zero-mean white Gaussian noise with variance σ 2 .
From (1), we can easily obtain diversity gain by using simple linear combination as
described in [2]
Xˆ k ,0 = ( H k0 ) ⋅ Yk ,0 + H k1 ⋅ ( Yk ,1 )
* *
( 2
= H k0 + H k1
2
)⋅ X k ,0 + ( H k0 ) ⋅ Wk0 + H k1 ⋅ (Wk1 )
* *
(2)
Xˆ k ,1 = ( H k1 ) ⋅ Yk ,0 − H k0 ⋅ (Yk ,1 )
* *
( 2
= H k0 + H k1
2
)⋅ X k ,1 + ( H k1 ) ⋅ Wk0 − H k0 ⋅ (Wk1 ) .
* *
With the residual CFOs, the received signal for symbol duration 0 and 1 can be ex-
pressed as [6]
An Efficient ICI Cancellation Method for Cooperative STBC-OFDM Systems 121
N −1
Yk ,0 = Γ 00 ⋅ H k0 ⋅ X k ,0 + Γ10 ⋅ H k1 ⋅ X k ,1 + ∑ ( Γ 0k − i ⋅ H i0 ⋅ X i ,0 + Γ1k − i ⋅ H i1 ⋅ X i ,1 ) + Wk0
i =0
i≠k
N −1
Yk ,1 = −Γ 00 ⋅ H k0 ⋅ X k*,1 + Γ10 ⋅ H k1 ⋅ X k*,1 + ∑ ( −Γ 0k − i ⋅ H i0 ⋅ X i*,0 + Γ1k − i ⋅ H i1 ⋅ X i*,1 ) + Wk1 (3)
i=0
i≠k
sin (π ⋅ ε j )
Γ kj = (
⋅ exp jπε j ( N − 1 N ) − jπ ( ε j − k ) N )
(
N ⋅ sin π ( ε j − k ) N )
where εj is the residual CFO from the j-th link which is normalized by the subcarrier
spacing, and the second parts in (3) means the ICI components represented by a sum
for i ≠ k on subcarrier k caused by CFOs.
Fig. 2 shows the performance degradations caused by ICI in cooperative STBC-
OFDM systems. In spite of perfect channel estimation, from Fig. 2, we can see that
there are severe performance degradations due to the loss of orthogonality between
subcarriers.
1.E+00
1.E-01
1.E-02
1.E-03 CFO=±0.1
CFO=±0.2
CFO=±0.3
1.E-04 CFO=±0.4
Eb/No [dB]
Ignoring system complexity, for the perfect cancellation of ICI, we can simply multi-
ply the inverse of ICI matrix to the received signal matrix (zero forcing method).
However, as you can see in (4), the implementation complexity of the inversion op-
eration of 2 N × 2 N size ICI matrix is too high.
To reduce the complexity of matrix inversion operation in zero forcing method, [7]
proposed the ICI cancellation method using SMD introduced in [10]. The major por-
tion of ICI components in (5) are distributed in the diagonal or near diagonal position
of ICI matrix. Therefore, by using SMD, inversion of ICI matrix can be transmitted
into a series of inversions of its diagonal sub-blocks as illustrated in Fig. 3.
By using the SMD method, (4) can be approximated as
⎡ Yn`.0 ⎤ ⎡ A n B n ⎤ ⎡ X`n ,0 ⎤ ⎡ Wn`,0 ⎤
⎢ `* ⎥ = ⎢ ⎥⎢ ` ⎥+⎢ ` ⎥
⎢⎣ Yn,1 ⎥⎦ ⎣Cn Dn ⎦ ⎢⎣ X n ,1 ⎥⎦ ⎢⎣ Wn,1 ⎥⎦
T
X`n,i = ⎡⎣ X n ,0 X n +1,0 L X n + 2 q ,0 ⎤⎦ (6)
T
where Yn`,i = ⎡⎣ Yn ,0 Yn +1,0 L Yn + 2 q ,0 ⎤⎦ , 0 < n ≤ N − (2q + 1)
T
Wn`,i = ⎡⎣Wn,0 Wn +1,0 L Wn + 2 q ,0 ⎤⎦
where n is sparse matrix index, 2q+1 is the tap size of diagonal sub-blocks, and
An is shown in (7). an, n is the elements of A, Bn, Cn, and Dn can be obtained
similarly [7].
An Efficient ICI Cancellation Method for Cooperative STBC-OFDM Systems 123
0 0
1
2
1
N − 2q − 2
N − 2q − 1 N − 2q − 1
⎡ an , n a n , n +1 L an , n + q L 0 ⎤
⎢a an +1, n +1 M ⎥
⎢ n +1, n ⎥
⎢ M 0 ⎥
⎢ ⎥
A n = ⎢ an + q , n an + q , n + 2 q ⎥ (7)
⎢ 0 M ⎥
⎢ ⎥
⎢ M an + 2 q −1, n + 2 q −1 an + 2 q −1, n + 2 q ⎥
⎢ 0 L an + 2 q , n + q L an + 2 q , n + 2 q −1 an + 2 q , n + 2 q ⎥⎦
⎣
⎡Xˆ ⎤ ⎡ A B ⎤ −1 ⎡ Yn ,0 ⎤ ⎡ X ⎤ ⎡ A B ⎤ −1 ⎡ W ⎤
⎢ ⎥ = ⎢ n n⎥ ⎢ * ⎥ = ⎢
n ,0 n ,0 n ,0
⎥+⎢
n n
⎥ ⎢ ⎥. (8)
⎢⎣ X n ,1 ⎥⎦ ⎣ n n ⎦ ⎢⎣ Yn ,1 ⎦⎥ ⎣ n ,1 ⎦ ⎣ n n ⎦ ⎣ n ,1 ⎦
ˆ C D X C D W
In the case of the above SMD method, although the total number of complex multipli-
cations for the matrix inversion is reduced from (2N)3 to ( N − 2q ) ⋅ ( 2 ( 2q + 1) ) , re-
3
quired complexity remains high because the tap size q increases as increasing CFOs
and FFT size N.
However, considering time and frequency selectivity due to mobile movement and
multi-path delay, respectively, each subcarriers are unequally interfered each other.
Therefore, there is no necessity for the size of sparse matrix to be fixed. If each sparse
matrix has adequately small tap size q* within maximum tap size q depending on the
amount of ICI from each neighboring subcarriers, the total amount of complex multi-
plications for the matrix inversion is reduced to less than ( N − 2q ) ⋅ ( 2 ( 2q + 1) ) .
3
the efficient tap size q*. Without loss of generality, we assume that all transmitted
signals have average power 1, i.e., E[XHX]=I, where ( )H denotes the Hermitian
transpose.
In the proposed ASMD method, the decision procedure of efficient tap size q*n of
the n-th sparse matrix can be expressed as
⎧ ⎧ 2
⎫
⎪min ⎪u Γ ± ( u +1) ⋅ H k ± (u +1) < Ψ j ⎪ , u = 0,1,L q − 1,
j
⎪ ⎨ 2 q⎬
⎪ Γ ⋅ ⎪⎭
j
⎪ H
Ψ nj = ⎨ ⎩ 0 k
⎪
2
Γj ⋅H (9)
⎪ q, if ± ( u +1) k ± (2u +1) ≥ Ψ qj for all u = {0,1,L , q − 1}
⎪ Γ 0j ⋅ H k
⎩
(
Ψ qj = Γ qj + σ 2
2
) 2
Γ 0j , k = n + q
2 2
where Γ ±j ( u +1) ⋅ H k ± (u +1) Γ 0j ⋅ H k means signal to interference ratio (SIR) of the k-th
subcarrier to the ±(u+1)-th subcarrier in frequency selective fading channel with CFO,
Ψ qj is based on SINR of the k-th subcarrier to the (k+q)-th subcarrier in frequency flat
channel, and noise power σ 2 is inserted to the threshold Ψ qj to reduce unnecessary
operation in low signal to noise ratio (SNR) environments.
From (8) and (9), with the assumption of quasi-static channel for two symbol pe-
riod, efficient tap size q*n of n-th sparse matrix can be determined as follows
Therefore, by using the proposed ASMD method, the total amount of complex multi-
plications for ICI cancellation in cooperative STBC-OFDM systems can be approxi-
( (
mately reduced from ( N − 2q ) ⋅ ( 2 ( 2q +1) ) to ( N − 2q) ⋅ 1+ 2⋅ 4⋅ (q* +1) + 2( 2q* +1)
3
) ),
3
q*≤q.
For each (N-2q) times of sparse matrix, 1+ 2 ⋅ 4 ⋅ (q`+1) times of complex multiplica-
2
tion is needed for the calculation of k-th subcarrier power Γ 0j ⋅ H k and SINR
comparisons.
4 Simulation Results
In this section, we investigate the performance of the proposed method on multipath
channel for cooperative STBC-OFDM system. The simulation parameters used in this
paper are shown in Table 1. The multi-path fading channel is generated by COST 207
TU model and in order to avoid inter symbol interference (ISI) we set the length of
cyclic prefix (CP) long enough.
An Efficient ICI Cancellation Method for Cooperative STBC-OFDM Systems 125
Parameters Values
FFT size 128
Cyclic prefix (CP) length FFT size / 4
System bandwidth FFT size * 15 kHz
Subcarrier spacing 15 kHz
Center frequency 2 GHz
Channel model COST 207 TU
Channel estimation Perfect
Data modulation QPSK
Normalized frequency offsets 0.1
1.E+00
FFT size: 128
CP length: FFT size / 4
Channel model: COST 207 TU
1.E-01
CFOs: ±0.1
1.E-02
BER
1.E-03
Without CFOs
Zero forcing
SNR [dB]
Fig. 4 shows the bit error rate (BER) performances comparisons between conven-
tional SMD and proposed ASMD for ICI cancellation. From Fig. 4, we can see that
zero forcing method shows almost the same performance with and without CFOs.
Also, we can see that each performance of the proposed ASMD method is very close
to that of the conventional SMD method for the same number of taps.
Fig. 5 shows the average number of complex multiplications (ACM) and relative
ACM for matrix inversion per each subcarrier. In Fig 5, in low SNR environment,
overall ACM is dramatically reduced by using proposed ASMD, because threshold
Ψ qj depends on the noise variance σ 2 (σ 2 ≈ 1) . And in high SNR environment, Ψ qj
126 K. Won et al.
1.E+05
1.E+04
1.E+03
ACM
1.E+02
0 5 10 15 20 25 30 35
SNR [dB]
(a) ACM versus SNR
70
FFT size: 128
CP length: FFT size / 4
60 Channel model: COST 207 TU
CFOs: ±0.1
30
20
10
0
0 5 10 15 20 25 30 35
SNR [dB]
(b) Relative ACM versus SNR
6.E-05
Without CFOs FFT size: 128
CP length: FFT size / 4
Conventional SMD
Channel model: COST 207 TU
5.E-05
Proposed ASMD CFOs: ±0.1
SNR: 30 dB
Performance [BER]
4.E-05
3.E-05
2.E-05
1.E-05
0.E+00
0.E+00 1.E+04 2.E+04 3.E+04 4.E+04 5.E+04 6.E+04
Complexity [ACM]
Fig. 6. BER performances versus AMC
5 Conclusions
In this paper, we proposed an efficient ICI cancellation method for cooperative
STBC-OFDM system. Through the frequency domain SINR measurement and adap-
tive tap allocation, we showed that the overall complexity for ICI cancellation can be
efficiently reduced with approximately same performance compared with the conven-
tional method. We expect that the proposed ASMD scheme can be applicable for
various ICI suppression or cancellation methods.
128 K. Won et al.
References
1. Tarokh, V., Seshadri, N., Calderbank, A.R.: Space-Time Codes for High Data Rate Wire-
less Communication: Performance Criterion and Code Construction. IEEE Trans. Inf. The-
ory
44(2), 744–765 (1998)
2. Alamouti, S.M.: A Simple Transmit Diversity Technique for Wireless Communications.
IEEE J. Sel. Areas Commun. 16(8), 1451–1458 (1998)
3. Sendonaris, A., Erkip, E., Aazhang, B.: User Cooperation Diversity-Part 1. IEEE Trans.
Commun. 51(11), 1927–1937 (2003)
4. Laneman, J.N., Wornell, G.W.: Distributed Space Time Block Coded Protocols for Ex-
ploiting Cooperative Diversity in Wireless Networks. IEEE Trans. Inf. Theory 49, 2415–
2425 (2003)
5. Lee, K.F., Williams, D.B.: A Space-Time Coded Transmitter Diversity Technique for Fre-
quency Selective Fading Channels. In: IEEE Sensor Array and Multichannel Signal Proc-
essing Workshop, March 2000, pp. 149–152 (2000)
6. Mingqian, T.Z., Premkumar, A.B., Madhukmar, A.S.: Performance Investigation of
STBC-OFDM Systems with Frequency Offset and a Semi-Blind Approach for the Correc-
tion. In: IEEE Vehicular Technology Conference, May 2004, pp. 1836–1839 (2004)
7. Li, Z., Qu, D., Zhu, G.: An Equalization Technique for Distributed STBC-OFDM System
with Multiple Carrier Frequency Offsets. In: IEEE Wireless Communications and Net-
working Conference, April 2006, vol. 2, pp. 839–843 (2006)
8. Zhang, W., Qu, D., Zhu, G.: Performance Investigation of Distributed STBC-OFDM Sys-
tem with Multiple Carrier Frequency Offsets. In: IEEE Personal, Indoor and Mobile Radio
Communications Symposium, September 2006, pp. 1–5 (2006)
9. Kim, Y.Z., Lee, H., Chung, H.K., Cho, W.S.: An Iterative Decoding Technique for Coop-
erative STBC-OFDM systems with Multiple Frequency Offsets. In: IEEE Personal, Indoor
and Mobile Radio Communications Symposium, September 2007, pp. 1–5 (2007)
10. Jeon, W.G., Chang, K.H., Cho, Y.S.: An Equalization Technique for Orthogonal Fre-
quency-Division Multiplexing Systems in Time-Varying Multipath Channels. IEEE Trans.
Commun. 49, 27–32 (2001)
Low-Cost Two-Hop Anchor Node-Based
Distributed Range-Free Localization
in Wireless Sensor Networks
Abstract. Due to the fact that most of the sensor network applications
are based on the location information of sensor nodes, localization is
an essential research area. Some localization schemes in the literature
require sensor nodes to have additional devices to measure the distance
or angle between two sensor nodes, but it is not suitable for low-cost
sensor nodes. On the other hand, other localization schemes use only
the connectivity information of sensor nodes so that localization is not
much accurate enough. In this paper, we propose a range-free localization
scheme, called Low-cost Two-hop Anchor Node-based Distributed Range-
free Localization (LADL), which offers a higher accuracy with lower cost
than the previous works. LADL exploits a small portion of anchor nodes
which know their own location beforehand. In LADL, sensor nodes collect
the location information of the anchor nodes within two-hop distance and
calculate their own location using a grid-scan algorithm. The simulation
results show that LADL has a maximum of 12% lower delivery cost of
location information messages and 25% higher accuracy than DRLS [8].
1 Introduction
D. Taniar et al. (Eds.): ICCSA 2010, Part III, LNCS 6018, pp. 129–141, 2010.
c Springer-Verlag Berlin Heidelberg 2010
130 T. Kim et al.
2 Related Work
Centroid [5] is a fundamental scheme for range-free localization. In Centroid, the
normal nodes find their position by collecting location information of neighboring
anchor nodes and calculating average coordinate of them, but if the anchor
node ratio is low, the accuracy of estimated positions becomes very low. To
estimate the position of the normal nodes more accurately, CPE was proposed
which is a centralized range-free localization in [6]. Anchor nodes in this scheme
broadcast their position information, and normal nodes estimate their positions
by constructing an Estimative Rectangle (ER). However, traffic load for message
exchange is heavy, and thus it is not suitable for large scale sensor networks.
As mentioned above, DRLS [8] is a distributed range-free localization scheme.
It calculates a position of a normal node based on the fact that both one-hop
anchor nodes and two-hop anchor nodes can affect the position accuracy of
the normal node. Fig. 1 shows how a two-hop anchor node, A3 , can affect the
position estimation of N. A normal node can exist anywhere within the data
transmission radio range of two one-hop anchor nodes, A1 and A2 , but not
within that of the two-hop anchor node, A3 . The distribution of these anchor
nodes enables the normal node to calculate its position more accurately. However,
such a distribution of anchor nodes is not always possible.
DRLS consists of three steps: beacon message exchange, grid scan, and re-
finement steps. In the beacon message exchange step, the normal node should
know the position information of the anchor nodes existing within a two-hop
distance from their neighboring anchor nodes by two-hop flooding and one ad-
ditional broadcast. Thereafter, in the grid scan step, each normal node creates
an estimative rectangle (ER) with the use of the transmission radio range of
the one-hop anchor nodes and divides the ER into a set of small grid cells.
Fig. 2 illustrates that how a normal node obtains a one-hop anchor-based ini-
tial estimated position using a grid scan algorithm. In the refinement step, the
normal node that measured the initial position decides its own final estimated
position via vector-based refinement. The direction of the vector, called virtual
132 T. Kim et al.
force (V F ), is from a two-hop anchor node to the initial estimated position, and
−−→
its size becomes | V F |= ab · N c/ac. Fig. 3 shows how a normal node refines its
initial position based on a two-hop anchor node.
As DRLS conducts localization with the use of more information of the an-
chor nodes compared to that in the previous studies, it shows relatively high
accuracy; however, it has several problems. First, as it uses two-hop flooding to
collect the position information of two-hop anchor nodes, the message transmis-
sion cost becomes higher than that in the existing range-free methods. Second,
its square-root computation for calculating the vector in the refinement step
is not appropriate for low-cost sensor nodes [9]. Third, as it accomplishes the
vector-based refinement, its location accuracy becomes lower if the distribution
of the anchor nodes is not even. Fig.4 shows an example of the refinement which
eventually lowers the position accuracy. Suppose that the initial estimated posi-
tion of normal node N is N . The sum of two vectors made by the anchor nodes
A3 and A4 exists in an area where the normal node cannot be placed. Thus, the
position accuracy becomes low. Followings are the motivations of this study to
enhance DRLS by tackling the above-mentioned problems:
Fig. 4. The influence of farther anchors that are not distributed uniformly
3 Preliminaries
We assume the followings in this study:
– Normal nodes that do not have information of their own positions, and a few
anchor nodes that have information of their positions, constitute the entire
sensor field.
– All the sensor nodes are distributed randomly in a large sensor field, and
cannot change their positions once these have been decided.
– All the sensor nodes can be identified through their unique IDs and have the
same transmission radio and sensing ranges.
– The collision that can occur during data transmission is not considered, and
there are no interferences such as fading and noise. In other words, a sensor
node receives data without fail as it transmits data to its neighbor sensor
nodes within the transmission radio range.
134 T. Kim et al.
Anchor nodes within a one-hop distance from a normal node are called one-
hop anchor nodes while anchor nodes within a two-hop distance from a normal
node are called two-hop anchor nodes. Normal nodes can enhance localization
accuracy by separating the one-hop anchor nodes from the two-hop anchor nodes
among their neighboring anchor nodes. Fig. 5 shows where normal nodes can be
located depending on the one-hop anchor nodes and the two-hop anchor nodes.
Anchor node A is within a one-hop distance from normal node N1 , thus becoming
a one-hop anchor node of N1 . On the other hand, it becomes a two-hop anchor
node of N2 since it is two hops away from normal node N2 . Accordingly, the
region where N1 can exist is a circular area with a radius of transmission radio
range r whereas the region where N2 can exist is the circular area with a radius
of 2r minus the circular area with a radius of r.
Fig. 5. The region that a normal node can be located from an anchor node
In the anchor node selection step, each anchor node transmits its own ID and
position information to its neighboring nodes. Each anchor node enables two-
hop flooding by assigning 2 to the Time to Live (TTL) value of its own position
information message. Thus, the normal node can identify either one-hop or two-
hop anchor nodes based on the value of TTL.
Fig. 6 shows an example of two-hop flooding. In Fig. 6(a), anchor nodes A1 ,
A2 , and A3 transmit their position information with TTL=2 to their neighboring
Low-Cost Two-Hop Anchor Node-Based Distributed Range-Free Localization 135
sensor nodes. In Fig. 6(b), every sensor node that has received the position
information with TTL=2 transmits the position information of neighbor anchor
nodes with TTL=1 to its neighboring sensor nodes. Via two-hop flooding, normal
node N1 selects A1 and A2 as one-hop anchor nodes, normal node N2 selects A2
as a one-hop anchor node while choosing A3 as a two-hop anchor node.
(a) (b)
Fig. 6. Two-hop flooding: (a) Flooding with TTL=2, (b) Flooding with TTL=1
4.2 ER Construction
Every normal node can construct an ER with collected position information of
both its one- and two-hop anchor nodes. To reduce the computation overhead
for the localization, the normal node uses the rectangle that circumscribes the
transmission radio range of each anchor node. Thereafter, an ER is constructed
by calculating the overlapped region of these rectangles.
Fig. 7 shows an example of the ER construction. Normal node N keeps the
position information of one-hop anchor node A3 and two-hop anchor nodes A1
and A2 , and constructs an ER (i.e., dark shaded area in the figure) applying
different rectangle drawing mechanisms to one- and two-hop anchor nodes, re-
spectively. Then, the normal node should be located inside the ER.
Fig. 7. ER made by the location information of one- and two-hop anchor nodes
136 T. Kim et al.
Fig. 8. The result of grid scan based on the location information of anchor nodes
Low-Cost Two-Hop Anchor Node-Based Distributed Range-Free Localization 137
in localization broadcast their own position information and that of the one-hop
anchor node to its neighboring nodes.
5 Performance Evaluation
5.1 Methodology
In this section, DRLS, which has come about upon the motivation of LADL,
will be compared with LADL to analyze the performance of LADL. We will use
the following two metrics to measure the performance of LADL.
1. Localization accuracy: It is defined as the difference between the real position
and the measured position of a normal node, and the data transmission radio
range of sensor node r is the basic unit. It shows through a numerical value
how accurately the normal node calculates the position compared with the
real position.
2. Message transmission cost: It is defined by the number of position informa-
tion messages of the anchor nodes used for each normal node to obtain infor-
mation of the anchor nodes. It indicates the position information exchange
cost of the anchor node to measure the positions of the normal nodes.
The simulator in this section is realized using JAVA. The data transmission radio
range of the sensor nodes is r, and the simulation is performed in an environment
where the sensor nodes were deployed randomly in the sensor field with 10rx10r
sizes. A collision on the link, which occurs by two-hop flooding, is not considered,
and when dividing the ER into specific grid sizes, it draws a grid with the size
of each cell being 0.1rx0.1r in consideration of the position estimation accuracy,
computational cost of grid scan, and memory capacity required. Moreover, it
should be compared with the performance of DRLS by using the average value
of the results derived from the 100 simulation repeats.
2.5
1.0
0.5
0.0
0 5 10 15 20 25 30 35 40 45 50
Fig. 9. Localization accuracy of normal nodes with different ratio of anchor nodes
1.2
Mean error (r)
1.0
0.8
0.6
0.4
0 2
0.2
0.0
0 5 10 15 20 25 30 35 40 45 50
Fig. 10. Localization accuracy of all normal nodes with different ratio of anchor nodes
Fig. 11 shows how much the number of total sensor nodes affects the position
estimation accuracy. The result of the accuracy of the position estimation that
was done while changing the number of total anchor nodes to 200, 400, and
600 with fixing the proportion of the anchor nodes at 20% indicates that LADL
has 16-49% better accuracy compared to DRLS. The increase in the number of
total sensor nodes leads to more accurate position estimation because the higher
density of the network causes that the amount of information of the anchor nodes
that a normal node receives increases.
0.8
0.6
0.2
0.0
Fig. 11. Localization accuracy in terms of the total number of sensor nodes
800
DRLS (200 nodes)
700 LADL (200 nodes)
600
The number of messages
500
400
300
200
100
0
0 5 10 15 20 25 30 35 40 45 50
The ratio of anchors (%)
Fig. 12. Anchor node selection cost with different ratio of anchor nodes
of the anchor nodes to the normal nodes. Moreover, as the proportion of the
anchor nodes increases, the number of messages increases significantly due to
the fact that the number of anchor nodes which are the starting point of two-
hop flooding increases. Fig. 9 shows that LADL provides the normal nodes with
information of the anchor nodes with a 5-12% lower message transmission cost.
Fig. 13 shows how much the number of total sensor nodes affects the mes-
sage transmission cost of two-hop flooding. The number of position information
messages is measured while changing the number of total anchor nodes to 200,
400, and 600 with fixing the proportion of the anchor nodes at 20%. The result
indicates that LADL has a 7-14% lower message transmission cost compared to
DRLS. The reason for DRLS’s poorer performance compared to LADL is that
DRLS broadcasts additional position information more than LADL does. Fur-
ther, the two-hop flooding cost increases rapidly as the number of total sensor
nodes increases because the increase in the network density rapidly increases the
number of sensor nodes existing within a one-hop distance from a sensor node.
140 T. Kim et al.
2500
DRLS (anchor ratio = 20%)
LADL (anchor ratio = 20%)
2000
1000
500
0
200 400 600
The number of sensor nodes
Fig. 13. Anchor node selection cost with changing of the total number of sensor nodes
6 Conclusions
References
1. Hightower, J., Borriello, G.: Location Systems for Ubiquitous Computing. Com-
puter 34(8), 57–66 (2001)
2. Pottie, G.J., Kaiser, W.J.: Wireless Integrated Network Sensors. Communications
of the ACM 43(5), 51–58 (2000)
3. Priyantha, N.B., Chakraborty, A., Balakrishnan, H.: The Cricket Location-Support
System. In: Proceedings of ACM MobiCom 2000, pp. 32–43 (2000)
4. Savvides, A., Han, C.-C., Strivastava, M.B.: Dynamic Fine-Grained Localization in
Ad-Hoc Networks of Sensors. In: Proceedings of ACM MobiCom 2001, pp. 166–179
(2001)
5. Bulusu, N., Heidemann, J., Estrin, D.: GPS-Less Low Cost Outdoor Localization
for Very Small Devices. IEEE Wireless Communications 7(5), 28–34 (2000)
6. Doherty, L., Pister, K.S.J., Ghaoui, L.E.: Convex Position Estimation in Wireless
Sensor Networks. In: Proceedings of IEEE INFOCOM 2001, vol. 3, pp. 1655–1663
(2001)
7. Niculescu, D., Nath, B.: Ad Hoc Positioning System (APS). In: Proceedings of
IEEE Globecom 2001, vol. 1, pp. 2926–2931 (2001)
8. Sheu, J.P., Chen, P.C., Hsu, C.S.: A Distributed Localization Scheme for Wireless
Sensor Networks with Improved Grid-Scan and Vector-Based Refinement. IEEE
Transactions on Mobile Computing 7(9), 1110–1123 (2008)
9. Acharya, M., Girao, J., Westhoff, D.: Secure Comparison of Encrypted Data in
Wireless Sensor Networks. In: Proceedings of IEEE Symposium on Modeling and
Optimization in Mobile, Ad Hoc, and Wireless Networks, pp. 47–53 (2005)
10. He, T., Huang, C., Blum, B.M., Stankovic, J.A., Abdelzher, T.: Range-Free Local-
ization Schemes for Large Scale Sensor Networks. In: Proceedings of ACM Mobi-
Com 2003, pp. 81–95 (2003)
11. Boukerche, A., Oliveira, H.A.B.F., Nakamura, E.F., Loureiro, A.A.F.: Localization
Systems for Wireless Sensor Networks. IEEE Wireless Communications 14(6), 6–12
(2007)
12. Akyildiz, I., Su, W., Sankarasubramaniam, Y., Cayirci, E.: A Survey on Sensor
Networks. IEEE Communications Magazine 40(8), 102–114 (2002)
13. Lederer, S., Wang, Y., Gao, J.: Connectivity-based Localization of Large Scale
Sensor Networks with Complex Shape. In: Proceedings of IEEE INFOCOM 2008,
pp. 789–797 (2008)
SecDEACH: Secure and Resilient Dynamic Clustering
Protocol Preserving Data Privacy in WSNs
Abstract. Clustering scheme enabling the efficient utilization of the limited en-
ergy resources of the deployed sensor nodes can effectively prolong lifetime of
wireless sensor networks. Like most of the protocols for wireless sensor net-
works, a clustering scheme is vulnerable to a number of security attacks. Espe-
cially, attacks involving cluster-heads are the most damaging since cluster-based
schemes rely on their cluster-heads for routing and data aggregation. Several
studies have proposed a secure clustering scheme and a resilient cluster leader
election protocol in wireless sensor networks. However, most of resilient clus-
tering schemes do not support dynamic clustering and most of secure clustering
schemes are not robust and resilient against node(specially, cluster-heads) cap-
ture. In this paper, we present a secure and resilient clustering protocol on the
basis of dynamic clustering protocol. Our scheme provides both a secure cluster-
ing and a resilient cluster-head election preserving dynamic clustering unlike the
existing secure clustering schemes. In addition, our scheme provides data privacy
through the concealed data aggregation. The analysis shows that our scheme of-
fers the security services against both an inside attack and an outside attack while
having the little memory, computation and communication overheads. The com-
parison with related works shows that our scheme has a great improvement in
security and the reduction of the overheads compared to the existing schemes.
1 Introduction
D. Taniar et al. (Eds.): ICCSA 2010, Part III, LNCS 6018, pp. 142–157, 2010.
c Springer-Verlag Berlin Heidelberg 2010
SecDEACH: Secure and Resilient Dynamic Clustering Protocol 143
makes them more vulnerable to attacks [3,9]. It is therefore crucial to devise security
solutions to these networks [3].
Sensor nodes are often organized into several clusters for the purpose of efficiency
and scalability. Clustering scheme enabling the efficient utilization of the limited en-
ergy resources of the deployed sensor nodes can effectively prolong lifetime of wireless
sensor networks [7,12]. Adding security to clustering scheme is challenging, as its dy-
namic and periodic rearrangement of clustering makes key distribution solutions inad-
equate [3,9]. Like most of the protocols for WSNs, a clustering scheme is vulnerable to
a number of security attacks including passive eavesdropping, jamming, spoofing and
replay attacks [3]. Especially, attacks involving CHs(cluster-heads) are the most dam-
aging since cluster scheme fundamentally relies on the CHs for data aggregation and
routing. If the adversary manages to become a CH, he can perform various attacks to
disrupt the function of the network such as routing and data aggregation. Therefore, to
guarantee for WSNs to function properly, a secure clustering protocol is required.
A number of secure routing protocols have been proposed recently in WSNs
[1,5,9,10,14]. These can be generally categorized into a secure cluster formation proto-
col [1,9] and a secure CH election protocol [5,14,10]. Secure cluster formation
protocols significantly improve the security of clustering in the presence of malicious
attacks. However, they only focus on the cluster formation and do not consider the se-
curity against disrupting CH election by the compromised node. Resilient cluster leader
election schemes in [5,14,10] are more resilient against a node capture than the above
mentioned secure cluster protocols. However, they do not support dynamic clustering
which is a heuristic method for an optimal clustering since they have assumption that a
cluster sector is pre-defined and not changed. This feature is not suitable for WSNs.
In this paper, we present a secure and resilient clustering scheme considering energy
efficiency and a concealed data aggregation in WSNs. Our scheme provides both the
resilient CH election preserving a dynamic clustering and the secure cluster formation
unlike the existing secure clustering schemes. In addition, our scheme provides data pri-
vacy through a concealed data aggregation. The comparison with related works shows
that our scheme offers the security services against both an inside attack and an outside
attack. Our scheme has a little memory and computation overhead but it is endurable in
sensor nodes.
This paper is organized as follows. Section 2 introduces the backgrounds of the pro-
posed scheme. Section 3 presents our goals and some assumptions required for describ-
ing the proposed scheme. Section 4 describes the procedures of the proposed scheme
in detail. Section 5 explains the results of analysis comparing the our proposed scheme
with the other schemes. At last, section 6 summarizes our work and discusses the future
works.
2 Backgrounds
2.1 DEACH and Its Vulnerabilities
DEACH. The basic operation of a clustering routing protocol in WSNs is to partition
network into several clusters. After clustering, the CH of each cluster aggregates the
sensed data from its member nodes for energy efficiency of data processing and then
144 Y.-J. Han, M.-W. Park, and T.-M. Chung
sends the aggregated data to the BS(base station) [11,15]. The most common technique
in famous clustering routing protocols is the probabilistic clustering algorithm based
on the randomized CH rotation for distributing the energy consumption among sensor
nodes in each cluster [11,15].
As our previous works, we have proposed DEACH(Distance based low-Energy
Adaptive Clustering Hierarchy Protocol) [16] which is an energy-efficient distance
based clustering scheme. In clustering scheme, how to control the balance of the energy
consumption among CHs can greatly affect the network lifespan which is generally de-
termined by the first dead-node [12]. Among the sources of energy consumption of a
CH, wireless data transmission is the most critical [6]. Because the energy consump-
tion of wireless data transmission increases generally in proportion with the distance
from sending node to receiving node [6], the energy consumption of CH depends on
its location from the BS. For example, in single-hop inter-cluster communication, the
CHs farther to the BS will die much faster than the other CHs, while in multi-hop
inter-cluster communication, the CH closest to the BS is burdened with a heavy relay
traffic load and die first(i.e., the hot spot problem). Most of clustering schemes such as
LEACH [15] and HEED [11] utilize mainly CH frequency or residual energy of each
node as criterion of CH election and do not consider the distance from the BS to the
CH. So, let those schemes cause the unbalanced energy consumption among CHs in
spite those schemes are able to prolong the network lifetime.
To solve this problem, DEACH considers the distance from the BS to CHs as well as
the residual energy as the criterion of CH election for the balanced energy consumption
among CHs. For a single-hop inter-cluster environment, we use Eq. 1 as probability
formula for electing a CH. That is, each sensor node has the probability, Pi , of becoming
a CH which is determined by the distance to the BS and its residual energy.
d(Si ,BS)−dmin
Popt (i) = 32 Popt × dmax −dmin
⎧
⎪
⎪ Popt (i) Eres (i)
⎨ × Ci (t) = 1 (1)
Pi = 1 − Popt (i) × (r mod
1
) Einit (i)
⎪
⎪
Popt (i)
⎩
0 Ci (t) = 0
where Popt is the optimal probability being a CH which is a system parameter, Popt (i)
is the optimal probability being a CH of sensor node i based on the distance from itself
to the BS, r is the current round of the CH election, Einit (i) and Eres (i) are the initial
energy and the residual energy of the sensor node, d(i, j) is the distance between node
i and j, dmax represents the distance of the farthest sensor node from the BS and dmin
represents the distance of the closest sensor node. Ci (t) presents whether or not sensor
node i has been a CH in the most recent r mod 1/Popt rounds. That is, Ci (t) = 0 if
node i has been a CH and one otherwise). Equation 1 shows that the closer the sensor
node is to the BS, the lower probability of becoming a CH it could have. In other words,
the CHs closer to the BS can have more members than the CHs farther to the BS since
the few CHs are elected in the area close to BS. Besides, among the nodes having the
same distance from the BS, the more the sensor node has the residual energy, the higher
probability of becoming a CH it could have. According to Eq. 1, every sensor node
SecDEACH: Secure and Resilient Dynamic Clustering Protocol 145
elects itself as a CH only once during 1/Popt rounds on average. In this way, the CHs
are selected randomly among the sensor nodes and the energy concentration on CHs is
distributed.
DEACH provides fully distributed manner by utilizing local information and good
energy-efficiency by performing the load balanced clustering. Through simulation ex-
periments, we have showed that the proposed scheme is more effective than the other
clustering protocols in prolonging the lifespan of WSNs [16].
Its Vulnerabilities. Adding security to clustering scheme is challenging, as its dynamic
and periodic rearrangement of clustering makes key distribution solutions inadequate
[3,9]. Like most of the protocols for WSNs, a clustering scheme is vulnerable to a
number of security attacks including passive eavesdropping, jamming, spoofing and
replay attacks [3]. Especially, attacks involving CHs are the most damaging since a
cluster scheme fully relies on the CHs for data aggregation and routing. If the adversary
manages to become a CH, he can perform various attacks to disrupt the function of the
network such as routing and data aggregation.
Firstly, in existing CH election protocols, nodes rely on their local status or on re-
ported status of peer nodes for electing their CH. For example, in LEACH [15], HEED
[11] and DEACH [16], a node can independently decide to become CH based on its
energy-level or a distance. However, such approaches have substantial limitations with
the respect to security. A fundamental observation is that any election mechanism based
on a concrete election metrics such as the residual energy level or distance, can be ma-
nipulated or forged. So, the adversary can become the CH or disturb a legitimate node
not to become a CH for the upcoming epoch by forging or replaying CH announcement.
Secondly, if the CH is compromised by the adversary, the adversary can learn the
sensed data and forge the data because the adversary knows key material even though
security primitive such as encryption is used to provide confidentiality of the sensed
data. This feature makes the application in WSNs very more critical and vulnerable.
with a compromised CH is not to guarantee confidentiality of the sensed data from sen-
sor nodes. Even though each sensor node encrypts its sensed data for confidentiality,
the compromised CHs can decrypt the data and finally they know which information
is sensed from sensor nodes. So, a concealed data aggregation is required against a
node capture. Finally, they have a single point of failure problem because they are the
centralized mechanism which fully depends on the BS for authenticating the CH.
A number of secure cluster leader election protocols have been proposed in [5,10,14].
All of them work in a decentralized way and use the light-weighted cryptographic al-
gorithms. The main technique of their schemes is that cluster member securely elects
its CH through the agreement of a elected CH within a pre-defined cluster. [5] and [10]
use a commitment based scheme. Each sensor sends its commitment to other sensors in
the peer-to-peer manner. A commitment is an encrypted random value using a shared
key and the random value is created by each sensor. Then, each sensor sends the ful-
fillment value(that is, its random value) to other sensors within its sector. The sensors
receiving the fulfillment values verify them using the shared key and sum them to make
an agreed random value. They divide the real sum of agreed random values by the num-
ber of sensors and get the remainder which indicates the position of CH node in the
cluster. Like [5,10], Dong’s scheme [14] also has the assumption that cluster sector is
pre-defined and not changed. In Dong’s scheme, the new CH is selected based on the
remaining energy on cluster members for the purpose of the load balancing. Their idea
for a resilient clustering is to let all member within a cluster have the identical list of
the candidate CHs. To control the list of each sensor node, they use two one-way hash
chains, YES chain and NO chain. In initialization phase, every sensor node sends its
YES commitment value to the other nodes within its sector. Each sensor node receiving
all YES commitment from the other nodes creates the list of the candidate CHs through
the shuffle algorithm. The shuffle algorithm makes the list of the candidate CHs of all
sensor nodes within the same sector identical. In each round of CH election, each sen-
sor node sends YES key or NO key according to its residual energy. If sensor node
does not have energy for serving as the CH, it releases NO key, otherwise it releases
YES key. If sensor node receiving NO key, it removes identifier(ID) of the sending node
from the list of the candidate CHs. Through this mechanism, all cluster members in the
list of the candidate CHs will server as the CH in a round-robin manner unless they
release their NO keys during the CH election. When the benign cluster members are
well-connected, all benign nodes that are eligible for the CH role will be included in
the list of the candidate CHs.
Resilient cluster leader election schemes are more resilient against a node capture
than the above mentioned secure clustering protocol. However, they do not support a
dynamic clustering which is a heuristic method for an optimal clustering since they
have the assumption that cluster sector is pre-defined and not changed. Figure 1 shows
the problem of a pre-defined clustering. In Fig. 1, WSNs are divided into 4 sectors.
Nodes A,B,C and D simultaneously can become the CHs in any round. In this case,
even though they are very close to each other, they are not clustered together because
they belong to different sectors. As a result, intra-cluster communications of them cause
the heavy overhead. This feature is not suitable for WSNs.
We assume an adversary may launch arbitrary attacks against the cluster formation
protocol and cluster based routing protocol except for completely jamming the commu-
nication channel. In this section, we classify adversaries in clustering routing protocol
as follows: external adversary and internal adversary.
– External adversary. An external adversary can randomly select some nodes and
extract key generation information from them. Then, he tries to eavesdrop the com-
munication between other ordinary nodes. Besides, he tries to inject and replay
packet to disrupt the cluster formation and cluster routing.
– Internal adversary. We assume that he already has knowledge on the algorithms or
protocols. Using this knowledge, he can selectively capture the nodes and then he
tries to eavesdrop the communication between other ordinary nodes. Then, he can
use any method to interrupt/corrupt cluster formation and to forge the sensed data or
the aggregated data, e.g., compromising some nodes, modifying the communication
message, replay attack etc.
Ensuring a good level of security for such types of networks is not a trivial task. As
WSNs use wireless communications, the threats and attacks against them are more di-
verse and often large-scale.
We denote some common notations for describing the our scheme as follows.
• Si is ith (i ∈ N ) sensor node.
• Si (id) is the identifier of the sensor node, Si .
• Si (r) is the maximum transmission range of the sensor node, Si .
• Pi is the probability of becoming a CH of the sensor node, Si .
• Popt (i) is the optimal probability of becoming a CH of the node Si referred in Eq. 1.
• CHi is ith cluster-head.
• SCH (i) is the set of the candidate CHs of the sensor node, Si .
• Eres (i) is the residual energy of the sensor node, Si .
• Eth is the minimum energy required for serving as CH.
• Sneighbor (i) is the set of neighborhood nodes of the node, Si , within Si (r).
• Qi (j) is the qualification information of the node, Sj , as a CH of the node, Si .
• Sq (i) is the set of the qualified nodes of the node, Si .
• di is the sensed data from the sensor node, Si .
• ermax is the maximum number of rounds for CH election.
The objective of SecDEACH is to provide for both a secure clustering protocol con-
sidering a concealed data aggregation and a resilient CH election against an internal
attacker and an external attacker. To prevent from security threats, WSNs must satisfy
the security requirements and design requirement as following:
Security Goal 1: Resilience against fabrication attacks, impersonation attacks and
replay attacks. To prevent an external attacker, a clustering protocol ensures that unau-
thorized sensor nodes should not join a cluster routing protocol. For doing this, it should
guarantee the basic security primitives such as authentication, confidentiality, freshness
and integrity.
Security Goal 2: Fault-tolerance. It is impossible to detect a compromised node act-
ing like a ordinary node. So, if a compromised node acts as a CH, the application is
damaged since a CH is responsible for data aggregation and data transmission schedule
of its member nodes. To prevent this attack, a cluster routing scheme guarantees that a
compromised node does not become a CH [14]. That is, the attacker cannot arbitrarily
increase or decrease the chance of any ordinary node being elected as a CH.
Security Goal 3: Concealed data aggregation in-networking. If a CH is compro-
mised, the compromised CH node can learn all of the sensed data from its member
nodes by decrypting them, even though its member nodes encrypt their sensed data. To
SecDEACH: Secure and Resilient Dynamic Clustering Protocol 149
prevent this attack, a clustering routing protocol guarantees confidentiality and privacy
of the sensed data which does not depend on CH.
Design Goal: Low communication, computation overheads and low storage re-
quirement. Adding to the security primitives in a clustering protocol makes commu-
nication and computation overheads larger. The stronger robustness of security is, the
overhead is larger. So, we consider the energy-efficient mechanism preserving the ro-
bustness of security. In addition, we consider the limitation of storage resulted from the
resource constraints of the sensor node.
4 SecDEACH
In this section, we will briefly review the security primitives used in the proposed
scheme and then describe its technical detail.
(a1 + a2 +, ..., +an ) = Deck (Enck1 (a1 ) + Enck2 (a2 ) + ... + Enckn (an )) (2)
(a1 + a2 +, ..., +an ) = Deck1 +k2 +...+kn (Enck1 (a1 ) + Enck2 (a2 ) + ... + Enckn (an ))
(3)
In our proposed scheme, we will use this for concealed data aggregation.
Initialization. Firstly, to use μTESLA [2] for the BS’s authenticated broadcast, the BS
creates BROADCAST key chain by using one-way hash key chain method in subsec-
tion 4.1, {K 0 , K 1 , ...., K brmax }, where brmax is the maximum number of broadcasts
from the BS. The BS creates a pairwise symmetric keys, {K1 , K2 , ..., KN }, for all
sensor nodes to protect the node-to-BS communication. The BS then maintains a sym-
metric bivariate polynomial f (·, ·) which is a key material for establishing a pairwise
key between sensor nodes after deploying and pre-distributes a polynomial share f (i, ·)
to every node Si . We will describe later how to generate sensor node id, Si (id).
Next, prior to deployment, each sensor node Si is assigned with node ID, Si (id),
two random one-way key chains, the A − key chain1 and the G − key chain2 , a
pairwise key shared with the BS, Ki , and K 0 which is shared by all members of the
network for BS’s authenticated broadcast. The G-key chain is used by this node to tell
other nodes that it has no energy left for serving as CH. Therefore, it only includes
a key chain commitment, Gi,0 , and G-key, Gi,1 . The A-key chain is used to inform
other nodes that it is a self-elected CH. Therefore, it includes a key chain commitment,
Ai,0 , and a number of A-keys ,{Ai,1 , Ai,2 , Ai,3 , ...., Ai,δ∗Popt ∗ermax }. We will discuss
in section 5.4 why we determine the number of A-keys as δ ∗ Popt ∗ ermax . Lastly, we
define that the ID of sensor node is the hash of the two key chain commitments, i.e.
Si (id) = H(Ai,0 Gi,0 ) like in [14].
1
ANNOUNCEMENT key chain.
2
GIVEUP key chain.
SecDEACH: Secure and Resilient Dynamic Clustering Protocol 151
After deployment, the sensor node Si firstly establishes a pairwise key with neigh-
borhood sensor nodes within their communication range Si (r) by using Blundo’s pair-
wise key pre-distribution scheme and initializes their neighborhood sensor nodes list,
Sneighbor (i). A pairwise key between two sensor nodes Si and Sj is denoted as Ki,j .
Next, the every sensor node calculates its Popt (i) for the CH election. Finally, the every
sensor node needs to exchange their key chain commitments (A-key commitment, Ai,0
and G-key commitment, Gi,0 ) between each other within its transmission range. For
doing so, each sensor node sends initM sg(Si ) = {Si (id)Eki,j (Ai,0 Gi,0 Popt (i))},
within its transmission range. The sensor node receiving this message can easily verify
these commitments according to the node ID. Whenever a sensor node Si receives the
verified A-key commitment from a node Sj , Si creates Qi (j). Qi (j) means that Sj
has a qualification being a CH of Si . It consists of node ID, 1/Popt (j) rounds, the last
round that the node Sj was a CH, the number of announcements within last 1/Popt (j)
rounds and last A-key. After collecting all those authenticated A-chain commitments
from other nodes, every sensor node Si creates a qualification list, Sq (i). Sq (i) consists
of array of Qi (j). This is meaningful because a dynamic clustering formation is only
performed within its communication range.
Secure Cluster Head Election. We use DEACH in section 2.1 as a fundamental clus-
tering method. DEACH consists of a cluster-setup phase and a data transmission phase.
The secure CH election and secure cluster formation phases belong to the cluster-setup
phase. For simplicity, we assume that no collision occurred in the MAC(Media Access
Control) layers of sensors during the clustering routing.
At the beginning of this phase, if the residual energy Eres (i) of the sensor node Si
is less than the threshold Eth , the sensor node Si broadcasts the key of its G-key chain
to announce its decision. Eth is the minimum energy required for serving as CH and
pre-determined system value. If Eres (i) is greater than Eth ,this node chooses a random
number between 0 and 1. If the random number is less than its Pi in Eq. 1, this node
becomes the CH and releases its A-key from A-key chain and broadcasts its announce-
ment message, advHeadM sg(CHi ) = {Si (id)Ci,c }, β times to tolerate the channel
loss within transmission range. c is the number of the cumulative announcements sent
by Si . If a sensor node Sj receives a key in the A-key chain of Si , node Sj can verify
it by computing Ai,c−1 = H(Ai,c ) and comparing with the key Ai,c−1 received ear-
lier. If the verification succeeds, node Sj extracts Qj (i) from its Sq (j) . And then Sj
determines whether Si has a qualification being a CH by using Eq. 4.
1 , if (r − r ) ≥ r mod (1/Popt (i))
V (i) = { (4)
0 , otherwise
where r is the current round, r is the last round in Qj (i). If V (i) is one, Sj adds Si into
SCH (i) since it means that Si is a non-compromised node. Otherwise, this announce-
ment is ignored. If this announcement continues over α times within 1/Popt (i), Si is
removed permanently from Sq (j). If node Sj receives the correct G-key Gi,1 from Si ,
Sj will remove Si from Sq (j). Finally, after SCH (j) is created, Sj chooses the closest
CHi among the CHs in SCH (j) and joins to the CHi as the member. After joining, it
updates the A-key and the last round in Qj (i). Table 1 shows the CH election process
in a sensor node having 5-node neighbors within its transmission range when α = 3.
152 Y.-J. Han, M.-W. Park, and T.-M. Chung
Table 1. Illustration of the resilient CH election in a sensor node having 5-node neighbors(α = 3)
Secure Cluster Formation. After electing the CHj , Si sends the joinM sg to the
CHj . The joinM sg(Si , CHj ) is as follows.
joinM sg(Si , CHj ) = {Si (id)Sj (id)nonceM ACki,j (Si (id)Sj (id)nonce)}
The nonce is used to prevent the replay attack. After receiving joinM sg, the CHs
send the time slot schedule to their cluster member nodes by using a pairwise key.
Secure Data Aggregation. In this step, Si encrypts the sensed data using a shared key
Ki with the BS for the concealed data aggregation of its CH and sends reportM sg.
The reportM sg(Si , CHj ) is as follows.
reportM sg(Si , CHj ) = {Si (id)Sj (id)Encki (di )nonce + l
M ACki,j (Si (id)Sj (id)Encki (di )nonce + l)}
The node-to-CH communication is protected by using the same key used to pro-
tect the joinM sg and guarantees integrity through MAC3 . A value computed from the
nonce and the reporting cycle(l) [9] is also included to prevent replay. The CHs can
now check the authenticity of sensing
reports they receive and then perform a con-
cealed data aggregation dsum = Encki (di ) for all encrypted sensing data using
symmetric double homomorphic encryption scheme in Eq. 3. That is, not only the CHs
do not need to decrypt the encrypted sensing data but also can not the CH decrypt the
encrypted data since the encryption is performed by using the shared key with the BS.
After performing the concealed data aggregation, the CH sends reportM sg to the BS.
The reportM sg(CHi , BS) is as follows where Lmembers is the IDs list of reporting
member nodes of CHi . It is required for the BS to decrypt the sum of encrypted data
by the CHi .
reportM sg(CHi , BS) = {Si (id)Lmembers (i)dsum counter
M ACki (Si (id)Lmembers (i)dsum counter)}
3
Message authentication code as keyed hash function.
SecDEACH: Secure and Resilient Dynamic Clustering Protocol 153
The aggregate result is protected by using the symmetric key shared between the CH
and the BS. For freshness, a counter (shared between the CH and the BS) is included
in the reporting message. The BS receiving the aggregated data can decrypt the data
using shared keys between sensor nodes using Eq. 2.
5 Analysis
5.1 Security Goal 1: Resilience against Fabrication Attacks, Impersonation
Attacks and Replay Attacks
This security goal requires authenticity, integrity, confidentiality and freshness to node-
to-node communications.
For authenticity, it has to guarantee that only legitimate cluster can participate in
cluster routing operation. In other words, we have to ensure that the unauthorized nodes
do not have to join the CH election and cannot impersonate any ordinary sensor node.
These two properties are actually achieved by using the Blundo’s key pre-distribution
protocol [4]. The ID of a sensor node is tied to its cryptographic key. It is thus not
possible for an unauthorized node to join the cluster routing procedure. That is, the
messages, initM sg, joinM sg and reportM sg, are encrypted with a pairwise key and
a successful decryption of this message allows the sensor node to conclude that the
message is originated from a legitimate node in the network. In addition, even though
advHeadM sg is not encrypted, the adversary cannot forge the announcements from a
legitimate node. The announcement contains the ID of the sensor node and the key in
A-key or G-key chains. To forge these keys, the attacker needs to either invert the one-
way hash function H or fool benign cluster members into accepting incorrect key chain
commitment. However, because of our ID assignment scheme, the ID of a sensor node,
Si (id), is the hash of its two key chain commitments, i.e., Si (id) = H(Ai,0 Gi,0 ).
Therefore, to forge the key chain commitment, the adversary has to forge the ID as
well. This requires the knowledge of keying materials related to the forged ID, which is
not available to the adversary. Therefore, the adversary has no way to make its forged
commitments accepted by any benign cluster member. Without the proper chained keys,
the adversary may try to replay the keys released earlier by the legitimate member.
However, this will not impact our scheme because our scheme uses either the fresh
A-key or the G-key in each round of CH election.
Because the messages, joinM sg and reportM sg, are encrypted by MAC using a
pairwise key and includes the nonce, the CH can also conclude that it is not changed and
it is not a stale message being replayed. The freshness of all subsequent sensor reports
from the ordinary nodes to their CH is guaranteed by nonce values that are incremented
each time. For CH-to-BS communication, the freshness is guaranteed by the counter
value shared between the CH and the BS. The counter value is also incremented each
time the CH sends a new report to the BS.
to identify those passive malicious node since there are no evidences of them being
malicious. As a result, this goal focuses on making sure that an adversary cannot sig-
nificantly impact the chance of any benign members elected as a new CH. Thus even
if the CH is compromised, it cannot continue to control properly the cluster. The CH
election protocol should ensure the fairness of the sensor nodes being elected as CHs.
For doing so, our idea is to ensure that each sensor node becomes a CH one per 1/Popt
rounds on average.
100
F−LEACH, GS−LEACH and
The number of the compromised CHs
SecLeach
80 Dong’ Scheme and
The proposed Scheme
60
40
20
0
0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1
The probability being compromised, f
In our scheme, the CH election is based on the neighbors list Sneighbor (i), which
is initialized at the beginning of the CH election. All nodes in the neighbors list has
the qualification serving as the CH. In our scheme, the sensor node can receive sev-
eral announcements being a CH from the neighborhood in each round of CH election.
The sensor node receiving them has to verify whether they have a qualification serving
as the CH and chooses its CH among the verified candidate CH nodes. To verify the
qualification, our scheme uses Eq. 4. If the announcement is advertised more than α
times within 1/Popt , the verification is failed. If the verification is failed for a node, the
node cannot become a CH. It ensures that each sensor node becomes a CH one per its
1/Popt rounds on average. Figure 2 shows the comparison our scheme with the exist-
ing secure cluster routing schemes [1,9] in the aspect of the compromised CH election.
Figure 2 shows the number of the compromised node being elected CH according to
the probability being compromised of the sensor node, f , where EX(1/Popt ) = 5 and
N = 100. The number of compromised CH in SLEACH, F-LEACH and SecLEACH
is very larger than that of our scheme and Dong’s scheme because the secure clustering
schemes like F-LEACH cannot know if the a self-selected CH is compromised. As we
can see in Fig. 2, our scheme provides resilient CH election like Dong’s scheme [14].
sensed data with a shared key with only BS. After receiving all reportM sg from all
member node in its cluster, the CH just only sum all the encrypted data and then sends
the summed data to the BS. The BS can extract the sensed data from the summed en-
crypted data since the BS knows all of the shared keys with sensor nodes. It offers a
concealed data aggregation as well as data privacy.
5.4 Overhead
Before deployment, every sensor node is assigned with an ID and a t-degree polynomial
share f (i, ·) . It also needs to store two key chains along with their commitments, one
contains a single key, and the other consists of δ ∗ Popt ∗ ermax keys. These chains
are similarly used in Dong’s scheme [14]. However,in Dong’s scheme, if a node has
capability being a CH, sensor node must inform its YES key per every cluster leader
election round since a YES chain is used to announce being a candidate CH. So, the
number of keys of a YES chain in Dong’s scheme is the same as the maximum number
of rounds for CH election, ermax . On the other hands, in our scheme, the sensor node
being a self-elected CH only broadcasts its A-key to announce the fact that it becomes
a CH. Without loss of generality, the ordinary senor node is elected itself as CH one
within last 1/popt rounds on average because the sensor node elects itself as the CH by
Eq. 1 which is optimized into Popt [16]. Thus, in our scheme, each node Si only needs
Popt (i) ∗ ermax keys for A-key chain unlike Dong’s scheme. So, the number of our A-
key chain is very smaller than that of Dong’s scheme. However, there is one limitation
to consider. To calculate Popt , each node has to know its location information which
will be deployed. However, we cannot calculate Popt (i) in the initialization phase since
each sensor node will be deployed randomly. So, we determine the length of A-key
chain as δ ∗ Popt ∗ ermax based on Popt . According to Eq. 1, Popt (i) is less than or
equal to 32 Popt . As a result, the adequate value of δ can be 2.
Figure 3 shows the comparison our scheme with Dong’s scheme in the aspects of
the memory and communication overheads. Figure 3(a) and Fig. 3(b) show the mem-
ory overhead for saving A-key chain and the communication overhead for sending
announcement by the number of CH where key size = 8bytes, ermax = 50 and
N = 100. As we can see in Fig. 3, our scheme is more light-weighted and energy-
efficient than Dong’s scheme. In addition, if there are n sensor nodes within a transmis-
sion range, a sensor node will also need to buffer the neighbors list of n IDs, (n − 1)
pairwise keys, and 2(n − 1) key chain commitments. These space requirements are
usually not a problem since the transmission range is limited.
The computational overhead mainly comes from three parts. First, with the Blundo’s
key pre-distribution protocol, a sensor node needs to compute a t-degree polynomial
to establish a pairwise key and verify the ID of another sensor node. Second, a sensor
node has to perform the hash function H one time to verify the key chain commit-
ment. Finally, every sensor node will need to perform symmetric key operations such
as encryption and authentication to protect the communication between sensor nodes.
Obviously, all these operations only involve light-weighted computation.
The proposed scheme only causes small communication overhead like Dong’s scheme
[14]. The Blundo’s key pre-distribution technique does not introduce any additional
communication overhead at all for two nodes to establish a pairwise key [14]. In the
156 Y.-J. Han, M.-W. Park, and T.-M. Chung
memory overhead for saving A−key chain
450 0.03
communication overhead
350 The proposed Scheme, δ=2
300 The proposed Scheme, δ=3 0.02
The proposed Scheme, δ=4
250
0.015
200
0.01
150
100 0.005
50
0
10 12 14 16 18 20 10 12 14 16 18 20
The number of CH The number of CH
initialization step, each node only needs to disclose its ID and two key chain commit-
ments. In each round of CH election, every sensor node is only required to broadcast
a single chained key, A-key. Different from centralized approaches [1,9] where sen-
sor nodes exchange messages with the BS that is far away, our scheme only involves
the communication between the sensor nodes that are physically close to each other.
Therefore, the communication cost of our protocol is not a big problem [14].
References
1. Ferreira, A.C., Vilaca, M.A., et al.: On the security of cluster-based communication protocols
for wireless sensor networks. In: Lorenz, P., Dini, P. (eds.) ICN 2005. LNCS, vol. 3420, pp.
449–458. Springer, Heidelberg (2005)
2. Perring, A., Szewczyk, R., Well, V., Culler, D., Tygar, J.D.: SPINS: Security Protocol for
Sensor Networks. Wireless Networks 8(5), 521–534 (2002)
3. Karlof, C., Wagner, D., et al.: Secure routing in wireless sensor networks: attacks and coun-
termeasures. Elsevier’s AdHoc Networks J. 1(2-3), 293–315 (2003)
SecDEACH: Secure and Resilient Dynamic Clustering Protocol 157
4. Liu, D., Ning, P.: Establishing pairwise keys in distributed sensor networks. In: CCS 2003,
Proceeding of the 10th ACM conference on Computer and communications security, pp.
52–61. ACM Press, New York (2003), Network
5. Wang, G., Cho, G.: Secure Cluster Head Sensor Elections using Signal Strength Estima-
tion and Ordered Transmissions, Sensors. MDPI (Molecular Diversity Preservation Interna-
tional) 9(2), 4709–4727 (2009)
6. Karl, H., Willig, A.: Protocols and Architectures for Wireless sensor Networks. John Wiley
& Suns, Ltd., Chichester (2005)
7. Al-Karaki, J.N., Kamal, A.E.: Routing techniques in wireless sensor networks: a survey.
IEEE Wireless Communications 11(6), 6–28 (2004)
8. Akyildiz, L.F., et al.: A Survey on Sensor Networks. IEEE Communications Magazine 40(8),
102–114 (2002)
9. Oliveira, L.B., Ferreira, A., et al.: Secleach-on the security of clustered sensor networks.
Signal Process 87(12), 2882–2895 (2007)
10. Sirivianos, M., et al.: Non-manipulable aggregator node election for wireless sensor net-
works. In: ICST wiOpt (January 2007)
11. Younis, O., Fahmy, S.: HEED: A Hybrid, Energy-Efficient, Distributed Clustering Approach
for Ad Hoc Sensor Networks. IEEE Transactions on Mobile Computing 3(4), 660–669
(2004)
12. Younis, O., et al.: Node Clustering in Wireless Snsor Networks: Recent Developments and
Deployment Challenges. IEEE Network 20(3), 20–25 (2006)
13. Tiwari, P.: Data Aggregation in Cluster-based Wireless Sensor Networks, A master thesis,
Deemed University (July 2008)
14. Dong, Q., Liu, D.: Resilient Cluster Leader Election for Wireless Sensor Networks. In: Pro-
ceedings of The IEEE Communications Society Conference on Sensor, Mesh and Ad Hoc
Communications and Networks (SECON), June 2009, pp. 1–9 (2009)
15. heinzelman, W., et al.: An application-Specific Protocol Architecture for Wireless Microsen-
sor Networks. IEEE Transactions on Wireless Communications 1(4) (October 2002)
16. Han, Y.-J., Park, S.-H., Eom, J.-H., Chung, T.-M.: Energy-Efficient Distance based Cluster-
ing Routing Scheme for Wireless Sensor Networks. In: Gervasi, O., Gavrilova, M.L. (eds.)
ICCSA 2007, Part II. LNCS, vol. 4706, pp. 195–206. Springer, Heidelberg (2007)
Avoidance of Co-channel Interference
Using Switched Parasitic Array Antenna
in Femtocell Networks
1 Introduction
According to recent surveys, 50 percent of phone calls and 70 percent of data
services have taken place indoors [1]. Therefore, network operators need to pro-
vide high data rate and quality of service (QoS) indoors. Femtocell has been
introduced to solve indoor coverage problem and QoS requirement. One defini-
tion of femtocell is a low-power wireless access point that operates in licensed
spectrum to connect standard mobile devices to the operators network using
a digital subscriber line (DSL) connection [2]. Femtocell provides benefits for
both subscribers and cellular poerators. Subscribers can achieve high data rate
indoors with low-cost with deployment of femtocell. Wireless operator can ex-
tend the indoor coverage with low-cost, increase cell capacity, and gain the high
frequency reuse efficiency. Moreover, cullular operators can reduce traffic for
macrocell, thus reducing infrastructure and operating cost.
Corresponding author.
D. Taniar et al. (Eds.): ICCSA 2010, Part III, LNCS 6018, pp. 158–167, 2010.
c Springer-Verlag Berlin Heidelberg 2010
Avoidance of Co-channel Interference 159
2 Related Work
If the macrocell and femtocell use the same frequency band, it causes co-channel
interference between them. Figure 1 shows that the co-channel interference sce-
nario between the macrocell and femtocell in downlink. The macrocell users are
160 Y. Jeong et al.
Femto UE
Femto BS
Macro BS
Macro UE
interfered from FBS, and the femtocell users are interfered from MBS as shown
in Figure 1. When a macrocell user approaches the femtocell coverage, the effect
of interference from FBS may grow up, and when the femtocell user approaches
MBS, it may suffer heavy interference from MBS. Such cases, the performance
of macrocell and femtocell are decreased. If the macrocell user gets into the
femtocell coverage, communication with MBS becomes impossible, because the
interference to macrocell user in femtocell coverage is much bigger than received
signal strength from MBS.
In the previous study, FBS uses omni-directional antenna to transmit the sig-
nal to all direction irrespective of its user’s position [7, 8]. Therefore, macrocell
users are interfered from FBS when they are located nearby the femtocell cover-
age. Additionally, co-channel interference is getting bigger as transmitting power
of FBS increases. If FBS makes beam pattern to its user’s location, the signal
level of FBS is very low in the rest of its user’s location. If the macrocell user is
located outside of femtocell coverage, the co-channel interference from FBS can
be avoided based on macrocell user’s location.
In [9], a multi-element antenna solution is proposed to decrease the interfer-
ence in macrocell users. It can control the antenna beam direction by switching
of two patch antennas and IFA antennas. The effect of interference in macrocell
users is reduced by controlling the transmission power Pf emto of FBS according
to the equation given below:
Pf emto = min (Pmacro + G (θ) − Lmacro (d) + Lf emto (r) , Pmax ) (1)
where Pmacro is the transmitting power of MBS, G(θ) is the antenna gain of
MBS, L( macro)(d) is average outdoor path loss when the distance between
macrocell and femtocell is d. Lf emto (r) is average indoor path loss when the
distance between FBS and its user is r. Pmax is the maximum transmitting
Avoidance of Co-channel Interference 161
power of FBS. The method of (1) means that FSB increases the transmitting
power to overcome the interference by macrocell when it is closely located to
MSB and conversely the FBS far away from MBS lowers its power.
Figure 2 shows omni-directional antenna and SPA antenna. In this paper,
FBS uses the SPA antenna in Figure 2 (b). A monopole antenna located at the
center of the ground plane, and four identical parasitic array elements are placed
surrounding the monopole symmetrically. Switches are inserted between array
elements and ground plane to control the main beam direction of the monopole
antenna by alternating the ground connection of the parasitic elements. A bias
applied to the knee in the element provides an RF shorting path to the ground
plane, and the shorted parasitic elements acts as a “director” to guide the main
beam direction to its direction. A bias applied to the end of element provides a
long RF path to the ground plane when the switch at the knee is turned-off. In
this case, the parasitic element works as a “reflector.”
If two switches in the parasitic element are turned-off, the element is “in-
visible.” Four near orthogonal beam patterns can be generated by sequentially
switching two element pair placed in a diagonal position as front director and
back reflector. In addition, eight directional patterns can be also obtained by
utilizing two adjacent reflectors and two directors. All directional beam patterns
are distinctly identified from each other according to their direction with the
same shape and easily synthesized by controlling the switches. Even radiation
pattern and reasonable front-back ratio for each direction makes it possible to
simplify the power control algorithm to reduce the interference.
3 Proposed Scheme
In this section, we explain the procedure of data transmission to its user in FBS
using SPA antenna. We assume that FBS are configured to limit access to only
a few authorized users to protect limited resources. If unauthorized users re-
quest to connect to the closest femtocell, FBS restricts connection to femtocell
for them. In this condition, our proposed scheme is as follows. The FBS mon-
itors the user location by measuring the signal strength of surrounding users.
Then, FBS performs 1-directional power control of transmitting signal with an-
tenna gain based on user’s position. Then, check the users’ ID to avoid unwanted
162 Y. Jeong et al.
data transmission to unauthorized users. Finally, FBS make the beam pattern
to transfer the signal to authorized user. These processes guarantee the same
performance of femtocell with FBS using omni-directional antenna. Addition-
ally, it reduces the co-channel interference to macrocell users nearby femtocell
coverage. The procedure of data transmission of FBS is shown in Figure 3.
8-direction Sensing
(Receive Mode)
1-direction
Power Control N
A subscriber
Certification
Y
Data Transmission
First, FBS measures the signal level of nearby users. Second, FBS confirms
the user’s location based on received signal strength. Third, according to the
measured distance and location, FBS controls the transmitting power as below:
is located. If the angle of the location between macrocell user and femtocell user
is quite different, macrocell user can avoid the co-channel interference from the
FBS. Otherwise, macrocell user may get interfered from FBS as the difference
of angle with femtocell user’s location.
Second, by using SPA antenna in FBS, we can obtain higher antenna gain
than FBS using omni-directional antenna. Thus, FBS using SPA antenna has
the same performance with omni-directional antenna while transmitting power
of FBS is lower than that of omni-directional antenna. Additionally it can avoid
unnecessary data transmission due to transmit the signal only for registered
users. When FBS uses SPA antenna, power consumption can be lower than that
of omni-directional antenna.
Parameters Values
System bandwidth 10MHz
Inter-macrocell distance 1000m
Inter-femtocell distance 40m
Radius of femtocell 10m
Macrocell BS power 20W
Femtocell BS power 20mW
Number of subcarriers 768
Frame duration 5ms
Number of OFDM symbols 48
Thermal noise(AWGN) density -174dBm/Hz
System frequency(λ) 2GHz
4 Performance Evaluation
In this paper, we performe the system level simulation in three cases: conven-
tional OFDMA macrocell with no femtocells; macrocell with femtocells using
omni-directional antenna; and macrocell with femtocells using SPA antenna. We
consider the reuse factor 1 for macrocell as 2-tier network that has 19 cells. We
assume that MBS is located at the center (0m) of the macrocell. Then, macrocell
is divided into two regions according to the distance from MBS. 0∼409m is inner
region, and 409∼577m is outer region. There are 30 users in each macrocell. We
set the radius of femtocell to 10m; each femtocell has one user. We assume that
FBS is loacaed at the center (0m) of each femtocells. Totally, 409 Femtocells are
distributed uniformly in a macrocell and the distance between adjacent FBS is
40m. The table 1 shows parameters used in performance evaluation.
The path loss model used in performance evaluation is as follows. First, we
use the modified COST 231-Walfisch-Ikegami urban macro model to know path
loss of macrocell user [13].
Pout [dB] = 31.81 + 40.5 log10 (d) + ψout (3)
164 Y. Jeong et al.
where the d [m] is distance between MBS and the its user. ψout [dB] is outdoor
shadowing modeled as log-normal distribution with zero mean, 8dB standard
deviation.
Second, we use Modified COST 231-multiwall model to know path loss of
femtocell user [14].
where the d [m] is distance between the FBS and the its user, L[dB] is path loss
by wall located in the edge of femtocell. We set to L = 8dB in this simulaiton.
ψin [dB] is indoor shadowing modeled as log-normal distribution with zero mean,
4dB standard deviation.
(a) radiation patterns with 0°, 90°, 180°, 270° (b) radiation patterns with 45°, 135°, 225°, 315°
where Pout [dB] is outdoor path loss model in Eq. (3). dout is the distance between
macrocell user and the wall located at the edge of femtocell. din is distance
between the FBS and the wall located in the edge of femtocell.
Figure 4 shows beam patterns of SPA antenna. The mesh size is λ/10 (λ
is wavelength.), the simulation is executed by the method of moments (MoM)
which is a full wave solution of Maxwell’s equation in the frequency domain. As
we can see in Figure 4, SPA antenna can radiate in one direction which has about
10dB front-to-back ratio symmetrically [10-12]. SPA has symmetric radiation
pattern for all directions and reasonable front-back ratio for each direction. Thus,
by controlling beam pattern of FBS according to its user’s location, we can
Avoidance of Co-channel Interference 165
1.0 1.0
CDF
0.4 0.4
0.2 0.2
0.0 0.0
0 10 20 30 40 50 60 -4 -2 0 2 4 6 8
Inner cell user SINR(dB) Outer cell user SINR(dB)
20
no femtocell
18
with femtocell
16 with femtocell and SPA
14
Throughput(Mbps)
12
10
0
inner region outer region
case, because there is less probability that macrocell users get interfered from
FBS than the second case. Third case shows that our proposed scheme improves
the throughput of macrocell users more than 4 times compared to the second
case.
5 Conclusion
Femtocells are small cellular base stations which extend indoor coverage and
provide high data rate. However, there exists co-channel interference between
macrocell and femtocells when they are operating in the same frequency band.
This co-channel interference causes severe performance degradation both macro-
cell and femtocells. Especially as FBS are using omni-directional antenna, they
cause co-channel interference to nearby macrocell users regardless of their loca-
tions. And macrocell user’s SINR gets lower as it gets closer to FBS. Therefore,
in this paper we have proposed procedure of data transmission and power control
for FBS using SPA antenna to reduce the co-channel interference with macrocell
users. In performance evaluation, we confirmed that macrocell user’s SINR is
improved by 60 ∼ 80% and macrocell throughput is increased more than four
times compared to FBS using omni-directional antenna.
Acknowledgment
References
1. Mansfield, G.: Femtocells in the US Market -Business Drivers and Consumer Propo-
sitions. FemtoCells Europe, ATT, London (2008)
2. Femto Forum, http://www.femtoforum.org/femto/
3. Chandrasekhar, V., Andrews, J., Gatherer, A.: Femtocell Networks: A Survey.
IEEE Comm. Mag. 46, 59–67 (2008)
4. Lopez-Perez, D., Valcarce, A., Roche, G., Zhang, J.: OFDMA Femtocells: A
Roadmap on Interference Avoidance. IEEE Comm. Mag. 47, 41–48 (2009)
5. Wang, Y., Kumar, S., Garcia, L., Pedersen, K., Lovacs, I., Frattasi, S., Marchetti,
N., Mogensen, P.: Fixed Frequency Reuse for LTE Advanced Systems in Local
Area Scenarios. In: VTC 2009-Spring, pp. 1–5 (2009)
6. Yavuz, M., Meshkati, F., Nanda, S., Pokhariyal, A., Johnson, N., Raghothaman,
B., Richardson, A.: Interference management and performance analysis of
UMTS/HSPA+ femtocells. IEEE Comm. Mag. 47, 102–109 (2009)
7. Claussen, H.: Performance of Macro- and Co-Channel Femtocells in a Hierarchical
Cell Structure. In: IEEE PIMRC 2007, pp. 1–5 (2007)
8. Guvenc, I., Jeong, M., Watanabe, F., Inamura, H.: A hybrid frequency assignment
for femtocells and coverage area analysis for co-channel operation. IEEE Commun.
Lett. 12, 880–882 (2008)
9. Claussen, H., Pivit, F.: Femtocell Coverage Optimization Using Switched Multi-
Element Antennas. In: IEEE ICC 2009, pp. 1–6 (2009)
10. Scott, N., Miles, O.: Diversity Gain from a single-Port Adaptive Antenna Using
Switched Parasitic Elements Illustrated with a Wire and Monopole Prototype.
IEEE Trans. Antennas Propagat. 47, 1–10 (1999)
11. Wennstom, M., Svantesson, T.: An antenna solution MIMO channels: The switched
parasitic antenna. Signal, Systems and Computer 2000 2, 1617–1621 (2000)
12. Mitilineos, S., Thomopoulos, S.: Development of a Compact SPA for 2.4GHz Ap-
plications using Commercially Available Elements: Design and Experimental Val-
idation. In: Loughborough Antennas and Propagation Conference 2009, pp. 16–17
(2009)
13. Baum, D., Hansen, J., Salo, J.: An interim channel model for beyond-3G systems:
extending the 3GPP spatial channel model (SCM). In: VTC 2005-Spring, pp. 3132–
3136 (2005)
14. European Cooperation in the Field of Scientific and Technical Reserch, EURO-
COST231, Digital Mobile Radio Towards Future Generation Systems, COST 231
final report, http://www.lx.it.pt/cost231/
15. IEEE 802.16m-08/004r5, IEEE 802.16m Evaluation Methodology Document, EMD
(2009)
16. Alipour, A., Hassani, H.: A Novel Omni-Directional UWB Monopole Antenna.
IEEE Trans. Antennas Propagat. 56, 3554–3857 (2008)
User Policy Based Transmission Control Method in
Cognitive Wireless Network
1 Introduction
Recently the development of wireless networks has been growing in various fields,
such as cellar phone, digital TV, and wireless Internet services. Those technologies
make us to expect the new application like the advanced traffic system, the disaster
prevention system, and so on. However, such a remarkable spread of wireless services
cause lack and inefficient usage of radio frequency wireless spectrum, we're in front
of a spectrum explosion, that is, one may not have enough wireless resources for new
wireless applications. According to FCC reports [6], wireless resources are limited
D. Taniar et al. (Eds.): ICCSA 2010, Part III, LNCS 6018, pp. 168–177, 2010.
© Springer-Verlag Berlin Heidelberg 2010
User Policy Based Transmission Control Method in Cognitive Wireless Network 169
and may not be efficient, because wireless device depends on access to the radio
frequency (RF) wireless spectrum, and spectrum has been chronically limited ever
since transmissions were first regulated in the early 20th century. Therefore, new
technologies that use spectrum more efficiently and more cooperatively, unleashed by
regulatory reforms, will soon overcome the spectrum shortage.
Cognitive wireless network (CWN) by J. Mitola III [8] from software defined
wireless was originally considered to improve spectrum utilization, and generally
considered as a technology to identify the opportunities using the “spectrum holes”
for telecommunications [1] [2]. In other words, Mitola defined that CWN is an
intelligent wireless communication system that is aware of its surrounding
environment, and uses the methodology of understanding by-building to learn from
the environment and adapt its internal states to statistical variations in the incoming
RF stimuli by making corresponding changes in certain operating parameters in real-
time, with two primary objectives in mind: highly reliable communication whenever
and wherever needed; efficient utilization of the wireless spectrum.
Also, CWN is consisted of the cogitation cycle; observing the environment,
orienting itself, creating plans, deciding, and then acting reconfigurations. The
cognition cycle is continually responding stimuli from environment, and it makes
successful link or path of network.
However, CWN still has some problems to realize like some algorisms, control
methods, and technical problems to attain efficient transmission. First of all, it is
recently known that CWN needs not only ordinal spectrum management or
Physical/Mac layer observation control but also new control method including upper
layers. Also, CWN still has technical problems like how to set radio frequency, how
to observe its environment, and how to select proper transmission protocol or radio
frequency, and so on [3][4][5]. Moreover, CWN needs the QoS algorism to select one
of wireless links and network route for each application or user environment.
In this paper, we consider the selecting method of the proper wireless link which
each node has several wireless links such as IEEE802.11a/b, and IEEE802.16
(WiMax), and the proper route in which network environment is changing through the
time. First, user policy and network parameters are observed at the observation stage.
At this stage, we set various policies for video, VoIP, text, disaster applications, and
so on. By each policy, the weight parameter is decided for the decision algorism
which based on Analytic Hierarchy Process (AHP). Network parameters like
throughput, latency, jitter, packet error rate, bit error rate, and electric field strength
are continuously measured and also used for the calculation of AHP. Secondly, at the
decision stage, the results of AHP for each communication link are compared, and the
proper link is selected when the results are changed. Also, if there is no proper
wireless links at the observing nodes, proper route is analyzed by extend Ad hoc On-
Demand Distance Vector (AODV) algorism. Finally, in the acting stage, the selected
link or route is applied for supposed network, and then simulation is carried out for
our supposed methods.
In the followings, in section 2, network model and system architecture of our proposal
communication method are defined. Section3 deals with the observation stage which is
the method of observing network parameters and user policy. Section 4 describes
wireless link selection, route selection and decision algorism which based on AHP.
Section 5 explains how to change link and route at supposed network. At section 6, the
170 N. Uchida, Y. Shibatta, and K. Takahata
simulation is held for the moodel by the calculation for the effectiveness of the suggessted
control method. Finally sectiion 6 derived our conclusion and future works.
2 Network Model
A. Network Configuration
The network configuration n of our suggested cognitive wireless network consistss of
several wireless nodes as sh
hown in Fig. 1.
Wireless nodes are supp posed as wireless terminal, and they have routing, adhhoc,
and multi-hop functions. Every
E node has multi wireless links like IEEE802.111a,
IEEE802.11b, IEEE802.16 (WiMax), and IMT2000. The lowest RF link is used as the
control links for the purpose of sending a network characteristic data or messagee of
switching links. Also, Antennna of each node is supposed as non-directional.
Moreover, the network condition is changed over time by node’s movementt or
radio interference like treees or buildings. A data transmission is carried out by
ordering user request from a server to a user terminal.
B. System Architecture
The system architecture is organized three layers, including the physical layer, the
network layer, and the appliication layer in Fig 2.
When transmission data sent from sender to receiver, network data from each layer
is observed through link0 that is the control link as explained in the previous section.
In the application layer, a user application is detected for the decision of user
policy. A user policy is assumed to various types like video, VoIP, text, connectivity
(emergence purpose like disaster) and so on. These policies are used for the decision
formula which will be mentioned in the later section.
The system layer collects type of coding and performs coding functions including
transcoding of a video coding to another one, such as from/to Motion JPEG to/from
MPEG1.
The network layer observes network conditions such as the value of PER,
throughput, delay, jitter, BER, electric field strength, and so on.
Network data from each layer is observed, and the decision making is held with
crossing through layers. Then, the message of prefer link or route is sent through
link0, and a reconfiguration procedure is acted at both sender and receiver nodes.
3 Methodology
As the previous research [8], the cognition cycle is introduced to our proposal
method. The cognition cycle is consisted of three stages; the observation stage, the
decision stage, and the acting stage. Each stage is continuously cycled in order to
perform link or route configuration.
A. Observation Stage
Network data is continuously observed through each layer at this observation stage.
Wireless network condition varies depending on the movement of nodes or radio
interference. Therefore, CWN needs to parse these stimuli to select the available
solution for providing the performance from user requests.
In this paper, our supposed system observes application types in order to decide
user policy which is depending on the specific services or media. Also, physical
characteristics like coding, PER, throughput, delay, jitter, BER, and electric field
strength are observed. Those parameters are used for understanding stimuli from user
environment.
B. Decision Stage
Decision making is held to maintain QoS in this decision stage. When the network
condition is changed, the proposal system will seek the suitable link and route by the
calculating from the values of network characteristics like user policy, throughput,
BER, and so on. We introduce AHP for the calculation of link, and extend AODV for
the decision of suitable route.
1) Link Selection
AHP is one of multi-attribute decision making and structured techniques for dealing
with complex decisions. It was developed by Thomas L. Saaty in the 1970s [9]. By
structuring a decision problem hierarchy, AHP provides us quantifications of its
elements and evaluations of alternative solutions.
For example, when the suitable link between neighbor nodes is solved by AHP, the
hierarchy of the problem is first structured. That is, goal (To decide suitable link of
wireless node), criteria (network characteristics such as delay, PER, throughput, and
172 N. Uchida, Y. Shibatta, and K. Takahata
⎛ n -l ⎞
Si = ⎜⎜ i i ⎟⎟ ×10 (In case of throughput) (4)
⎝ ui-li ⎠
⎛ n -l ⎞
Si = ⎜⎜1 − i i ⎟⎟ ×10 ( In case of BER, PER, latency, etc) (5)
⎝ ui-li ⎠
In the above formula, Si is the weight of each alternative, ui and li are the upper and
lower limits of the alternativ
ve, and ni is the observed value from network.
We also adopt the valu ue of consistency index (C.I.). C.I. expresses the chaara-
cteristics of polynominal of A, and we accept the estimation of w if C.I. is less tthan
certain threshold.
In this paper, we propose th he calculation of AHP based on each user policy like viddeo,
VoIP, or connectivity for decision
d making of the suitable link. For example, in ccase
of emergency like disaster,, connectivity is more important as user policy. Thus, the
criteria like electric field strrength and BER are weighed and calculated by the form mula
(3). Then, priorities of alteernatives is calculated by (4) and (5), and the results are
inserted for (3). Finally, eaach value of alternative is calculated by alternative prioority
multiplied with criteria prio ory, and the link with largest value will be decided as the
best suitable link.
2) Route Selection
Our proposed system would change network route if the suitable link is not foundd or
minimum requirement for user is not satisfied. When a network route needss to
change, we introduced extend AODV for the decision of suitable route.
AODV is a routing pro otocol for mobile ad hoc networks (MANETs) and otther
wireless ad-hoc networks.[10][11] It is a reactive routing protocol, which means tthat
it establishes a route to a deestination only on demand base. Therefore, the connecttion
is slower than proactive routing
r protocols like OLSR or TBRPF. But AODV V is
superior when the network condition changes so often or changes so slowly. [7].
AODV builds routes using a route request (RREQ) and route reply (RREP). When
a source node requires a route to a destination for which it does not have a route, the
source node broadcasts RREQ packets across the network. When the other nodes
receive those packets, and then update their routing information for the source node
and set up backwards pointers to the source node in the route tables.
The RREQ contains the source node's IP address, current sequence number,
broadcast ID, and the most recent sequence number for the destination node. A node
receiving the RREQ may send RREP packets if it is either the destination or it has a
route to the destination with corresponding sequence number greater than or equal to
that contained in the RREQ. If this is the case, RREP packets are sent back to the
source node with setting the routing tables relatively.
3) Extended AOCV
We propose a extended AODV protocol by adding the link values and network
conditions on each node for RREQ and RREP packets on each node from the source
to the destination. The link values of each node are calculated by AHP as the
previous sections, and those are set to the information of next node ID. Also, network
characteristics values of each node such as delay, PER, throughput, and so on are
added to RREQ and RREP packets. Then, a destination node receives all of the
RREQ packets from the possible routes during certain unit time. Then, possible routes
are compared to select the best route,
1) that can provide the maximum End-to-End throughput among all of the routes
for video service, or
2) that can provide the minimum End-to-End delay time among all of the routes for
VoIP, and
3) that can optimize the policy based AHP for Web service.
Through those comparisons, the best suitable route is decided by the policy of
transmission data.
We propose to add the link values by AHP for RREQ and RREP packets, and those
are set in the routing tables like Fig.5.
C. Acting Stage
After the decision making of the suitable link or the route, a link or route will be
changed in the acting stage.
In the proposed system, a control wireless link is assumed to use for a transmission
of the control information. A control link is set to the lowest frequency bandwidth,
and a node sends the information with a suitable link to the next node through the
User Policy Based Transmission Control Method in Cognitive Wireless Network 175
control link. Then, both nodes act the link change at the same time. A route change
also acts as the same way of a link change.
The cognition cycle goes back to the observation stage after the acting stage, and it
works to a link or route change relatively.
4 Simulation
In the simulation, ns2 (Network Simulator 2) was examined to evaluate the
effectiveness of our suggested method. IEEE802.11a, IEEE802.11b, IEEE802.11g is
used as the wireless links, and IMT2000 is used as the control link. The number of
wireless nodes is set to two, and every node has the same wireless links. Then, only a
source node moved with the same speed, and video data were transmitted from the
source node to the destination node. The simulation conditions are the following
table.
In this paper, a link change is especially focus on the simulation. First, the
simulation of IEEE802.11a, b and g is simulated under table1 conditions.
The Fig.6 shows the results of each node simulation. At Fig.6, throughput of
IEEE802.11a is quickly down about 18sec, and the link is disconnected about 23sec.
Also, IEEE802.11g is gradually down and disconnected about 145sec. IEEE802.11b
keeps link connection although throughput is not so high.
On the contrary, the results of our proposal method are showed in Fig.7.
Fig.7 shows that our proposed method keeps high throughput because the link
selection is effectively worked under video transmission scenario. The wireless link is
switched from IEEE802.11a to IEEE802.11g at 14 sec, and also switched IEEE802.
11g to IEEE802.11b at 132 sec. This results showed the proposed methods change the
best suitable link to maximize the transmission rate with user policy.
176 N. Uchida, Y. Shibata, and K. Takahata
5 Conclusions
In this paper, the transmission control methods in cognitive radio network
considering with cross layers including user policies are introduced. Compared with
giving fixed transmission rate, the our suggested methods can determines the
transmission rate dynamically to achieve the maximum throughput by changing
wireless links. The simulation result based on the video based data show that the
proposed methods change the best suitable link to maximize the transmission rate
with user policy. Therefore, the proposed method is shown in practical, efficient and
reliable to support data transmission.
User Policy Based Transmission Control Method in Cognitive Wireless Network 177
References
[1] Chen, K.C., Peng, Y.J., Prasad, N., Liang, Y.C., Sun, S.: Cognitive radio network
architecture: part I – general structure. In: ICUIMC 2008: Proceedings of the 2nd
international conference on Ubiquitous information management and communication
(January 2008)
[2] Chen, K.C., Peng, Y.J., Prasad, N., Liang, Y.C., Sun, S.: Cognitive radio network
architecture: part II – trusted network layer structure. In: ICUIMC 2008: Proceedings of
the 2nd international conference on Ubiquitous information management and
communication (January 2008)
[3] Cordeiro, C., Challapali, K., Ghosh, M.: Cognitive PHY and MAC layers for dynamic
spectrum access and sharing of TV bands. In: TAPAS 2006: Proceedings of the first
international workshop on Technology and policy for accessing spectrum (August
2006)
[4] Weingart, T., Sicker, D.C., Grunwald, D.: Evaluation of cross-layer interactions for
reconfigurable radio platforms. In: TAPAS 2006: Proceedings of the first international
workshop on Technology and policy for accessing spectrum (August 2006)
[5] Kliazovich, D., Granelli, F.: Packet concatenation at the IP level for performance
enhancement in wireless local area networks. Wireless Networks 14(4) (2008)
[6] Renk, T., Kloeck, C., Burgkhardt, D., Jondral, F.K., Grandblaise, D., Gault, S., Dunat,
J.-C.: Bio-inspired algorithms for dynamic resource allocation in cognitive wireless
networks. In: Mobile Networks and Applications, October 2008, vol. 13(5) (2008)
[7] Staple, G., Werbach, K.: IEEE Spectrum: The End of Spectrum Scarcity,
http://www.spectrum.ieee.org/telecom/wireless/
the-end-of-spectrum-scarcity
[8] Sugimoto, T., Yamaguhi, S., Asatani, K.: A Proposal for System Selection Scheme
Using Multiple Handover Triggers in Heterogeneous Wireless Networks, Technical
Report of IEICE
[9] He, W., Nahrstedt, K., Liu, X.: End to End Delay Control of Multumedia Applications
over Multihop Wireless Links. ACM Transaction on Multimedia Computing,
Communications and Applications 5(2), Article 16 (November 2008)
[10] Perkins, C.E., Royer, E.M.: Ad hoc On-Demand Distance Vector Routing. In:
Proceedings of the 2nd IEEE Workshop on Mobile Computing Systems and
Applications, February 1999, pp. 90–100 (1999)
[11] Chakeres, I.D., Royer, E.M.: AODV Routing Protocol Implementation Design. In:
Proceedings of the International Workshop on Wireless Ad Hoc Networking (WWAN)
(March 2004)
Development of a Hybrid Case-Based
Reasoning for Bankruptcy Prediction
1 Introduction
Bankruptcy prediction is one of the most important issues influencing financial and
investment decision-making [1]. While affecting the entire life span of a corporation
or an organization, a business failure can even undermine social functions and agitate
economical structures [2]. Therefore, the evaluation of business failure has emerged
as a professional area prompting academics and professionals to develop optimal pre-
diction models based on their specific interest for practical applications in the busi-
ness community.
Methods developed, such as univariate statistical model, Multiple Discriminant
analysis, Linear Probability models, Logistic Regression, and Probit analysis [3][4],
are now generally regarded as obsolete statistical methods [5]. Statistical methods
*
Corresponding author.
D. Taniar et al. (Eds.): ICCSA 2010, Part III, LNCS 6018, pp. 178–188, 2010.
© Springer-Verlag Berlin Heidelberg 2010
Development of a Hybrid Case-Based Reasoning for Bankruptcy Prediction 179
require a few assumptions and normally being constrained by the linearity, thus they
are quietly difficult to deal with the mass and complicated data [6]. To meet the var-
ied and restrictive assumptions, such as a large number of samples, normal distributed
independent variables, and linear relationship between all variables, scholars in re-
lated fields have come up with alternatives methods to predict the risk of business
failure, notably Artificial Neural Network (ANN), Case Based-Reasoning (CBR), and
Data Mining techniques [7][8][9][2].
Although reported to perform fairly well in terms of predictive accuracy, ANN
classifies every object by its computational characteristic and is therefore often criti-
cized as a “black box” approach for its lack of transparency [10]. In contrast to the
“black box” scenario, CBR, developed by Schank and Abelson [11] in the 1970s, uses
“similar” cases stored in a case base to build a solution to a new case based on the
adaptation of solutions to past similar cases [12]. Literature indicates that CBR has
been widely used in diagnosis domain and enhanced some of the deficiencies in statis-
tical models and ANN [13][14]. However, accuracy and effectiveness may be reduced
when CBR deals with too many attributes.
To solve the problem of excessive variables (attributes), Rough Set Theory (RST),
a data mining technique that has been successfully applied to solve classification
problems, can be adopted to reduce unnecessary attributes. Attributes, once identified,
need to be further prioritized by taking their weights into due consideration and this
task can be accomplished by performing Grey relational analysis (GRA) that com-
pares quantitative analysis to the development between every variable in the grey sys-
tem dynamically. The method describes the relation between the main factor (the
business status) and other variables in the grey system.
This study accordingly proposes a hybrid model for predicting business failure,
called RG-CBR, which incorporates three stages starting with applying RST to extract
key attributes, adopting GRA to obtain the weights of the key attributes, and culmi-
nating in feeding the retrieved key attributes and weights into CBR to improve the
model’s prediction accuracy.
Following the present section as the introduction, four major sections are incorpo-
rated in the paper. Section 2 outlines the backgrounds to RST, GRA, and CBR. Sec-
tion 3 explains the research methodology. Section 4 centers on experiment results and
model evaluations while Section 5 presents our conclusions.
2 Research Background
RST, a powerful mathematical tool introduced by Pawlak [15], can discover facts
from imperfect data [16]. Studies showed that RST has played an important role in
coping with classification problems [9][2][7]. It also has been successfully applied to
real-world classification problems with the following advantages [7]: (a). Identifying
potentially important information contained in a large number of cases and reducing
the information into a model including a generalized description of the knowledge;
(b). Requiring no interpretation as the model is a set of easily understandable
180 R.-H. Lin and C.-L. Chuang
Deng [18] pioneered a mathematical method called Grey Relational Analysis (GRA)
which is characterized by good performance in analyzing few data and many vari-
ables to examine the relationship among factors in observable systems and construct
the prediction model. GRA has been successfully used in a wide range of fields [19],
including agriculture, traffic, industrial engineering, education. Only a few studies,
however, have applied it to tackle business prediction problems.
An analogical reasoning method developed by Schank & Abelson [11] in the 1970s,
CBR can serve as a useful research paradigm for exploring complex problems [12] by
adapting past experiences to acquire potential solutions. As a flourishing problem-
solving method, CBR has been successfully applied to a wide spectrum of different
fields, such as medical diagnosis [20], bankruptcy prediction [8], fault diagnosis [21],
risk analysis [22], and marketing [2]. The study by Bryant [8] that adopts CBR to
bankruptcy prediction modeling in particular underlines the importance of selecting
proper variables in the successful application of the method. Complete reviews of CBR
with comprehensive details and applications for further works can be found in the
study of Watson & Marir [23].
Data Processing
3.1 RST
At the beginning, data processing is initiated to collect samples and establish criteria
for sampling and coding. RST is adopted to reduce attributes; some redundant attrib-
utes are eliminated without any information loss to induct the minimal variables sub-
sets named “reducts.” The following sections present some basic concepts of RST.
A finite set of objects U and a finite set of attributes A can be considered as a re-
lational system S = (U , A) . Each attribute a belongs to the considered set of attributes
A ( a ∈ A ); a (x ) represents the value of attribute a of object x. For every set of at-
tributes B∈A, an indiscernibility relation Ind (B ) is defined in the following way:
Ind ( B ) = {( x , x ) ∈UXU ; b ( x ) = b ( x ) ; b ∈ B}
i j i j (1)
Grey relational analysis (GRA) can measure the degree of interrelationships between
one major sequence and another sequence in a given system [24]. The relational grade
is regarded as high if the two sequences tend towards concordance; otherwise, the
relational grade is regarded as low. Assume that the objective sequence is defined as
Given the objective sequence X 0 (k ) and the reference sequence X i (k ) with the
* *
normalized form, the grey relational coefficient r0i (k ) between them at any data
point k is defined as
Δ min + ρΔ max
r0i (k ) = . (6)
Δ 0i (k ) + ρΔ max
Δ = x *0 (k ) − x *i (k ) (7)
min min min
i k
Δ min and Δ max are respectively the minimum and maximum distances for all factors
in all sequences. ρ is the distinguishing coefficient (0 ≤ ρ ≤ 1) , and usually
ρ = 0.5 . After the grey relational coefficient is obtained, the grey relational grade,
i.e. the mean of the coefficient, can be measured by
1
μB F and μ R F is called the quality of approximation of partition F by set of attributes B
and R.
Development of a Hybrid Case-Based Reasoning for Bankruptcy Prediction 183
1 n
γ (X 0 , X i ) = ∑ r0i (k )
n k =1
(9)
∑ w × sim( f
i =1
i i
I
, fiR )
(10)
Similarity Score = n
∑w
i =1
i
samples. A qualified sample, i.e. a failed firm, refers to a company with the transac-
tion of its listed securities placed under the “altered-trading method” category, sus-
pended or terminated. More specifically, a firm is cited as failed for (a). Suffering
from credit crisis, (b). Having net operating loss, (c). Failing to pay debts, or (d). Vio-
lating related regulations.
In order to improve the comprehensiveness and accuracy in predicting, this study
will concern simultaneously non-financial (qualitative) and financial (quantitative)
attributes and then apply them to diagnose the business status with different models.
21 attributes (13 financial ones and 8 non- financial) used in previous researches are
selected for our study and coded and described in Table 1.
The sample data sets consist of an equal number of healthy and failed companies
(321 companies in each category). Sample companies are grouped by binary assign-
ment into a decision class (healthy or failed, coded by 1 and 0, respectively). The data
sampling follows the “pairwised sampling method” first proposed by Beaver [3]. Each
data set is split into two subsets: a training set of 70% and a validation set of 30% of
the data, respectively. The training data are used as a case base, and the validation data
not used in constructing the case base are adopted to test the model. Five different
validation sets, each with six entries of data, are sampled and then used to validate the
model performance; that is, a total of 97 test samples are used for validation.
As RST is concerned with discrete values, the experimental study uses the software
ROSETTA that includes a method for discretizing continuous attributes called en-
tropy heuristic approach. The approach will subject continuous data to a process of
discretization.
Attributes Description
A1 Return on assets (%)
A2 Return on equity (%)
A3 Net income (%)-except disposed
A4 Gross margin (%)
A5 Net income (%)
A6 Current ratio (%)
A7 Acid test ratio (%)
A8 Liabilities ratio (%)
A9 TCRI credit ranking
A10 Cash flow operation to current liabilities (%)
A11 Total equity growth ratio (%)
A12 Return on total assets growth ratio (%)
A13 Accounts receivable turnover (day)
A14 Inventory turnover (%)
A15 Earning per share
A16 Added value per person
A17 Manager-director
A18 Director and supervisor shareholding
A19 Inflation rate (%)
A20 Business cycle
A21 Rediscount Rate (%)
Development of a Hybrid Case-Based Reasoning for Bankruptcy Prediction 185
D A6 A7 A8 A9 A16 A18
1 57.91 27.46 26.5 8 1289.5 19.41
1 32.57 21.25 80.2 9 590.4 12.37
1 63.69 54.94 100.55 10 1188.81 7.45
1 144.5 88.71 43.55 8 2494.46 8.59
1 80.98 35.69 78.36 9 181.06 11.4
1 143.61 85.9 71.51 10 47.41 24.49
1 249.79 86.01 85.27 10 619.21 35.29
1 71.01 51.43 82.89 10 1105.08 25.36
1 2.35 1.24 94.2 10 0.78 8.89
1 512.16 438.14 11.32 8 489.64 14.74
1 199.93 151.11 44.92 5 8.68 5.7
1 117.09 63.13 52.87 9 978.6 13.46
1 225.23 143.6 44.52 6 669.41 10.47
1 30.87 64.94 97.29 10 635.85 11.61
1 209.79 158.66 50.24 10 2017.14 14.75
1 298.29 242.31 18.77 5 424.59 35.36
1 135.29 128.63 63.15 7 243.78 8.88
1 133.45 121.01 55.77 8 1377.4 9.4
1 147.06 123.46 69.68 6 747.69 16.41
1 167.64 137.2 46.55 2 1852.75 13.61
The discernibility matrix can be used to find the minimal subset(s) of attributes
called reducts who have the same quality as the complete set of attributes. As a result,
we further apply RST to reduce the redundant attributes and to combine with CBR for
lowering misclassification rate. At the beginning, there were 78 reducts computed.
These reducts were employed from the total 21 attributes. According to the results
obtained from RST, the measuring data of each key attribute is shown in Table 2. The
weights of those key attributes are developed by GRA in Excel, as described in Table
3 which shows the grey relational coefficient of each attribute adjusted its weight. The
six attributes are A6 (Current ratio), A7 (Acid test ratio), A8 (Liabilities ratio), A9
(TCRI credit ranking), A16 (Added value per person) and A18 (Director and supervi-
sor shareholding).
The six attributes are ranked based on their weights into the following order: TCRI
credit ranking, Liabilities ratio, Director and supervisor share-holding, Acid test ratio,
Current ratio, and Added value per person. An attribute is ignored during similarity
computation if it is found to have zero weight, whereas an attribute with a higher
weight exerts a greater impact on determining similarity. As indicated above, TCRI
credit ranking emerges as the most significant attribute influencing business failure
prediction with a highest weight of 0.529.
Identifying key attributes is a sub-process of investigating the important attributes
of factors critical for identifying analogous cases and of predicting the value of the
target variable. RST for identifying key features and GRA for ranking key features by
their weights are applied to compensate the weaknesses of CBR. And the enhanced
CBR forms the core of the proposed RG-CBR model. The model then proceeds to
search in the case base the case that best matches with the target case and to adapt the
retrieved case to fit the problem and produce the solution for the problem. In the ex-
periment, the overall predictive accuracy of the validation data emerges to be 83.3%
as shown in Table 4.
To evaluate the prediction accuracy of RG-CBR model, Table 4 compares the re-
sults of the three different models of RG-CBR, CBR, and RST-CBR (combining RST
and CBR with equal weights). Among the three models, RG-CBR shows the highest
level of accuracy (83.3%) in the given validation data. Therefore, the result concludes
that RG-CBR outperforms CBR and RST-CBR and testifies to its qualification to
serve as a promising alternative for business failure prediction.
5 Conclusions
CBR, as indicated in the literature, has been widely used in diagnosis domain by re-
trieving most similar cases more accurately and effectively. The accuracy and effec-
tiveness, however, may be reduced when there are too many attributes. It is therefore
not easy to obtain successful results by adopting CBR alone to predict business fail-
ure. In order to enhance the performance of the CBR system, this study proposes a
hybrid model, RG-CBR, for predicting business failure. RG-CBR can refer the user to
the similar cases as references for making decisions, and the attributes retrieved and
assigned different weights by mathematical methods (RST with GRA) are more ob-
jective than those decided by experts’ subjective experiences. As suggested by the
results of the experiment, RG-CBR outperforms other comparative algorithms such as
CBR, and RST-CBR in terms of accuracy and serves a promising alternative for busi-
ness failure prediction.
Development of a Hybrid Case-Based Reasoning for Bankruptcy Prediction 187
References
[1] O’Leary, D.E.: Using neural networks to predict corporate failure. International Journal
of Intelligent Systems in Accounting, Finance and Management 7, 187–197 (1998)
[2] Ahn, B.S., Cho, S.S., Kim, C.Y.: The integrated methodology of rough set theory and ar-
tificial neural network for business failure prediction. Expert Systems with Applica-
tions 18, 65–74 (2000)
[3] Beaver, W.H.: Financial ratios as predictors of failure- empirical research in accounting:
selected studies. Journal of Accounting Research 4, 71–111 (1966)
[4] Ohlson, J.A.: Financial ratios and the probabilistic prediction of bankruptcy. Journal of
Accounting Research 18, 109–131 (1980)
[5] Balcaen, S., Ooghe, H.: 35 years of studies on business failure an overview. The British
Accounting Review 38, 63–93 (2006)
[6] Akahoshi, M., Amasoki, Y., Soda, M.: Correlation between fatty liver and coronary risk
factors: a population study of elderly men and women in Nagasaki, Japan. Hypertens Re-
search 24, 337–343 (2001)
[7] Slowinski, R., Zopounidis, C.: Application of the rough set approach to evaluation of
bankruptcy risk. International Journal of Intelligent Systems in Accounting, Finance and
Management 4, 27–41 (1995)
[8] Bryant, S.M.: A case-based reasoning approach to bankruptcy prediction Modeling. In-
ternational Journal of Intelligent Systems in Accounting, Financial and Management 6,
195–214 (1997)
[9] McKee, T.E.: Developing a bankruptcy prediction model via rough set theory. Interna-
tional Journal of Intelligent Systems in Accounting, Finance and Management 9, 159–
173 (2000)
[10] Kumar, P.R., Ravi, V.: Bankruptcy prediction in banks and firms via statistical and intel-
ligent techniques – A review. European Journal of Operational Research 180, 1–28
(2007)
[11] Schank, R. (ed.): Dynamic Memory: A Theory of Learning in Computers and People.
Cambridge University Press, New York (1982)
[12] Kolodner, J.: Improving human decision making through case-based decision aiding. AI
Magazine 12, 52–68 (1991)
[13] Barletta, R.: An introduction to case-based reasoning. AI Expert 6, 42–49 (1991)
[14] Park, T.S., Han, J.Y.: Derivation and characterization of pluripotent embryonic germ
cells in chicken. Molecular Reproduction and Development 56, 475–482 (2000)
[15] Pawlak, Z.: Rough sets. International Journal of Information and Computer Sciences 11,
341–356 (1982)
[16] Walczak, B., Massart, D.L.: Rough sets theory. Chemometrics and Intelligent Laboratory
Systems 47, 1–16 (1999)
[17] McKee, T.E., Lensberg, T.: Genetic programming and rough sets: A hybrid approach to
bankruptcy classification. European Journal of Operational Research 138, 436–451 (2002)
[18] Deng, J.: Control problems of grey systems. System and Control Letters 1, 288–294
(1982)
[19] Lai, H.H., Lin, Y.C., Yeh, C.H.: Form design of product image using grey relational
analysis and neural network models. Computers & Operations Research 32, 2689–2711
(2005)
[20] Varma, A., Roddy, N.: ICARUS: Design and deployment of a case-based reasoning sys-
tem for locomotive diagnostics. Engineering Applications of Artificial Intelligence 12,
681–690 (1999)
188 R.-H. Lin and C.-L. Chuang
[21] Yang, B.S., Han, T.H., Kim, Y.S.: Integration of art-kohonen neural network and case-
based reasoning for intelligent fault diagnosis. Expert Systems with Applications 26,
387–395 (2004)
[22] Jung, C., Han, I., Suh, B.: Risk analysis for electronic commerce using case-based rea-
soning. International Journal of Intelligent Systems in Accounting, Finance and Man-
agement 8, 61–73 (1999)
[23] Watson, I., Marir, F.: Case-based reasoning: A review. The Knowledge Engineering Re-
view 9, 327–354 (1994)
[24] Hsu, Y.T., Chen, C.M.: A novel fuzzy logic system based on N-version programming.
IEEE Transactions on Fuzzy Systems 8, 155–170 (2000)
[25] Kolodner, J.: Case-based reasoning. Morgan Kaufmann, San Mateo (1993)
[26] Shin, K.S., Han, I.: A case-based approach using inductive indexing for corporate bond
rating. Decision Support Systems 32, 41–52 (2001)
[27] Burke, E.K., MacCarthy, B., Petrovic, S., Qu, R.: Structured cases in case-based reason-
ing-re-using and adapting cases for timetabling problems. Knowledge-Based Systems 13,
159–165 (2000)
The Time Series Image Analysis of the HeLa Cell Using
Viscous Fluid Registration
Abstract. Optical microscopy image analysis is important in the life science re-
search. To obtain the motion of the cell, we use the viscous fluid registration
method based on fluid dynamics. Viscous fluid registration deforms an image at
time t to the next image at time t+1. In this algorithm, there is a problem that an
object cannot be divided into two. In other words, the divided objects from one
are connected by thin line because the velocity field on the connected thin line
is zero. To solve this problem, we suggest a new viscous fluid registration algo-
rithm for the object division. This algorithm is only added similarity maximiza-
tion step to correct the displacement in the near pixels in the original viscous
fluid registration. Using this method, one object is divided into two, and divided
objects are not connected. We experiment the anaphase detection based on a
nucleus identification using laser scanning microscope HeLa cell images. Ex-
perimental result shows that 74 in 76 cells are tracking well and 6 cells in the
anaphase are detected. In three scenes in the cell division which can not be
divided into two using original viscous fluid registration, suggested algorithm
can be divided into two cells completely.
Keywords: HeLa cell, cell division, cell tracking, viscous fluid registration.
1 Introduction
Microscopy image analysis is important in the life science research. Cell phase identi-
fication and analysis of the cell features are attracted. The abnormal cell detection
method helps early detection of a cancer, diabetes, and the gene deficit disease.
Therefore early treatment for disease is enabled and helps medical care.
In the field of the cell analysis, various image processing methods are performed in
cell segmentation and the phase identification. For example, it is suggested the accu-
rate automated method that cell segmentation using watershed algorithm and the cell
phase identification using Markov model by Zhou et al. [1] and Chen et al. [2]. Dzyu-
bachyk et al. suggested the multiple level-set framework that each cell have each
level-set function [3].
We think that the cell movement can be expressed by the viscous fluid, so the cell
movement can be described as the time series image matching technique. We use the
D. Taniar et al. (Eds.): ICCSA 2010, Part III, LNCS 6018, pp. 189–200, 2010.
© Springer-Verlag Berlin Heidelberg 2010
190 S. Tokuhisa and K. Kaneko
image registration method applied in the medical image processing such as the CT
image and MRI image. Though this method has a large calculation cost, it is obtained
the smooth displacement because it is based on the fluid dynamics. The displacement
contains vectors xt – xt-1. It means that a pixel at position xt at time t is moved from a
pixel at position xt-1 at time t-1.
There are some kinds in a microscopy images such as an optical microscopy image
and confocal laser scanning microscope. The three-dimensional cell image can be
constructed using confocal laser scanning microscope. In this experiment, we use
these cell images in Fig.1 that are obtained by live cell imaging [4] of the HeLa cell
with mCherry-NLS[5] via the confocal laser scanning microscope created by Yuki
Tsujimura in RIKEN.
When an object is divided into two objects, viscous fluid registration cannot to be
separate correctly. So, we suggest the new algorithm that is added one calculation step
in the viscous fluid registration which step is the similarity maximization step to cor-
rect the displacement in the near pixels.
When this method is used, we can divide an object to two objects completely. Us-
ing viscous fluid registration with similarity maximization step, it will be performed a
smooth cell movement analysis. The displacement can be use for the time series im-
age analysis like Fig.2. The displacement is useful for anaphase detection because the
corresponding time series cell at time t and cell at time t+1 can be detected. The cell
in anaphase is defined the first time of the cells called daughter cell which was di-
vided one cell into two in this article. From the cell identification result, it can be
confirmed the movement of the daughter cells after the cell division. Cell identifica-
tion result shows that 74 in 76 cell in 17 images is success in tracking.
t =1 t =2 t =3 t =4 t =5 t =6
Fig. 2. Example of the time series image analysis. The displacement calculated by viscous fluid
registration shows the white arrow in right image named “Overlapped image”. These arrows
show a pixel in time01 is moved to a pixel in time02.
Sample01
Sample02
Fig. 3. Example of a force field, a velocity field and a displacement calculated first time
192 S. Tokuhisa and K. Kaneko
sample01 sample02
viscosity 20
rMIN 1
rMAX 8
calculation num.100
Overlapped image Deformed sample01
parameter
input Under calculation Output
Fig. 5. Overlapped current frame image (orange) on next frame image (purple)
Viscous fluid registration is the image registration method to align two images by
deforming one image to the other [7] [8]. A viscous fluid registration is based on the
theory of a viscous fluid dynamics. This algorithm contains the calculation of three
vector fields: a force field, a velocity field, and a displacement. These vector fields are
a set of two-dimensional vectors that are x-axis and y-axis values because we perform
two-dimensional viscous fluid registration for all images. We use sum of squared
difference as an image similarity measure. Sum of squared difference is used to calcu-
late the force field. The force field is equal to the derivative of image similarity meas-
ure. The force field equation can be written as follow:
f ( x , u( x, t )) = − [T ( x - u( x, t )) − S ( x )]∇T ( x - u( x, t )) (1)
The Time Series Image Analysis of the HeLa Cell Using Viscous Fluid Registration 193
where f(x, u(x, t)) is the force field at point x and at time t that depends on the dis-
placement u(x, t),. T(x) is the pixel value at point x on template image T, and S(x) is
the pixel value at point x on study image S. ∇T is calculated by sobel filter. The
velocity field is equal to the convolution of the force field with the Gaussian kernel.
The velocity field v(x,u(x,t)) can be written as follow:
v ( x , u( x, t ) ) = Gσ (f( x, u( x, t )) (2)
The Gaussian kernel Gσ has two parameters: window size w and standard deviation σ.
We define this standard deviation σ as the viscosity. The value of the viscosity deter-
mines the degree of freedom of the deformation to limit the movement of the pixel at
point x. The window size w is defined as follow:
6σ
w= (3)
r
where r is the resolution level that allows for reducing the image size. If the multi-
resolution method is not used, the resolution level is set to 1. Using this window size
w, the velocity at x is convolved in 99.74% force field affected to the viscosity at x
according to Gaussian distribution. Displacement u(x, t) is calculated as follow:
u( x, t i +1 ) = u( x, t i ) − (t i +1 − t i ) (I − ∇u (x, t i )) v(x, t i ) (4)
where the velocity field v(x,u(x,t)) is equal to the Lagrange derivative of displace-
ment u(x, t). Matrix I is the identity matrix. The time (ti+1−ti) is calculated by the
following equation:
MAX (| v ( x , t i ) |) (t i +1 − t i ) = du max ⋅ r (5)
where dumax is the maximal movement allowed in one iteration. We set dumax as a
fixed value of 0.7 (pixel). If |I – ∇ u(x, t)| is less than 0.5, the displacement u(x, t) is
applied to the image, and the displacement u(x, t) and time t are initially set to 0,
because the displacement is singular for large curved deformation. The total dis-
placement is calculated as follow:
utotal ( x, t i ) = u( x, t i ) + utotal ( x − u( x, t i ), t i -1 ) (6)
Viscous fluid registration algorithm requires four input parameters: the maximum
resolution level rMAX, the minimum resolution level rMin, the calculation number k,
and the viscosity σ. We use a multi-resolution method that enables fast and robust
image registration. In the multi-resolution method, image Ir in the current resolution
level r is used. Image Ir is reduced in size by resolution level r. If resolution level r is
2, the image size is reduced by half. Image Ir is defined as follows:
I r ( x r , y r ) = I 1 (r ⋅ x r , r ⋅ y r ) (7)
1≤ x r ≤
width
r
1 ≤ yr ≤
height
r
(8)
194 S. Tokuhisa and K. Kaneko
where xr and yr are array numbers and they are constrained by (9). The current resolu-
tion level r is taken as 1, 2, 4, 8,…,2n. The values of the width and height are obtained
by the image width and image height at resolution level 1.
The viscous fluid registration method is calculated three vector field. In the first
step, it calculates the force field value using data from two images. In the second step,
it calculates the velocity field using the force field. In the third step, it calculates the
displacement using the velocity field. When these three steps are finished, it incre-
ments the iteration number and repeats the three steps. If the iteration number is equal
to the calculation number, the resolution level is divided by 2, and the iteration num-
ber is initialized to 0. After initialization, the three calculation steps are repeated. The
resolution level indicates the scale of the image. For example, if the resolution level is
1, it indicates the original image, and if the resolution parameter is 4, it indicates that
an image size is quarter. At the beginning of a calculation, the resolution level is set to
the maximum resolution level. If the resolution level is equal to the minimum resolu-
tion level, it generates output data, including the total displacement and the image
deformed using displacement.
To divide one object into two, we suggest adding the similarity maximization step in
viscous fluid registration. To confirm this technique, we used two image set in Fig.6,
left image set is shown that one circle is divided into two circles, and right image set is
shown that one square is divided into two squares. Using original viscous fluid regis-
tration, two objects are connected by the thin line (see 40calculation in Fig.7). The
force field over and under the line have in the opposite direction (see Fig.8). Because
the velocity is calculated by the convolution of the force field, the velocity field on the
line is calculated to zero. And the displacement is in proportion to the velocity, dis-
placement on the connected line is zero. After all calculation, there are gaps between
the displacement on the connected line and over and under the line. Therefore I intro-
duce a similarity maximization step to approximate the displacement at the point x by
near the point x. The similarity maximization step is described as follow:
u total ( x, t i ) = MIN{(T (x - u( x, t )) − S (x) ) | x − 1 ≤ x ≤ x + 1}
2
(9)
This equation maximizes an image similarity measure by minimizing the square error
of the difference. This step is added after the calculation of the total displacement utotal
in (6). Using this step, deformed result is shown in Fig.9, and the deformed grid im-
age is shown in Fig.10. Fig.10 shows that image edge is enhanced in circle grid image
using the new step. This means that fluency of the motion vector is decreased, and the
issue of connected line was solved.
Fig. 6. two image set. Left is circle division example. Right is square division example
The Time Series Image Analysis of the HeLa Cell Using Viscous Fluid Registration 195
䊶䊶䊶
Connected line
Fig. 7. Cconnected line is not go disappear using original viscous fluid registration. Top images
are circle image result. Bottom images are a square image result.
Fig. 8. Force field, velocity field, and displacement at 7calculation in circle image
Fig. 9. Disappear connected line adding Similarity maximization step in registration. Top line
images are circle image result. center and bottom line images are sqare image result.
196 S. Tokuhisa and K. Kaneko
Fig. 10. Deformed grid image. Left two images are deformed circle grid result. Right two im-
ages are deformed square grid result. In these set, left is not using step result, right is using step
result.
3 Experiment
We experiment the detection of the anaphase in the time series HeLa cell images.
Anaphase detection consists of the three steps: image processing, viscous fluid regis-
tration and cell identification. We use 17 bitmap images in Fig.1. These images are
350*350 size and 8bit grayscale data created by Yuki Tsujimura in RIKEN. These
images contain 76 cells. The resolution of the image is 0.286um in x and y direction
and time span between images is 1 hour.
At first we generate a label image by image processing. This step is described in
3.1. In the next step, we perform viscous fluid registration time series images. This
second step is described in 3.2. Finally, we perform time to time cell identification by
using of the displacement calculated by viscous fluid registration. If corresponding
cell is not found after the registration, these cells make pair of a nearest cell in the
previous time image. After three steps, we make cell identification in all time, so we
can detect a cell in anaphase because these cells are divided into two. This last step
and anaphase detection are described in 3.3.
For the cell identification, we generate a label image from grayscale image using
hybrid image processing. This image processing is implemented by the octave lan-
guage. This method contains five processes: thresholding, median filter, erosion,
labeling and dilation(Fig.11). Thresholding process make binary image from gray-
scale image. Median filter process removes white noise in the cell from binary image.
Erosion process makes the cell area shrink. Labeling process add label data on the
cell. Dilation process makes the cell area inflated. Dilation and erosion is used for
division cell because some cells are touching together. After labeling, label image is
displayed by a color using color table. So, label image have pixel values such as 0 on
the background and 1, 2, 3… on each independent cell. In this process, 17 label im-
ages are created. Regulating parameters at each time are described in Table 1.
time 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17
threshold 30 30 30 30 30 30 30 50 70 70 70 30 30 30 30 60 60
erosion and diration
repeat numver 0 0 0 0 0 0 0 5 6 5 5 0 0 0 0 0 0
The Time Series Image Analysis of the HeLa Cell Using Viscous Fluid Registration 197
In this step, we input two time series images into viscous fluid registration to get the
displacement. The label images are not used in this step. The result of the original
viscous fluid registration is shown in Fig.12, and the result of the suggested viscous
fluid registration added the similarity maximization step is shown in Fig.13. We can
find that a cell is divided obviously in suggested technique. As for this technique,
limitation of the transformation becomes lax, so the deformed cell shape almost same
as input study image.
By the next step, a label image in Fig.11 is deformed using the displacement, and
determined the correspondence of the pair of the cells in time t and time t+1 image. If
the cells in time t and time t+1 image exist in the same position, these cells are treated
as the correspondence of the pair. And when cell division occurs, cell at time t is
divided into two cells at time t+1, so we can distinguish the cells in Anaphase by
checking label numbers are same value or not.
Fig. 12. Original viscous fluid registration result when cell division is occurred
Fig. 13. Suggested viscous fluid registration result when cell division is occurred
198 S. Tokuhisa and K. Kaneko
In this step, cell identification and anaphase detection is performed. The label images
and the displacement is used for cell identification. Current time label of cell is taken
the correspondence of the next time label. The correspondence of the label is found
from the relations of a cell position between deformed current time label and next
time label. We calculate histogram of deformed current time label image and next
time label image. And next time label value is changed the value of the most over-
lapped label from deformed current time label image (Fig.14 left). Most overlapped
label is found from the histogram. If there are same labels in the next time label im-
age, it found that anaphase is detected. Once anaphase is detected in one label value,
Anaphase
Fig. 15. Result of a detection of anaphase. Dividing cell in anaphase is shown in the red circle
The Time Series Image Analysis of the HeLa Cell Using Viscous Fluid Registration 199
create flag on this label value. This mean anaphase is already found on this label. So,
if flag is created, anaphase is not detected on this label. There is the case that registra-
tion is fault because the next time cell and deformed cell is not overlapped well. In
other words, the corrected label value candidate has the background value (Fig.14
right). When a label candidate became the background, nearest neighbor label is
given. In 17 images, two cells are detected by nearest neighbors. This nearest
neighbor process is adapted only center of the image. If the center of gravity of a label
exists on the area of 40 pixels from boundary of the image, this cell is not regard as in
anaphase because this cell moves image boundary to inside and outside. Fig.15 shows
the result of corrected all label images.
Fig. 16. Overlapped time series label images. This image shows the dividing nucleus move-
ment. The boundary line of nucleus that appears the first is white and that appear the last time is
black. A cell color from the second and before last nucleus is semitransparent. Dark blue nu-
cleus is divided and daughter nuclei are moved to left and right (yellow allow), green daughter
ones are moved to top and right-top (red allow), and blue ones are moved to top and bottom
(purple allow).
5 Conclusion
We can find that divided objects are connected by a thin line using original viscous
fluid registration. So we suggest a new viscous fluid registration algorithm for an
object division. This algorithm is to add new step in original viscous fluid registration.
200 S. Tokuhisa and K. Kaneko
We confirmed that connected line which appeared in original viscous fluid registra-
tion result is disappeared by adding new step in two sample image set and time series
HeLa cell images. We examine the anaphase detection to confirm that new algorithm
is worked well. By creating label image and deforming label image using displace-
ment, the cell identification is success 74 in 76 cells. 2 of the remainder cell can be
tracked by nearest neighbor processing. A cell in anaphase is detected be cell identifi-
cation, and daughter cell movement is observed by the overlapped time series label
image. Displacement calculated by viscous fluid registration is useful in various ways.
Using displacement, we can observe cell movement like Fig.2.
References
1. Zhou, X., Li, F., Yan, J., Wong, S.T.: A novel cell segmentation method and cell phase
identification using Markov model. IEEE transactions on information technology in bio-
medicine 13, 152–157 (2009)
2. Chen, X., Zhou, X., Wong, S.T.: Automated segmentation, classification, and tracking of
cancer cell nuclei in time-lapse microscopy. IEEE Transactions on Biomedical Engineer-
ing 53, 762–766 (2006)
3. Dzyubachyk, O., Niessen, W., Meijering, E.: Advanced level-set based multiple-cell seg-
mentation and tracking in time-lapse fluorescence microscopy images. Biomedical Imag-
ing 21, 185–188 (2008)
4. Goldman, R.D., Spector, D.L.: Live Cell Imaging: A Laboratory Manual. Cold Spring Har-
bor Laboratory Press (2005)
5. Shu, X., Shaner, N.C., Tarbrough, C.A., Tsien, R.Y., Remington, S.J.: Novel Chromophores
and Buried Charges Control Color in mFruits. Biochemistry 45, 9639–9647 (2006)
6. Zitova, B., Flusser, J.: Image registration methods: a survey. Image and Vision Comput-
ing 21, 977–1000 (2003)
7. D’Agostino, E., Maes, F., Vandermeulen, D., Suetens, P.: A Viscous Fluid Model for Mul-
timodal Non-rigid Image Registration Using Mutual Information. In: Dohi, T., Kikinis, R.
(eds.) MICCAI 2002. LNCS, vol. 2489, pp. 541–548. Springer, Heidelberg (2002)
8. Bro-Nielsen, M., Gramkow, C.: Fast fluid registration of medical images. Visualization in
Biomedical Computing, 267–276 (1996)
Matrices Representation of Multi Soft-Sets and
Its Application
1 Introduction
Soft set theory [1], proposed by Molodtsov in 1999, is a new general method for
dealing with uncertain data. In recent years, research on soft set theory has been
active, and great progress has been achieved, including the works of theoretical soft
set, soft set theory in abstract algebra, parameterization reduction, decision
making and forecasting. Let S = (U , A, V , f ) be an information system as in [2].
The “standard” soft set deals with a binary-valued information system
S = (U , A, V{0,1} , f ) . For a multi valued information system, we introduced a concept of
multi soft-sets [3]. The idea is based on a decomposition of a multi-valued
information system S = (U , A,V , f ) , into A number of binary-valued information
systems S = (U , A,V{0,1} , f ) , where A denotes the cardinality of A. Consequently, the
A binary-valued information systems define multi-soft sets, denoted by
(F , A) = {(F , a i ) : 1 ≤ i ≤ A } . In [4], we used the concept of multi-soft sets and AND
operation for finding reducts in a multi-valued information system. However, the
comparison of the proposed reduct has not been presented yet, especially with rough-
set based reduct [2,5]. In this paper, we present the notion of matrices representation
of multi-soft sets. We present a definition of AND and OR operations on a collection
D. Taniar et al. (Eds.): ICCSA 2010, Part III, LNCS 6018, pp. 201–214, 2010.
© Springer-Verlag Berlin Heidelberg 2010
202 T. Herawan, M.M. Deris, and J.H. Abawajy
of such matrices. Further, we apply the notion of AND operation for finding reducts
and core attributes in a multi-valued information system. We prove that the proposed
technique of reduct is equivalent with Pawlak’s rough reduct.
The rest of this paper is organized as follows. Section 2 describes the notion of
information system. Section 3 describes fundamental concept of soft set theory.
Section 4 describes multi soft sets construction in a multi-valued information system.
Section 5 describes matrices representation of multi-soft sets, AND and OR operations,
as well a simple example. Section 6 describes applications of matrices representation
of multi-soft sets for finding reducts and core of attributes. Further, we prove that our
proposed technique for reduction is equivalent to Pawlak’s rough reduction. Finally,
we conclude and describe future activities of our works in section 7.
2 Information System
Data are often presented as a table, columns of which are labeled by attributes, rows
by objects of interest and entries of the table are attribute values. By an information
system, we mean a 4-tuple (quadruple) S = (U , A,V , f ) , where U = u1 , u 2 , u 3 ,L, u U { } is
{
a non-empty finite set of objects, A = a1 , a 2 , a3 ,L, a A } is a non-empty finite set of
attributes, V = U V , V is the domain (value set) of attribute a, f : U × A → V is an
a∈ A
a a
U a1 … ak … aA
u1 f (u1 , a1 ) … f (u1 , a k ) … ( )
f u1 , a A
u2 f (u 2 , a1 ) … f (u 2 , a k ) … f (u , a )
2 A
M M O M O M
uU (
f u U , a1 ) … (
f u U , ak ) … (
f u U ,a A )
A relational database may be considered as an information system in which rows
are labeled by the objects (entities), columns are labeled by attributes and the entry in
row u and column a has the value f (u, a ) . We note that a each map
(
f (u , a ) : U × A → V is a tupple ti = f (ui , a1 ), f (ui , a2 ), f (ui , a3 ),L , f ui , a A ( )) , for
1 ≤ i ≤ U . In many applications, there is an outcome of classification that is known.
Note that the tuple t is not necessarily associated with entity uniquely (refer to objects
3 and 4 in Table 3). In an information table, two distinct entities could have the same
tuple representation (duplicated/redundant tuple), which is not permissible in
relational databases. Thus, the concepts in information systems are a generalization of
the same concepts in relational databases.
Matrices Representation of Multi Soft-Sets and Its Application 203
□
From Proposition 2, it is easily to understand that a binary-valued information system
can be represented as a soft set. Thus, we can make a one-to-one correspondence
between (F , E ) over U and S = (U , A, V{0,1} , f ) .
Definition 3. (See [6].) The class of all value sets of a soft set (F , E ) is called value-
class of the soft set and is denoted by C( F , E ) .
204 T. Herawan, M.M. Deris, and J.H. Abawajy
{
A = a1 , a 2 ,L , a A } { }
into the disjoint-singleton attribute {a1 }, {a 2 }, L , a A . Here, we
only consider for complete information systems. Let S = (U , A, V , f ) be an
information system such that for every a ∈ A , Va = f (U , A) is a finite non-empty set
and for every u ∈ U , f (u , a ) = 1 . For every a i under i th-attribute consideration,
a i ∈ A and v ∈ Va , we define the map a vi : U → {0,1} such that a vi (u ) = 1 if
f (u , a ) = v , otherwise a vi (u ) = 0 . The next result, we define a binary-valued
information system as a quadruple S i = (U , ai , V{0,1} , f ) . The information systems
S i = (U , ai , V{0,1} , f ) , 1 ≤ i ≤ A is referred to as a decomposition of a multi-valued
information system S = (U , A, V , f ) into A binary-valued information systems, as
depicted in Figure 1. Every information system S i = (U , ai ,Vai , f ) , 1 ≤ i ≤ A is a
deterministic information system since for every a ∈ A and for every u ∈ U ,
f (u , a ) = 1 such that the structure of a multi-valued information system and A
number of binary-valued information systems give the same value of attribute related
to objects.
(
We define (F , A) = (F , a1 ), (F , a 2 ), L , F , a A ( )) as a multi-soft sets over universe U
representing a multi-valued information system S = (U , A, V , f ) .
U a1 a2 … ak … aA
u1 f (u1 , a1 ) f (u1 , a 2 ) … f (u1 , a k ) … ( )
f u1 , a A
u2 f (u 2 , a1 ) f (u 2 , a 2 ) … f (u 2 , a k ) … f (u , a )
2 A
u3 f (u 3 , a1 ) f (u 3 , a 2 ) … f (u 3 , a k ) … f (u , a )
3 A
M M M O M O M
uU (
f u U , a1 ) (
f u U , a2 ) … (
f u U , ak ) … (
f uU ,a A )
Binary-valued information system-1
a1
Va11 Va12 … Va1k … Va1n
U
u1 0 1 … 0 … 0
u2 0 0 … 1 … 0
u3 0 1 … 0 … 0
M M M O M O M
uU 1 0 … 0 … 0
M
Binary-valued information system- A
aA
Va A 1 Va A 2 … Va A k … Va A n
U
u1 0 0 … 1 … 0
u2 0 0 … 0 … 1
u3 0 1 … 0 … 0
M M M O M O M
uU 1 0 … 0 … 0
The concept of matrix representing multi-soft sets is described here. In this sub-
section, the notation (F, A) represents a multi-soft set over universe U of a multi-
valued information system S = (U , A, V , f ) .
⎧ (F , a1 ) ⇔ S 1 = (U , a1 ,V{0,1} , f ) ⇔ M a
⎪ (F , a )
⇔ S 2 = (U , a2 ,V{0 ,1} , f ) ⇔ M a
1
(F , A) ⇔ ⎪⎨ M 2 M M
2
.
⎪
( ) ( )
⎪ F , a A ⇔ S A = U , a A ,V{0,1} , f ⇔ M A
⎩
Definition 5. Matrix M a , 1 ≤ i ≤ A is called matrix representation of soft set (F , a1 )
( )
i
⎧⎪0, if f (u, α ) = 0
aij = ⎨ , 1 ≤ i ≤ U , 1 ≤ j ≤ Vai , u ∈ U and α ∈ Vai .
⎪⎩ 1, if f (u, α ) = 1
M A
{
= M a :1 ≤ i ≤ A .
i
}
Matrices Representation of Multi Soft-Sets and Its Application 207
{
C M = {u : f (u , α1 ) = 1}, L, u : f u , α Va
ai
{ ( i
) = 1}},
where 1 ≤ i ≤ Vai , u ∈ U and α ∈ Vai .
Clearly CM ⊆ P(U ) .
ai
M A
{
= Ma ,Ma ,Ma ,Ma ,Ma ,
1 2 3 4 5
}
where
⎡1 0 0⎤ ⎡1 0 0⎤ ⎡1 0⎤
⎢1 0 0⎥⎥ ⎢0 1 0⎥⎥ ⎢1 0⎥
⎢ ⎢ ⎢ ⎥
⎢0 1 0⎥ ⎢0 1 0⎥ ⎢1 0⎥
⎢ ⎥ ⎢ ⎥ ⎢ ⎥
0 0 1⎥ 0 0 1⎥ 1 0⎥
M Diploma = ⎢ , M Experience = ⎢ , M French = ⎢ ,
⎢0 0 1 ⎥ ⎢ 1 0 0⎥ ⎢1 0⎥
⎢ ⎥ ⎢ ⎥ ⎢ ⎥
⎢0 0 1⎥ ⎢0 0 1⎥ ⎢1 0⎥
⎢1 0 0 ⎥ ⎢0 0 1⎥ ⎢0 1 ⎥
⎢ ⎥ ⎢ ⎥ ⎢ ⎥
⎢⎣0 1 0⎥⎦ ⎢⎣0 1 0⎥⎦ ⎢⎣0 1⎥⎦
⎡1 0 0⎤ ⎡1 0⎤
⎢0 1 0⎥⎥ ⎢0 1⎥⎥
⎢ ⎢
⎢0 0 1⎥ ⎢0 1⎥
⎢ ⎥ ⎢ ⎥
0 1 0⎥ 1 0⎥
M Reference = ⎢ and M Decision = ⎢ .
⎢0 1 0 ⎥ ⎢ 0 1⎥
⎢ ⎥ ⎢ ⎥
⎢1 0 0⎥ ⎢1 0⎥
⎢0 0 1 ⎥ ⎢1 0⎥
⎢ ⎥ ⎢ ⎥
⎢⎣1 0 0⎥⎦ ⎢⎣0 1⎥⎦
where
CM Diploma
= {{1,2,7}, {3,8}, {4,5,6}} , C M Experience
= {{1,5}, {2,3,8}, {4,6,7}} ,
CM French
= {{1,2,3,4,5,6}, {7,8}} , C M Reference
= {{1,6,8}, {2,4,5}, {3,7}} ,
and C M Decision
= {{1,4,6,7}, {2,3,5,8}} ,
respectively.
The AND and OR operation of the matrices are described in this sub-section.
208 T. Herawan, M.M. Deris, and J.H. Abawajy
1 ≤ n ≤ Va be two matrices in
i
M A
. The " AND" operation M a ANDM a of matrices i j
⎡1 0 0⎤ ⎡1 0 0⎤ ⎡1 0 0 0 0 0 0 0 0⎤
⎢1
⎢ 0 0⎥⎥ ⎢0
⎢ 1 0⎥⎥ ⎢⎢0 1 0 0 0 0 0 0 0⎥⎥
⎢0 1 0⎥ ⎢0 1 0⎥ ⎢0 0 0 0 1 0 0 0 0⎥
⎢ ⎥ ⎢ ⎥ ⎢ ⎥
0 0 1⎥ 0 0 1⎥ ⎢0 0 0 0 0 0 0 0 1⎥
M Diploma ANDM Experience = ⎢ AND ⎢ = ,
⎢0 0 1 ⎥ ⎢ 1 0 0 ⎥ ⎢0 0 0 0 0 0 1 0 0⎥
⎢ ⎥ ⎢ ⎥ ⎢ ⎥
⎢0 0 1⎥ ⎢0 0 1 ⎥ ⎢0 0 0 0 0 0 0 0 1⎥
⎢1 0 0⎥ ⎢0 0 1 ⎥ ⎢0 0 1 0 0 0 0 0 0⎥
⎢ ⎥ ⎢ ⎥ ⎢ ⎥
⎢⎣0 1 0⎥⎦ ⎢⎣0 1 0⎥⎦ ⎢⎣0 0 0 0 1 0 0 0 0⎥⎦
where
CM Diploma ANDM Experience = {{1}, {2}, {3,8}, {4,6}, {5}, {7}} .
where
CM Diploma ORM Experience = {{1}, {2}, {3,8}, {4,6}, {5}, {7}} .
d. M a OR M a ORM a
i j k ai aj ak associative
6 Application
In this section, we present an application of matrices representation of multi-soft sets
for dimensionality reduction and finding core attributes.
In [4], we presented the application of multi-soft sets for finding reducts. The
proposed approach is based on AND operation in multi-soft sets [3]. In this sub-
section, we explore the applicability of matrices representation for finding reducts and
core attributes. We show that the reducts obtained are equivalent with that in [4]. The
notion of reduct in [4] is redefined and given as follow.
Definition 13. Let (F , A) = ((F , ai ) : 1 ≤ i ≤ A ) be multi-soft set over U representing a
multi-valued information system S = (U , A, V , f ) . A set of attributes B ⊆ A is called
a reduct for A if
CF (b ×L×b ) = CF (a ×L×a ) and CF (b ×L×b ) ≠ CF (a ×L×a ), B* ⊂ B ,
1 B 1 A 1 B* 1 A
where
(F , a × a ) = (F , a )AND(F , a ) .
i j i j
A core of A is defined as
CORE ( A) = I RED( A) ,
versa. □
Example 15. For simple example of soft reduction in multi-valued information
system, we consider a small dataset as used in [4].
U a1 a2 a3 a4
1 low bad loss small
2 low good loss large
3 high good loss medium
4 high good loss medium
5 low good profit large
⎢0 1⎥ ⎢0 1 ⎥ ⎢1 0 ⎥ ⎢0 0 1⎥
⎢1 0⎥⎦ ⎢0 1 ⎥ ⎢0 1 ⎥ ⎢0 1 0⎥⎦
⎣ ⎣ ⎦ ⎣ ⎦ ⎣
Since A = {a1 , a2 , a3 , a4 } , then
⎡1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0⎤
⎢0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0⎥
M ANDa :1≤i≤4 = ⎢0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0⎥
i ⎢ ⎥
⎢0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0⎥
⎢⎣0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 0⎥⎦
where
CM ANDai :1≤ i ≤ 4
= {{1}, {2}, {3,4}, {5}} .
Matrices Representation of Multi Soft-Sets and Its Application 211
⎢0 0 0 0 0 0 1 0⎥
⎢0 0 0 1 0 0 0 0⎥⎦
⎣
and
⎡1 0 0 0 0 0⎤
⎢0 1 0 0 0 0⎥⎥
⎢
M a ANDM a = ⎢0 0 1 0 0 0⎥
⎢ ⎥
3 4
⎢0 0 1 0 0 0⎥
⎢0 0 0 0 1 0⎥⎦
⎣
where
CM a1 ANDM a 2 ANDM a3 = {{1}, {2}, {3,4}, {5}} and CM a1 ANDM a 2 ANDM a3 = {{1}, {2}, {3,4}, {5}} ,
In rough set theory [2,5], the aim of dimensionality reduction is to keep only those
attributes that preserve the indiscernibility relation and, consequently, set
approximation. The remaining attributes are redundant since their removal does not
worsen the classification. There are usually several such subsets of attributes and
those which are minimal are called reducts. While computing equivalence classes is
straightforward, finding all reducts is NP-hard. This reduction is different in soft set
theory. The existing techniques of soft parameterization are still based on a binary
information system [8,9]. Soft parameterization reduction is obtained based on the
optimal and sub-optimal choice related to each object. Thus, the idea of reduction
under rough set theory generally cannot be applied directly in reduction under soft set
theory. In this sub-section, we show that our proposed technique on dimensionality
reduction in an multi-valued information system is equivalent with that in [2,5]. Let,
firstly, we recall the notion of rough reduction as follow.
Let S = (U , A, V , f ) be an information system and let B be any subset of A. Two
elements x, y ∈ U are said to be B-indiscernible (indiscernible by the set of attribute
B) if and only if f (x, a ) = f ( y, a ) , for every a ∈ B . Obviously, every subset of A
induces unique indiscernibility relation. Notice that, an indiscernibility relation is an
equivalence relation. The partition of U induced by IND(B ) in S = (U , A, V , f ) is
denoted by U / B .
212 T. Herawan, M.M. Deris, and J.H. Abawajy
a. U / B =U / A
b. U / (B − {b}) ≠ U / A, ∀b ∈ B
A core of A is defined as
CORE ( A) = I RED( A) ,
where RED( A) is the set of all reducts of A.
Proposition 17. Soft reduction in Definition 13 is equivalent with rough reduction in
Definition 16.
Proof. (⇒ ) It is clear that (F , A) = (U , A,V , f ) , thus CF (a ) = U / ai , 1 ≤ i ≤ A . From
i
CF (b ×L×b ) = CF (a ×L×a )
(F , b )ANDL AND(F , b ) = (F , a )ANDL AND(F , a )
1 B 1 A
1 B 1 A
CF (b ) I L I C F (b ) = C F (a ) I L I C F (a )
1 B 1 A
I 1≤i ≤ B
CF (b ) = I 1≤i≤ A CF (a )
i i
I 1≤ i ≤ B
U / bi = I 1≤ i ≤ A U / ai
U / B =U / A
CF (b ×L×b ) ≠ CF (a ×L×a )
1 B* 1 A
U / (B − {b}) ≠ U / A
(⇐) Obvious. □
⎡1 0 0⎤ ⎡1 0 0⎤ ⎡1 0⎤ ⎡1 0 0⎤
⎢1 0 0⎥⎥ ⎢0 1 0⎥⎥ ⎢1 0⎥⎥ ⎢0 1 0⎥⎥
⎢ ⎢ ⎢ ⎢
⎢0 1 0⎥ ⎢0 1 0⎥ ⎢1 0⎥ ⎢0 0 1⎥
⎢ ⎥ ⎢ ⎥ ⎢ ⎥ ⎢ ⎥
=
⎢0 0 1⎥
AND ⎢
0 0 1⎥
AND ⎢
1 0⎥
AND ⎢
0 1 0⎥
⎢0 0 1 ⎥ ⎢ 1 0 0 ⎥ ⎢ 1 0 ⎥ ⎢ 0 1 0⎥
⎢ ⎥ ⎢ ⎥ ⎢ ⎥ ⎢ ⎥
⎢0 0 1⎥ ⎢0 0 1⎥ ⎢1 0⎥ ⎢ 1 0 0⎥
⎢1 0 0 ⎥ ⎢0 0 1 ⎥ ⎢0 1⎥ ⎢0 0 1⎥
⎢ ⎥ ⎢ ⎥ ⎢ ⎥ ⎢ ⎥
⎣⎢0 1 0⎦⎥ ⎣⎢0 1 0⎦⎥ ⎣⎢0 1⎦⎥ ⎣⎢1 0 0⎦⎥
The calculation of the above AND operation is given as follows
⎛ ⎡1 0 0⎤ ⎡1 0 0⎤ ⎞ ⎛ ⎡1 0⎤ ⎡1 0 0⎤ ⎞
⎜⎢ ⎥ ⎢ ⎥⎟ ⎜⎢ ⎥ ⎢0 ⎟
⎜ ⎢1 0 0⎥ ⎢0 1 0⎥ ⎟ ⎜ ⎢1 0⎥ ⎢ 1 0⎥⎥ ⎟
⎜ ⎢⎢ ⎥ ⎢ ⎥⎟ ⎜⎢ ⎟
⎜ ⎢0 1 0⎥ ⎢0 1 0⎥ ⎟ ⎜ ⎢1 0⎥ ⎢0 0 1⎥ ⎟
⎥ ⎢ ⎥ ⎥ ⎢ ⎥
⎜ ⎢0 0 1 ⎥ ⎢0 0 1 ⎥ ⎟ ⎜ 1 0⎥ 0 1 0⎥ ⎟
= ⎜⎢ ⎥ AND ⎢ ⎥ ⎟AND⎜ ⎢ AND ⎢ ⎟
0 0⎥ ⎟ ⎢ 0⎥ ⎢0 1 0⎥ ⎟
⎜ ⎢⎢0 0 1⎥
⎥
⎢1
⎢ ⎥ ⎜ ⎢1 ⎥ ⎢ ⎥
⎜ ⎢0 0 1⎥ ⎢0 0 1⎥ ⎟ ⎜ ⎢1 0⎥ ⎢1 0 0⎥ ⎟
⎜⎢ ⎥ ⎢ ⎥⎟ ⎜⎢ ⎥ ⎢0 ⎟
⎜ ⎢⎢1 0 0⎥ ⎢0 0 1⎥ ⎟ ⎜ ⎢0 1 0 1⎥ ⎟
⎥ ⎢ ⎥ ⎥ ⎢ ⎥
⎜ ⎢0 ⎟ ⎜ ⎢0 1⎦⎥ 0 0⎦⎥ ⎟⎠
⎝⎣ 1 0⎦⎥ ⎢0
⎣
⎥
1 0⎦ ⎠ ⎝⎣ ⎣⎢1
⎛ ⎡1 0 0 0 0 0 0 0 0⎤ ⎞ ⎛ ⎡1 0 0 0 0 0⎤ ⎞
⎜⎢ ⎥⎟ ⎜⎢ ⎟
⎜ ⎢0 1 0 0 0 0 0 0 0⎥ ⎟ ⎜ ⎢0 1 0 0 0 0⎥⎥ ⎟
⎜ ⎢⎢ ⎥⎟ ⎜⎢ ⎟
⎜ ⎢0 0 0 0 1 0 0 0 0⎥ ⎟ ⎜ ⎢0 0 1 0 0 0⎥ ⎟
⎥ ⎥
⎜ ⎢0 0 0 0 0 0 0 0 1 ⎥ ⎟ ⎜ 0 1 0 0 0 0⎥ ⎟
= ⎜⎢ ⎥ ⎟ AND⎜ ⎢ ⎟
⎢ 1 0 0 0 0⎥ ⎟
⎜ ⎢⎢0 0 0 0 0 0 1 0 0⎥ ⎟
⎥ ⎜ ⎢0 ⎥
⎜ ⎢0 0 0 0 0 0 0 0 1⎥ ⎟ ⎜ ⎢1 0 0 0 0 0⎥ ⎟
⎜⎢ ⎥⎟ ⎜⎢ ⎟
⎜ ⎢⎢0 0 1 0 0 0 0 0 0⎥ ⎟ ⎜ ⎢0 0 0 0 0 1⎥ ⎟
⎥ ⎥
⎜ ⎢0 ⎥ ⎟ ⎜ ⎢0 0 0 1 0 0⎦⎥ ⎟⎠
⎝⎣ 0 0 0 1 0 0 0 0⎦ ⎠ ⎝⎣
⎡1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 ⎤
⎢0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 ⎥
⎢0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
⎥
= ⎢ ⎥
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0
⎢0 ⎥
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
⎢0 0⎥
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
⎢⎣0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0⎥⎦
Thus we have
CM Diploma ANDM Experience ANDM French M Reference = {{1}, {2}, {3}, {4}, {5}, {6}, {7}, {8}} ,
214 T. Herawan, M.M. Deris, and J.H. Abawajy
7 Conclusion
The “standard” soft set deals with a binary-valued information system. For a multi-
valued information system, the notion of multi-soft sets has been proposed. The idea
of multi-soft sets is based on a decomposition of a multi-valued information system
into binary-valued information systems. In this paper, we have presented the notion of
matrices representation of such multi-sot sets. The AND and OR operations in a
collection of matrices sets also presented. The AND and OR operations satisfy
idempotent and associative properties, neither closed nor commutative. Further, we
have proven that AND operation can be applied for finding reducts. Finally, we have
proven that our proposed technique provide the same results with that rough reduct.
For future research, we further elaborate our approach for finding functional, identity
and attribute dependencies in information systems. Since, the calculation of matrices
is complex, we will use MATLAB to solve this problem. We believe that our
proposed approach can be used for feature selection and data cleansing in multi-
valued information system under soft set theory.
Acknowledgement
This work was supported by the FRGS under the Grant No. Vote 0402, Ministry of
Higher Education, Malaysia.
References
1. Molodtsov, D.: Soft Set Theory-First Results. Computers and Mathematics with
Applications 37, 19–31 (1999)
2. Pawlak, Z., Skowron, A.: Rudiments of Rough sets. Information Sciences 177(1), 3–27
(2007)
3. Herawan, T., Mustafa, M.D.: On Multi-soft Sets Construction in Information Systems. In:
Huang, D.-S., Jo, K.-H., Lee, H.-H., Kang, H.-J., Bevilacqua, V. (eds.) ICIC 2009. LNCS
(LNAI), vol. 5755, pp. 101–110. Springer, Heidelberg (2009)
4. Herawan, T., Rose, A.N.M., Mustafa, M.D.: Soft Set Theoretic Approach for
Dimensionality Reduction. In: Communication of Computer and Information Sciences,
vol. 64, pp. 180–187. Springer, Heidelberg (2009)
5. Pawlak, Z.: Rough Sets: A Theoretical Aspect of Reasoning about Data. Kluwer Academic
Publisher, Dordrecht (1991)
6. Maji, P.K., Biswas, R., Roy, A.R.: Soft Set Theory. Computers and Mathematics with
Applications 45, 555–562 (2003)
7. Komorowski, J., Ohrn, A.: Modeling prognostic power of cardiac tests using rough sets.
Artificial Intelligence in Medicine 15, 167–191 (1999)
8. Chen, D., Tsang, E.C.C., Yeung, D.S., Wang, X.: The Parameterization Reduction of Soft
Sets and its Applications. Computers and Mathematics with Applications 49, 757–763 (2005)
9. Zhi, K., Gao, L., Wang, L., Li, S.: The Normal Parameter Reduction of Soft Sets and Its
Algorithm. Computers and Mathematics with Applications 56, 3029–3037 (2008)
Clustering Analysis of Water Quality for Canals in
Bangkok, Thailand
1 Introduction
Water will become the major constraining resource for sustainable development of
large areas in the world. Water quality is one of the main characteristics of a river,
even when its purpose is other than human water supply [4]. Water quality in the
superficial waters has started to degenerate as a result of waste-water being let go to
the receiving ground and surface water without any controls [5].
Bangkok is the capital city, as well as, the economic center of Thailand. It’s
activities, which include commercial, industrial and service have caused the expansion
of the city and its population to accumulate environmental pollution to the point that
nature cannot cope with the pollution loading, especially for water quality [21].
At present, the canal water pollution is very severe as the canals still act as sewers
for direct discharge of waste-water in most areas of Bangkok even though there is
already a requirement for large buildings to treat waste-water and houses must have at
least septic tanks. However, other waste-waters are discharged without treatment to the
public sewers which drain into the canals [22].
D. Taniar et al. (Eds.): ICCSA 2010, Part III, LNCS 6018, pp. 215–227, 2010.
© Springer-Verlag Berlin Heidelberg 2010
216 S. Areerachakul and S. Sanguansintukul
The network of canals (shown in Fig.3) is important for the daily life of the people
in Bangkok. These canals are used for consumption, transportation and recreation.
Therefore, the rapid growth of industry, condominium, high-rise and low-rise building,
and other infrastructure, have a significant effect on the canal water. The major impact
can be classified into 3 types: [21]
• Impact to tourism activities (shown in Fig.1): Bangkok is the capital city and
the center of tourism. Therefore, deterioration of canal water has direct
impact to tourism activity. As pollution in canals is in the inner area of
Bangkok, it certainly gives negative impression to the tourists who travel and
stay in Bangkok.
• Impact to aquatic life (shown in Fig.2): Basically, water pollution causes
deaths of aquatic creatures either by its toxicity or reducing dissolved oxygen
concentration. The toxicity may come from high concentration of sulfide or
ammonia. It is observed that when water is polluted there are usually fall in
the total number of species of organisms generally more sensitive than
fishes, a change in the type of species present and a change in the numbers of
individuals of each species in the water. In severe cases, where dissolved
oxygen concentration becomes zero, all aquatic creatures die and this occurs
in the highly polluted canals in Bangkok.
• Impact to public health: poor sanitary conditions prevail in many parts of
Bangkok. Records of some water related diseases generally associated with
sanitary conditions are discussed in the Annual Epidemiological Surveillance
Report of the Epidemiological Subdivision, Disease Control Division
Department of Health [23]. In case of acute diarrhea, about 45,000 and
50,000 cases were found in 1996 and 1997 respectively.
Fig. 1. The impact to tourism activities Fig. 2. The impact to aquatic life
218 S. Areerachakul and S. Sanguansintukul
3 Methodology
Clustering is an important data mining technique that has a wide range of applications
in many areas like biology, medicine, market research, image processing, and
geographical information systems, among others [24].
Clustering Analysis of Water Quality for Canals in Bangkok, Thailand 219
In this study, the water qualities of canals in Bangkok were grouped according to
their characteristic forming clusters. The clustering process was carried out using a K-
means algorithm and Fuzzy c-means algorithm.
J = ∑∑ x −c
i j
(1)
j =1 i =1
where x is the data point , c is the j cluster centroid , n is the number of data
x −c
i j
is a chosen distance measure between a data point x and the cluster
centroids c .
Fuzzy c-means (FCM) is a method of clustering which allows one piece of data to
belong to two or more clusters. This method is frequently used in pattern recognition
[18]. It is based on minimization of the following objective function:
∑
4
∑
This iteration will stop when
where is a termination criterion between 0 and 1 and k are the iteration steps. This
procedure converges to a local minimum or a saddle point of .
The algorithm is composed of the following steps.
1. Initialize U = [ ] matrix, U(0)
2. At k-step: calculate the centers vectors C(k)=[ ] with U(k)
3. Update U(k), U(k+1)
4. If 1 < then STOP; otherwise return to step 2.
The criteria widely accepted for partitioning a data set into a number of clusters are
the separation of the cluster and their compactness. The optimum case implies
parameters that lead to partitions that are as close as possible (in term of similarity) to
the real partitions of the data set [13]. A reliable quality assessment index should
consider both the compactness and the separation. One of the quality measures that
can be used in clustering is described as follows [8], [13].
The compactness of spatial data set , called , the value of the p-th dimension
is defined as follows: [8], [13]
1
σ x x 6
n
x ∑ x , x x 7
The total compactness of spatial data set with respect to c cluster is:
⁄
Further the term is defined as: = , where is a vector
The average compactness of c cluster, Comp is :
Comp = / (10)
The average scattering of data set compactness, Scatt_Comp is:
Scatt_Comp = Comp/ (11)
The more compact the cluster are the smaller the Scatt_Comp. Thus, for a given
spatial data set, a smaller Scatt_Comp indicates a better clustering scheme. The
distance between cluster is defined by the average distance between the centroid of
specified clusters, that is:
∑ ∑
12
1
The larger d is the more separated the clusters are according to definitions, a quality
measure for clustering was defined as follows:
_ / (13)
The definition of CD indicated that both criteria of “good” clustering (i.e.,
compactness and separation) are properly combined, enabling reliable evaluation of
clustering results [8]. Small values of CD indicate all the cluster in clustering scheme
are overall compact and separated.
The K-means clustering and Fuzzy c-means clustering methods were applied in this
study intending to divide the groups of several sites of canals. Distances between site
222 S. Areerachakul and S. Sanguansintukul
and clusters are computed as Euclidean distances in the K-dimension space, a number
of clusters N were designated arbitrary. As a result, very different means are obtained
between clusters for most K parameters used in the analysis [10].
Because the distance in the K-dimensional space is greatly affected by differences
in scale among the variables from which the distance are computed, each variable has
been standardized before implementation of cluster analysis.
The cluster analysis of each initial matrix has been carried out in several steps,
increasing the cluster number K from 2 to 6. From determination of optimized cluster
small values of CD indicate all the cluster in clustering scheme are overall compact
and separated that both criteria of “good” clustering.
Number of K Value of CD
K-means Clustering Fuzzy c-means Clustering
K=2 1.599 2.256
K=3 1.445 2.433
K=4 1.193 2.463
K=5 0.682 1.963
K=6 0.965 1.971
Cluster
Cluster Cluster Cluster Cluster Cluster
1 2 3 4 5
Parameters
Temp 29.19 29.19 27.96 28.87 29.22
pH 6.91 6.97 7.14 6.98 7.10
DO 0.04 1.08 2.17 0.37 2.77
H2S 1.76 0.11 0.08 0.76 0.01
BOD 40.92 12.40 9.33 23.81 6.05
COD 100.89 46.92 42.99 67.21 33.94
SS 26.68 33.05 39.08 23.61 28.56
TKN 13.98 7.15 4.64 9.59 4.34
NH3N 9.42 2.15 1.40 3.71 0.78
NO2N 0.04 0.12 0.11 0.05 0.11
NO3N 2.28 1.69 0.66 1.48 1.67
T-P 1.75 0.99 0.67 1.37 0.47
T.Coliform 6.42E+09 2.18E+08 1.53E+06 2.64E+08 3.03E+07
Table 4. Water quality of records of site and number of sites in each cluster groups
1 Klong Kumuengderm 12 - 4 4 2 2
2 Klong Robkrung 24 1 4 7 8 4
3 Klong Mahanark 12 - 1 5 1 5
Klong
4 20 - 2 9 4 5
PadungKrungkasem
5 Klong Samsen 16 2 3 8 2 1
6 Klong Sansab 40 - 9 21 4 6
7 Klong pramprachakorn 32 1 6 10 8 7
8 Klong Bangsue 12 1 1 3 5 2
9 Klong Satorn 8 1 2 2 1 2
10 Klong Somdetjaopraya 8 - 1 4 3 -
11 Klong Huai Khwang 16 - 4 7 4 1
12 Klong Lat phrao 16 - 4 11 - 1
13 Klong Mon 8 - 2 4 2 -
14 Klong Bangkoknoi 20 - - 15 4 1
15 Klong Prapa 8 - - 5 3 -
16 Klong Bangkokyai 20 - 2 16 2 -
17 Klong Bang khen 20 - 1 15 4 -
18 Klong Bangyikon 12 - 1 11 - -
19 Klong Bang Po 8 - - 8 - -
20 Klong Tair 8 - - 8 - -
21 Klong Lat Yao 8 - - 8 - -
22 Klong Suan Lhong 8 - - 8 - -
23 Klong Bangkabue 4 - - 4 - -
24 Klong Songtonnoon 4 - - 4 - -
Total 344 6 47 197 57 37
In this study, water quality data were collected during 2005-2008. There are 24
canals of 344 records of sites. Bangkok’s canals are preferred for city transportation
by many locals. The network of Bangkok Canals branches out of the Chao Phraya
River, "the river of the kings", finally meeting with the ocean. The canals in Bangkok
are hundreds of miles long. Not as many canals are used nowadays, and today both
sides of the Chao Phraya River are totally different from what they used to be in the
old days. Now, alongside the older homes and buildings, there are beautiful hotels.
Table 4 shows cluster 1 consisting of 6 of 344 records of site, cluster 2 consists of 47
of 344 records of site, cluster 3 consists of 197 of 344 records of site, cluster 4
consists of 57 of 344 records of site and cluster 5 consists of 37 of 344 records of site,
respectively.
Cluster 1 represents 6 records of site with a high value of BOD and low value of
DO in surface water canals. The reasons for higher BOD and lower DO are site of
canals connecting to industrial and populated accommodations. Most of the buildings
in this area are the commercial and high rise buildings.
Cluster 2 represents 47 records of site and cluster 4 represents 57 records of site
don’t show the various site in the study area.
Cluster 3 represents 197 records of site and all sites of canal number 19 to 24
appeared in this cluster. The site of those canals are close to the sludge treatment
centers that are knowledgeable about waste-water treatment including impact of
pollution in surface water of canals. As a result, the canals of cluster 3 according to
low value of BOD and high value of DO.
Cluster 5 represents 37 records of site and each site lowest value of BOD, COD
and TKN but highest value of DO. Because, each site of canals in cluster 5 is near to
pumping stations, non-point pollution from extractive and heavy industries, and some
site collected to rainy season of high flow water, it can be summarized that the best
water quality is in cluster 5.
6 Conclusions
In this paper, a comparison between clustering methods, K-means and Fuzzy c-means
was performed. K-means algorithm is preferred since the CD values are lower than
the Fuzzy c-means for all experimental clusters (2-6 clusters). Therefore, K-means
clustering is applied to cluster surface water quality of canals in Bangkok, Thailand.
The results indicate that five relatively cluster suitable for this sample site and
conclude to parameters.
Some of the differences corresponded to different site of canals, or water surfaces
separated by plugs or gates, that were held at different elevation and differences in
source water. The BOD and DO value parameters can show water quality for each
cluster. Better comprehension of the information may be helpful in further developing
the water quality management. As a result, Cluster analysis is found to be a useful and
efficient tool for stakeholders to manage natural resources, corresponding to the
government’s policy in water management.
References
1. Bengraine, K., Marhaba, T.F.: Using principle component analysis to monitor spatial and
temporal changes in water quality. Journal of Hazardous Materials 100(1-3), 179–195
(2003)
2. Guler, C., Thyne, G.D.: Hydrologic and geologic factors controlling surface and
groundwater chemistry in Indian wells-Owens Valley area, southeastern California.
Journal of Hydrology 285(1-4), 177–198 (2004)
3. MacQueen, J.B., Foster, I., Kesselman, C.: Some Methods for classification and Analysis
of Multivariate Observations: Procee C.: The Grid: Blueprint for a New Computing
Infrastructure. In: Proc. Fifth Berkeley Symp. on Math. Statist. and Prob., vol. 1, pp. 281–
297. Univ. of. Calif. Press, Berkeley (1967)
4. Dogan, E., Sengorur, B., Koklu, R.: Modeling biological oxygen demand of the Melen
River in Turkey using an artificial neural network technique. Journal of Environmental
Management (2008)
5. Areerachakul, S., Sanguansintukul, S.: Water Classification Using Neural Network: A
Case Study of Canals in Bangokok. In: The 4th International Conference for Internet
Technology and Secured Transactions (ICITST 2009), Thailand. IEEE Press, United
Kingdom (2009)
6. Anazawa, K., Ohmori, H.: The hydrochemistry of surface waters in Andesitic Volcanic
area, Norikura Volcano, central Japan. Chemosphere 59(5), 605–615 (2005)
7. Astel, A., Tsakovski, S., Simeonov, V., Reisenhofer, E., Piselli, S., Barbieri, P.: Multi-
variate classification and modeling in surface water pollution estimation. Analytical and
Bioanalytical Chemistry 390(5), 1283–1292 (2008)
8. Binbib, H., Fang, T., Guo, D.: Quality assessment and uncertainty handling in spatial data
mining. In: Proc. 12th Conference on Geoinformations, Sweden (2004)
9. Helena, B., Pardo, R., Vega, M., Barrado, E., Fernández, J.M., Fernández, L.: Temporal
evolution of groundwater composition in an alluvial aquifer (Pisuerga River, Spain) by
principal component analysis. Water Research 34(3), 807–816 (2000)
10. Kanungo, T., Mount, D., Netanyahu, N., Piatko, D.C., Silverman, R.: An Efficient K-
means Clustering Algorithm Analysis and Implementation. IEEE Transaction on Pattern
Analysis and Machine Intelligence 24(7) (July 2002)
11. Laaksoharju, M., Gurban, I., Skarman, C., Skarman, E.: Multivariate mixing and mass
balance (M3) calculations, a new tool for decoding hydrogeochemical information.
Applied Geochemistry 14(7), 861–871 (1999)
12. Liu, C.W., Lin, K.H., Kuo, Y.M.: Application of factor analysis in the assessment of
groundwater quality in a blackfoot disease area in Taiwan. The Science of the Total
Environment 313(1-3), 77–89 (2003)
13. Halkidi, M., Vazirgiannis, M.: Cluster validity assessment: finding the optimal partitioning
of a dataset. In: Proc. ICDM Conference, San Jose, Californaia, USA, pp. 187–194 (1996)
14. Simeonov, V., Stratis, J.A., Samara, C., Zachariadis, G., Voutsa, D., Anthemidis, A.,
Sofoniou, M., Kouimtzis, T.: Assessment of the surface water quality in Northern Greece.
Water Research 37(17), 4119–4124 (2003)
15. Simeonova, P., Simeonov, V., Andreev, G.: Water quality study of the Struma River
Basin, Bulgaria. Central European Journal of Chemistry 2, 121–136 (2003)
16. Singh, K.P., Malik, A., Sinha, S.: Water quality assessment and apportionment of pollution
sources of Gomti River (India) using multivariate statistical techniques: a case study.
Analytica Chimica Acta 538(1-2), 355–374 (2005)
Clustering Analysis of Water Quality for Canals in Bangkok, Thailand 227
17. Vega, M., Pardo, R., Barrado, E., Deban, L.: Assessment of seasonal and polluting effects
on the quality of river water byex ploratory data analysis. Water Research 32(12), 3581–
3592 (1998)
18. Bezdek, J.C.: Fuzzy partition and relations: an axiomatic basic for clustering. Fuzzy Sets
and Systems 1, 111–127 (1978)
19. Liou, S., Lo, S., Hu, C.: Application of two-stage fuzzy set theory to river quality
evaluation in Taiwan. Water Research 37, 1406–1416 (2003)
20. Ministry of Natural Resource and Environment, http://www.mnre.go.th
21. Department of Drainage and Sewerage Bangkok Metropolitan Administration,
http://dds.bangkok.go.th/wqm/thai/home.html
22. Department of Pollution Control Bangkok, http://www.pcd.go.th
23. Ministry of Public Health, http://eng.moph.go.th
24. Luke, B.: K-Means Clustering,
http://fconyx.ncifcrf.gov/~lukeb/kmeans.html
A Review of Routing Protocols for UWB MANETs
1 Introduction
The practical application of MANETs is sometimes preferable over infrastructure base
networks due to the unavailability or high cost of wired infrastructure. They can be
used effectively in military applications such as the battle field, where the existence of
wired infrastructure is impossible. They are also very useful in industrial or commer-
cial settings involving cooperative mobile data exchange, or when establishing com-
munications in a disaster recovery situation such as a fire or rescue operation.
MANETs function in the absence of a base station where nodes communicate di-
rectly without the need of a centralised control system. Wireless routing protocols en-
able nodes within the MANET to relay data packets around the network. If nodes are
within transmission range, they can communicate directly, or if they are away from
each other, intermediate nodes are required to establish a multihop route between
sender and receiver.
Now it is the time for UWB to be given more attention due to its great capability in
high data rate, large BW, and low power consumption. Currently, hosts are equipped
with UWB system in order to achieve short range communication and exploit the
D. Taniar et al. (Eds.): ICCSA 2010, Part III, LNCS 6018, pp. 228–245, 2010.
© Springer-Verlag Berlin Heidelberg 2010
A Review of Routing Protocols for UWB MANETs 229
previous mentioned advantages, besides UWB has the ability to measure distance be-
tween hosts accurately.
As a result, researchers have been encouraged to design new routing protocols for
UWB networks to make it suitable for hosts to exploit features of UWB, but still needs
more and more work to overcome drawbacks of network such as overhead, power con-
sumption, unreliable route, etc…
UWB [1], [2] is a radio technology proposed to be used in personal area networks
(PAN) and appeared in the IEEE 802.15.3a draft PAN standard. UWB systems con-
sume very low energy levels, and can be used in short-range, high BW (BW>500
MHz, 20% of the center frequency). UWB operates in range 3.1–10.6 GHz (the mini-
mum BW is 55MHz). The maximum allowable power spectral density is -41.3
dBm/MHz corresponding to average transmitted power of 0.5 mW when operating in
band range 3.1-10.6 GHz.
UWB is best used for ad-hoc and sensor networks, it is used as part of location sys-
tems and real time location systems such as hospitals and healthcare, short broadcast
time, “see-through-the-wall” precision radar imaging technology, and replacing cables
between portable multimedia Consumer Electronics (CE) devices and in next-
generation Bluetooth Technology devices [3].
It provides relatively long radio range which is around 150 meters indoor and 1 km
outdoor and high data rate in excess of 100 Mbps with bit rates 55, 110 & 200 Mbps.
UWB uses pulse coded information with sharp carrier pulses (pulse-based systems),
the signal consists of a train of pulses, one single pulse may be used to represent
one bit.
There is duration between single pulses to avoid both pulse overlapping and ambi-
guity between pulse positions [3]. Each pulse occupies the entire UWB bandwidth,
thus resist the multipath fading. UWB systems are carrierless systems based on pulses
of very short duration transmission [4].
Generically, wireless routing protocols in MANETs can be classified into topology
and position based protocols. These are presented in sections II and III, and of particu-
lar relevance to UWB systems, section IV then considers power aware routing proto-
cols and then section V highlights the issues associated with routing protocols for
UWB applications.
(DSDV) [6], Cluster-head Gateway Switch Routing (CGSR) [7], Wireless Routing
Protocol (WRP) [8], Optimized Link State Routing (OLSR) [9], and Hierarchical State
Routing (HSR) [10]. The first three protocols are discussed bellow:
a) Destination-Sequenced Distance Vector (DSDV)
DSDV [6] is a flat proactive routing protocol which guarantees loop freedom, single
shortest path to destination. Each node forwards its information among the network;
neighbor nodes (intermediate nodes) receive this information and update their routing
tables. Thus, all nodes have routes to all reachable nodes in the network in advance.
DSDV uses a sequence number to prevent loops; here a node will not forward in-
formation with the same sequence number twice. The higher the sequence number the
newer the route; nodes update their routing tables whenever they receive route infor-
mation that contains a higher sequence number than the stored one in routing table, or
a route with same sequence number but shorter than the stored one.
When a source node wants to send data to destination, it checks its routing table and
selects the shortest path to the destination. DSDV uses two types of packets; Full
Dump packets which carry all available routing information and can require multiple
network protocol data units (NPDUs), and a route Broadcast packet which contains the
address of destination, number of hops, and the sequence number of the information
received regarding the destination.
DSDV can not work in scalable network, especially when nodes move; this causes a
high control packet update which increases network overhead.
b) Cluster-head Gateway Switch Routing (CGSR)
CGSR [7] is a hierarchical proactive routing protocol. Nodes are arranged in clusters, a
cluster head (CH) is elected in each cluster by a distributed algorithm. Transmission is
then achieved by CH inside clusters, which reduces overhead in network, but nodes
within cluster are busy in selecting CH rather than packet relaying; which affects pro-
tocol performance. While transmission between clusters is achieved by gateway nodes
that exist between the transmission ranges of two or more CHs as shown in Fig. 1.
Gateway
Cluster Head
Member node
Each node maintains two tables; a cluster member table for storing destination CH
of every node in the network, this table is broadcasted using the DSDV algorithm, and
another routing table for storing neighbor nodes.
c) Wireless Routing Protocol (WRP)
WRP [8] is a flat proactive routing protocol which guarantees loop freedom by consid-
ering predecessor information. It uses four routing tables; distance table, routing table,
link-cost table, and Message Retransmission List (MRL), table and thus it consumes
memory.
Each record in MRL contains a sequence number, retransmission counter, acknowl-
edgement required flag vector, and a list of update message that contains the destina-
tion, the distance to destination, and the predecessor of destination. An update message
is sent in two cases; after processing updates from neighbours or when detecting a
change in a link to neighbour.
WRP checks for nodes connectivity by sending hello messages. When a node re-
ceives hello message, it adds the node to its routing table, and then it sends a copy of
its routing table to the node. This process causes overhead and consumes power.
d) Summary
Proactive routing protocols, in general, decrease end to end delay, but provide the
shortest path. On the other hand, they increase network overhead, consume memory
and power, and so, they are only suitable to small size networks.
B. Reactive Routing Protocols
Reactive routing protocols maintain route information on demand, i.e. when source
node wants to send message to destination, it initiates route request (RREQ) to find
route to destination. When a route fails, a route maintenance process is launched to
repair the failed route.
Ad hoc On-Demand Distance Vector (AODV) [11], Dynamic Source Routing
(DSR) [12], Temporally Ordered Routing Algorithm (TORA) [8], and Associativity-
Based Routing (ABR) [8] are examples of reactive strategy.
More discussion about AODV and DSR and TORA is bellow:
a) Ad hoc On-Demand Distance Vector (AODV)
AODV [9] belongs to the class of Distance Vector (DV) routing protocols. In a DV,
every node knows about its neighbours and the costs incurred in order to reach them
using the Bellman-Ford algorithm [10].
AODV is a reactive shortest single path wireless routing protocol based on DSDV
protocol. When a source wants to send a message to a destination, it checks its routing
table, if there is a valid route to destination, it starts sending packets immediately. If
not, it broadcasts RREQ to all neighbor nodes. It should be noted that RREQ contains
the fields: hop count, source and destination sequence numbers, destination and source
addresses, RREQ ID, and other pre-determined fields. When an intermediate node re-
ceives a RREQ, it checks its routing table for a path to the destination, if it exists, it
unicasts a route reply (RREP) to the source, otherwise, it increases the hop count by
one, and adds its ID to the RREQ and then re-broadcasts it to its neighbours; the proc-
ess continues until the RREQ reaches its destination. The destination then selects the
first coming RREQ and unicasts RREP using the reverse path to source node. When
232 Y.H. Jazyah and M. Hope
the source receives several RREPs, it selects the route of the highest sequence number
and minimum hop count, and then it establishes the route and starts sending packets.
To guarantee loop freedom, the source node uses a sequence number and includes it
in RREQ. When a node receives a control message (RREQ, RREP, or RERR), it
checks its routing table for an entry to the specified destination, if there is no entry in
the routing table about the destination, it creates a new one. If there is an entry in the
routing table, the route is only updated if the new sequence number is either higher
than the destination sequence number in the routing table, the sequence numbers are
equal, but the hop count plus one is smaller than the existing hop count in the routing
table, or the sequence number is unknown. In addition, the source uses a time to live
(TTL) count to limit the flooding of RREQ packets and control the overhead associ-
ated with the network. Finally, a HELLO message is broadcasted periodically to in-
form neighbor nodes about node existence. When an active node (node on the active
route) detects a route failure (the neighbor node is unreachable; i.e. HELLO is not be-
ing received), it sends a route error (RERR) packet to the source node, which in turn,
initiates a new RREQ.
Overhead is a major drawback of AODV because of the flooding of control mes-
sages (HELLO and RREQ packets). However, the overhead is low when compared to
proactive routing protocols, and it also needs less memory, on the other hand it only
stores the routes on demand so, it is preferable in application to large scale networks.
b) Dynamic Source Routing (DSR)
DSR [13] is a reactive routing protocol; when a source node wants to send data to a
destination address; it initiates and broadcasts a RREQ that carries the full address
(every hop). When intermediate receives the RREQ, it checks its routing table (nodes
may store many routes to destination), if there is a valid path to destination, it sends
RREP to source, or it forwards the packet after adding its address if its address is not
existed in route record in order to avoid loop, and so on until RREQ reaches to destina-
tion which sends RREP to source using the reverse path.
A link failure is detected when no acknowledgement received by a node. RERR
message is sent to source, which in turn, looks for a valid route in its cache, if no valid
route is existed, source initiates RREQ again.
DSR does not use hello message, and so, it reduces the overhead and saves band-
width, which makes it suitable for scalable network. On the other hand, packet size
increases when route is long and network is large since every node adds its address to
route record of RREQ.
c) Temporally Ordered Routing Algorithm (TORA)
TORA is a distributed and reactive routing protocol. It is suitable for highly dynamic
topology network, it provides acyclic multipath but it suffers from long delay and does
not support shortest path calculation. It is based on directed acyclic graph to maintain
loop freedom.
When source want to forward packets to destination, it uses a height metric to estab-
lish direct acyclic graph and each node maintains information about one hop neighbors
in order to localize control message to a very small set of nodes near the occurrence to
topology change. When a link fails, a node generates a reference level which propa-
gates by neighbor nodes.
A Review of Routing Protocols for UWB MANETs 233
d) Summary
Reactive routing protocols reduce network overhead, consume less bandwidth, and
need less memory when compared to proactive protocols because they establish the
route on demand. But they need a longer time than proactive protocols to establish the
route between source and destination, and the also need to route provide maintenance
because there are no ready available backup routes in general. In general, reactive rout-
ing protocols are suitable to large scale network compared to proactive ones.
Topology based routing protocols use flat addressing scheme as mentioned in the
previous examples except CGSR which uses hierarchical addressing.
C. Hybrid Routing Protocols
Most networks are partitioned into zones (or zone-based) or nodes are grouped into
trees or clusters. Hybrid routing protocols merge proactive and reactive strategies
whereas proactive strategy is applied within zone (or intra zone), while a reactive strat-
egy is applied outside zone (or inter zone). The following are examples of hybrid rout-
ing protocols:
a) Zone Routing Protocol (ZRP)
In ZRP [14], network is divided into routing zones, and each node belongs to a zone
based on its geographical location. ZRP reduces the overhead of proactive protocols and
the delay associated with reactive protocols, and so, it is suitable for large networks.
The network is divided in non overlapping zones without zone head which reduces
overhead. Zone size depends on node mobility, network density, transmission power
and propagation characteristics.
If destination is moving within the zone, no need to search in other locations since
node ID and zone ID of destination are required for routing, and so, ZHLS may scale
well to large networks; nevertheless, all nodes must have a pre-programmed static zone
map, while geographical boundary of network is dynamic.
c) Summary
Hybrid routing protocols merge both proactive and reactive approaches together by
utilizing advantages of both of them. They reduce the overhead of proactive protocols
by applying reactive routing between zones and limiting flooding inside zones, and the
delay associated with reactive protocols by applying proactive routing inside zones,
thus it is suitable for large scale networks.
3 Position-Based Routing
Position based routing focuses on two issues: a location service to determine destina-
tion's position and forwarding strategy that can be greedy forwarding, restricted direc-
tional flooding, and hierarchical routing.
D. Location Services
Location service is used by sender node to learn the position information of a certain
node in order to include it in the forwarded packet if it does not know the position of
communication target. Each node registers its current position to a location service,
when a node needs the position information of another one; it sends a request to loca-
tion service. For example, Quorum-Based Location Service [5], Home-Zone [5],
Distance Routing Effect Algorithm for Mobility (DREAM) [15], and Grid Location
Service (GLS) [5], [16]. The last two location services are discussed bellow:
a) Distance Routing Effect Algorithm for Mobility (DREAM)
Each node maintains a position database that stores position information of each other
nodes that are part of the network (all-for-all), where packets are regularly flooded to
update position information [5].
DREAM [15] combines proactive and reactive approaches. Source node has to have
positions of it, destination, one hop neighbours, and the maximum speed of destination.
Source node uses this information to determine the direction towards destination in or-
der to send packets by finding one hop neighbor within the range between source and
destination according to a certain angle determined by speed and position of destination.
b) Grid Location Service (GLS)
GLS [5], [16] divides network into a hierarchy of squares. In each hierarchy, n-order
squares contain exactly four (n – 1)-order squares, forming a "quadtree". Each mobile
node periodically updates a small set of other nodes with their current locations by
maintaining a table of all other nodes within the local first-order square. (Classified as
all-for-some)
A Review of Routing Protocols for UWB MANETs 235
GLS uses geographical forwarding. Source must know the geographical positions of
any destination, and must label packets for that destination with its position, while in-
termediate node only needs to know its position and the positions of nearby nodes.
Geographical forwarding is similar to Cartesian routing. Each node broadcast a
HELLO message that contains its position and velocity; where neighbours within
transmission range store this information in a table. Node sends packet to the closest
neighbor of destination after checking the neighbor table. When a node does not find
closest node other than itself, it sends error indicating a hole in geographic distribution
of nodes. Greedy Perimeter Stateless Routing (GPSR) is a solution for the problem of
hole.
GLS is based on the idea that a node maintains its current location in a number of
location servers (each node acts as a location server on behalf of some other nodes)
distributed through over network, but they are no elected leaders. GLS makes balance
when location servers are selected and dense of network is considered.
c) Summary
Location service is used by source node to determine position of destination. Location
services are classified according to “how many servers maintain position information
about mobile nodes “into all for all, all for some, some for all, and some for some. For
example, when all nodes within the network have information about all nodes in the
network, then the location service is classified as “all for all”.
E. Position-Based Routing Protocols
Position based routing protocols use position information gathered by GPS system.
Each node in the wireless network knows its location, destination location, and other
geographical information in order to forward packets in geographical direction instead
of flooding as in topology based protocols.
Mainly, there are three forwarding strategies: First, greedy packet forwarding;
whereas the sender of packet includes the approximate position of recipient in the
packet, intermediate node forwards the packet to a neighbor which is lying in the gen-
eral direction of the recipient. Second, restricted directional flooding; the packet is
flooded in one-hop direction towards the destination where it is expected to be located;
this process is repeated until the packet reaches the destination. The direction is deter-
mined by calculating the expected region that covers destination. Third, hierarchical
Routing; where the complexity that each node has to handle can be reduced massively
by establishing some form of hierarchy.
Many position based routing protocols have been designed such as, Location Aided
Routing (LAR) [17], GRID [18], Greedy Perimeter Stateless Routing (GPSR) [19],
Adaptive Location Aided Roaming from Mines Protocol (ALARM) [20], Location-
Aware routing Protocol with Dynamic Adaptive of Request Zone (LARDAR)
[21],Compass [22],Angle Routing PROTOCOL (ARP) [23],Position-based Selective
Flooding (PSF) [24], Octopus [25], Improved Location-Aided Routing (ILAR) [26] ,
and Novel Routing Power Balance (NRMPB) [27] . The following is a discussion
about position based routing protocols.
a) Location Aided Routing Protocol (LAR)
LAR [17] is an on-demand routing protocol that uses the modified Dijkstra's Algorithm
to find the shortest path.
236 Y.H. Jazyah and M. Hope
Destination lies in the centre of a circular region of certain radius at a certain time
known as Expected Zone, as shown in Fig. 3, which indicates which zone of the net-
work should be reached by RREQ. GPS enables terminals to know its own position
and speed, while dissemination is performed by piggybacking location information in
all routing packets.
“Piggybacking is a bi-directional data transmission technique in the network layer.
It makes the most of the sent data frames from receiver to emitter, and adds the con-
firmation that the data frame sent by the sender was received successfully acknowl-
edge.” [28].
There are two proposed schemes of LAR:
First, LAR1 defines request zone that includes sender and receiver on opposite cor-
ner of a rectangle. Rectangle dimensions are estimated according to the receiver aver-
age speed at a certain time. Only nodes within the zone respond to the RREQ of
sender. LAR1 reduces network overhead, but causes delay when route is found.
Different shapes can be used other than rectangle such as circle, cone, and bar.
LAR2 is proposed based on the calculated distance between source position and the
estimated position of destination. When node receives RREQ, it calculates the distance
toward destination, if the distance is less than the distance from the previous sender
node to destination; it forwards the packet as shown in fig. 4. In this scheme, interme-
diate receiving node may be the closest node to destination, and so the algorithm
reaches to a dead-end.
b) GRID
GRID [18] is a reactive routing protocol which exploits location information. GRID
has strong maintenance capability; where intermediate node that attends to move per-
forms handoff operation to keep route alive.
Expected Zone
Destination
Source
Geographical area is divided into square grids; the nearest node to the centre of grid
is elected as a leader (gateway) to maintain routing. If leader leaves a grid, it broad-
casts routing table to the new leader. Position information of source and destination are
used to confine.
The search area, leader is responsible for searching procedure and setting up routing
table in grid by a grid manner; routing in GRID is performed grid by grid through
grids' leaders.
Source node calculates the searching range depending on its location and the loca-
tion, speed, and direction of destination in order to eliminate the effect of collection of
redundant rebroadcast, heavy contention, and collisions (problem of broadcast storm).
c) Greedy Perimeter Stateless Routing (GPSR)
In GPSR [19], the closest intermediate node to destination is selected as next hop, but
GPSR does not guarantee the existence of path, and so perimeter forwarding selects the
neighbor which is farther than itself from the destination in order to solve the dead-end
caused by greedy forwarding, GPSR switches between these two strategies. A modifi-
cation to GPSR is made by using flooding to solve the problem of dead-end.
Each node needs information about next hop only, which is exchanged by means of
periodic beacons broadcasted by each node.
d) Octopus
Octopus [25] is a fault-tolerant and efficient position-based routing protocol. Fault-
tolerance is achieved by employing redundancy (storing the location of each node at
many other nodes) and by keeping frequently refreshed soft state; nevertheless, it
achieves a low location update overhead by employing a novel aggregation technique.
Octopus is a typical routing scheme, assumes that each node gets its own location
information using like GPS. It divides the network into grid by horizontal and vertical
strips, each node maintains a table for neighbors (updated by hello message) and a ta-
ble for strips. Octopus forward RREQ in north and south directions to find destination,
if this process fails, it sends the RREQ to the horizontal direction. Octopus uses
Greedy Forwarding to forward data packets.
e) Improved Location-Aided Routing (ILAR)
ILAR [26] improves location-aided routing (LAR); when intermediate node receives
RREQ, it checks weather it is within the requested zone, if it is, it will send RREQ revise
(RREQ_R) to the transmitting node, RREQ_R includes the VDIST (vertical distance, it
is the distance from node to the line connecting source and destination; base line) as
shown in fig. 5. The node which receives RREQ_R will select the next hop which has the
smallest VDIST and has a distance to source larger than to the previous node.
This technique has less packets and better performance than LAR.
f) Novel Routing Power Balance (NRMPB)
NRMPB [27] resolves the problems of ILAR which are: the movement of nodes near
border of coverage broadcast which cause breakage of route, the other problem is se-
lecting the next hop node; NPMPB comply position information and power saving to
balance the energy of each nodes and prolong lifetime of network.
In ILAR protocol, the intermediate node which has the smallest vertical distance is
selected as next hop, but the selected node could be located near the boundary of
238 Y.H. Jazyah and M. Hope
Baseline
Source
g) Summary
Position based routing protocols eliminates overhead of reactive protocols by directing
flood of route request towards the expected location of destination, which leads to bet-
ter performance in all aspects; power consumption, stability, reliability, maintenance,
and bandwidth utilization. These specifications make position-based routing widely
trusted to large scale networks.
4 Power-Aware Routing
Power metrics are important issues when routing protocols are designed since mobile
nodes have limited resource of power which is represented by local batteries.
A Review of Routing Protocols for UWB MANETs 239
Generally, energy efficient issues are: 1) Sleeping mode; whereas nodes turn them-
selves off when not actively transmitting or receiving (sleep/awake). 2) Transmission
Power Control; which aims to minimize the power per bit required to transmit packet
from source to destination, this technique always selects the least-power cost routes; as
a result, nodes along those routes tend to die soon because of battery energy exhaus-
tion. 3) Power-aware route selection that find route at certain route discovery time such
that a cost function is minimized. Cost function depends on energy consumed per
packet, time to network partition, variance in node power levels, cost per packet, and
node cost metrics. And 4 ) Broadcast control.
Examples of power-aware routing protocols are: Power-Aware Multiple Access
Protocol (PAMAS) [29], Self-Healing-AODV (SHAODV) [30], Energy-Efficient Ad
hoc Routing (DEAR) [31], Energy-Aware Ad hoc On-Demand Distance Vector
(EAAODV) [32], Active Route Maintenance Protocol (ARMP) [33], and Distributed
Power Control (DPC) [34]. The following are more details about routing protocols
related to power.
a) Power-Aware Multiple Access Protocol (PAMAS)
PAMAS [29] is a shortest-cost routing algorithm, power-Aware Multiple Access pro-
tocol with Signalling, it uses a new power-aware metrics for determining routes in
wireless ad hoc networks based on battery power consumption of node.
Nodes’ radios are turned off (sleep mode) when they can not transmit or receive
packets during transmission period of other nodes; which in turn saves their power.
b) Self-Healing-AODV (SHAODV)
SHAODV [30] overcomes the problem of edge effect and route break without causing
heavy overhead. The path can be established for long lifetime by connection admission
control of route request. The unstable route can be healed without route break, flood
over the network, and overhead by link state inspection and active route local repair.
Signal to Noise Ratio (SNR) is measured and compared to a threshold in order to
overcome the overhead; it is calculated according to the received power which is based
on transmitted power, height and gain of antenna at both sides of transmitter and re-
ceiver, and distance between them.
Movable node requests the source to heal the unstable path before it moves and
causes break in the route. Process of reduction the repaired path is initiated after that by
broadcasting RREQ for reduction along the repaired path and accepting the first re-
ceived RREQ packet with link state value (SNR) greater than a threshold.
c) Energy-Efficient Ad hoc Routing (DEAR)
DEAR [31] controls the rebroadcast time of RREQ packets to prolong network lifetime
by balancing between minimum transmission energy consumption and fair node en-
ergy consumption in a distributed manner without additional control packets. It also
improves data packet delivery ratio.
The physical layer can save energy by adapting transmission power according to the
distance between nodes. While at the data link layer, energy conservation can be
achieved by sleep mode operation.
Network lifetime is defined as the time when a node runs out of its own battery
power for the first time.
240 Y.H. Jazyah and M. Hope
DEAR utilizes only the RREQ message to distribute the average residual battery
level of the entire network. Intermediate nodes control the rebroadcast time of the
RREQ packet whereas retransmission time is proportional to the ratio of average resid-
ual battery power of the entire network to its own residual battery power, and so, nodes
with larger battery energy will rebroadcast RREQ packets earlier.
Residual battery level of the entire network depends on the average residual power
of route, which in turn depends on the residual power of node and hop count. DEAR
control the rebroadcast time; if the residual battery power of node is smaller than the
average network residual power, the retransmission time will be longer and vice versa.
DEAR compromises between the min-hop path and the fair energy consumption path.
d) Energy-Aware Ad hoc On-Demand Distance Vector (EAAODV)
EAAODV [32] combines the on-demand routing capability of the AODV routing pro-
tocol with a distributed topology discovery mechanism – mobile agents, thus reaching
the energy balance condition. EAAODV eliminates the problem of low efficiency and
link breaking in routing due to the energy exhaustion of some nodes.
"A mobile agent is a software program that can suspend its execution on the host
computer, transfer itself to another agent-enabled host through the network, and
resume its execution on the new host." The key features of mobile agents which distin-
guish them from traditional distributed programming are: mobility; network aware-
ness; communication; intelligence; reactivity; autonomous; goal-oriented; temporally
continuous; learning; and flexible.
When intermediate node receives a RREQ, it decided to broadcast it according to its
residual energy, if it is under a threshold level, the node terminates the RREQ, other-
wise it broadcasts the RREQ. During the search process, the node checks its routing
table, if the table is fresh; the residual energy information of the next-hop node indi-
cated by the table is acquired by mobile agents, if the energy is above the threshold, the
routing table is valid and it will be used, otherwise, searching for the routing will be
finished, if the routing table is old, nodes will broadcast RREQ to their neighbors and
update the information of the table.
EAAODV is able to achieve longer surviving network duration and a larger
throughput compared to AODV.
e) Summary
Methods that concern power control issue can be classified into three categories: meth-
ods find the optimal transmitted power to control connectivity properties of network,
power aware routing, and algorithms based on shortest path rather than hop count. In
general, route maintenance is detected according to a threshold value which is com-
pared to a certain power metric.
data which depends upon the requested data rate, both represent power consumption,
and distance between two nodes. Besides Signal to Noise Ratio (SNR), Symbol Error
Probability (SEP), and Packet Error Probability (PEP) are taken into account when
multi-user interference (MUI) is considered.
The above technique is applied to DSR routing protocol. Whereas each intermedi-
ate node adds the cost of previous hop to the path cost. Destination selects the path
according to the cost information included in RREQ. Also it is applied to LAR rout-
ing protocol.
Adoption of position-aware routing strategies significantly improves the efficiency
in power management for UWB networks, for example, by optimizing the path selec-
tion procedure in DSR and trade-off between performance and power efficiency can
be achieved by adopting different shapes for the Request Zone of LAR.
Paper [39] proposes power-efficient routing strategy based on the ranging capabili-
ties offered by UWB.
The cost function of link is considered as in [38]. Minimum Cost strategy leads to
more robust system in a high-interference environment; the Minimum Cost strategy is
not effective in the case of low network connectivity since it increases the average
number of hops with no advantage in system performance. When network connec-
tivity grows, the Minimum Cost effectively increases the percentage of found connec-
tions by reducing the effect of multi-user interference.
Paper [41] proposes a novel packet routing scheme for UWB ad-hoc network. This
approach utilizes the distance measurement between nodes that UWB provides,
adopting the multihop transmission, determined spreading factor and the signal to
interference and noise ratio (SINR).
Spreading Factor (SF) is the ratio of the chips to baseband information rate.
Spreading factors vary from 4 to 512 in FDD UMTS. Spreading factor in dBs indi-
cates the process gain, whereas the lower the spreading factor the higher the data rate.
Source node searches for node within its transmission range provided that the dis-
tance to destination is greater than the distance between the previous node and desti-
nation, the process is repeated by each node until destination is reached.
e) Summary
UWB is a new technology that is in MANET to transmit data packets in high speed.
Routing protocols for UWB MANET have been designed to use the high performance
and capabilities of UWB; such routing protocols are designed as master slave in general
to overcome network overhead and reduce power consumption, other routing techniques
consider a cost function that is accumulated until the RREQ reaches its destination,
whereas the destination selects the route of cost function that satisfy the criteria.
Many routing protocols have been designed for UWB MANET, nevertheless the
field needs more and more effort by researcher to design new protocols that can use
the capabilities of UWB.
6 Conclusion
This work summarises wireless routing protocols for MANETs which can be classified
into topology based and position based protocols. Topology based protocols are either
A Review of Routing Protocols for UWB MANETs 243
proactive or reactive, while hybrid strategies merge proactive and reactive approaches
in order to enhance the drawback of proactive protocols which utilize the advantages of
reactive ones.
Position based strategies are based on positional information gathered by GPS sys-
tems. They eliminate the overhead caused by topology based strategies by directing
flooding towards the destination.
Power metrics are included in both topology and position based routing in order to
enhance the performance of routing protocols in terms of power saving. They also take
part in route maintenance as detector of route failure before it occurs, or as a metric to
select the next hop.
UWB is a radio technology proposed to use in PANs and has appeared in the IEEE
802.15.3a draft PAN standard. UWB systems consume very low energy levels, and
they are applicable to short-range high data rate systems which make it particularly
useful to MANETs and sensor networks. Many conventional routing protocols have
been designed for MANETs. However routing strategies for UWB networks need
more work but will most likely employ aspect of power aware routing and geographic
position.
References
[1] Xia, L., Lo, A., Niemegeers, I., Baugé, T.: An Ultra-Wide Band Based Ad Hoc Network-
ing Scheme for Personnel Tracking in Emergencies. In: This work was performed in the
IST Framework 6 EUROPCOM project
[2] Kuladinithi, K., Timm-Giel, A., Görg, C.: Mobile Ad Hoc Communications in AEC In-
dustry (August 2004), http://www.itcon.org/2004/22/
[3] Benedetto, M., Giancola, G., Domenicali, D.: Ultra Wide Band radio in Distributed
Wireless Networks. IEEE, Los Alamitos (2006)
[4] Project SPEARHUW, Enhanced Sensor and WLAN Networks using UWB Communica-
tions (April 2005)
[5] Mauve, M., Widmer, J., Hartenstein, H.: A Survey on Position-Based Routing in Mobile
Ad Hoc Networks. IEEE Networks 15(6), 30–39 (2001)
[6] Perkins, C.E., Bhagwat, P.: Highly Dynamic Destination-Sequenced Distance-Vector
Routing (DSDV) for Mobile Computers. ACM SIGCOMM Computer Communication
Review 24(4), 234–244 (1994)
[7] Abolhasan, M., Wysocki, T., Dutkiewicz, E.: A review of routing protocols for mobile ad
hoc networks. Ad Hoc Networks 2, 1–22 (2003)
[8] Royer, E.M., Santa, B., Toh, C.: A Review of Current Routing Protocols for Ad hoc Mo-
bile Wireless Networks. IEEE Wireless Communications 6(2), 46–55 (1999)
[9] Clausen, T., Jacquet, P., Laouiti, A., Muhlethaler, P., Qayyum, A., Viennot, L.: Opti-
mized Link State Routing Protocol. In: Proceedings of IEEE INMIC (2001)
[10] Kilinkaridis, T.: Routing Protocols for Wireless Ad Hoc Networks Hierarchical routing
protocols
[11] Perkins, C., Royer, E.B., Das, S.: Ad hoc On-Demand Distance Vector (AODV) Routing
(2003), http://www.ietf.org/rfc/rfc3561.txt
[12] Baumann, R.: AODV, Ad hoc On Demand Distance Vector Routing Protocol. AODV
Presentation at ETH, Zürich (2002)
244 Y.H. Jazyah and M. Hope
[13] Johnson, D.B., Maltz, D.A., Broch, J.: DSR: The Dynamic Source Routing Protocol for
Multi-Hop Wireless Ad Hoc Networks. In: Ad Hoc Networking, ch. 5, pp. 139–172
(2001)
[14] Beijar, N.: Zone Routing Protocol (ZRP), http://www.netlab.tkk.fi/
opetus/s38030/k02/Papers/08-Nicklas.pdf
[15] Nardis, L.D., Giancola, G., Benedetto, M.: Power-aware design of MAC and routing for
UWB networks. In: Globecom Workshops. IEEE Communications Society, Los Alami-
tos (2004)
[16] Jinyang, L., Jannotti, J., Douglas, S.J., Karger, D.R., Morris, R.: A scalable location ser-
vice for geographic ad hoc routing. In: Proc. MOBICOM, pp. 120–130 (2000)
[17] Ko, Y.B., Vaidya, N.H.: Location-Aided Routing (LAR) in mobile ad hoc networks.
Wireless Networks 6, 307–321 (2000)
[18] Liao, W.H., Tseng, Y.C., Sheu, J.P.: GRID: A fully location-aware routing protocols for
mobile ad hoc networks. Telecommunication Systems 18(1-3), 37–60 (2001)
[19] Karp, B., Kung, H.T.: GPSR: Greedy Perimeter Stateless Routing for Wireless Networks.
In: MobiCom (2000)
[20] Boleng, J., Camp, T.: Adaptive Location Aided Mobile Ad Hoc Network Routing. In:
IEEE International Conference on Performance, Computing, and Communications, pp.
423–432 (2004)
[21] Shih, T.F., Yen, H.C.: Location-aware routing protocol with dynamic adaptation of
request zone for mobile ad hoc networks, October 9. Springer, Heidelberg (2006)
[22] Kranakis, E., Singh, H., Urrutia, J.: Compass Routing on Geometric Networks. In: Proc.
11th Canadian Conf. Computational Geometry (August 1999)
[23] Banka, R.K., Xue, G.: Angle Routing Protocol: Location Aided routing for Mobile Ad-
hoc Networks Using Dynamic Angle Selection. IEEE, Los Alamitos (2002)
[24] Abolhasan, M., Wysocki, T.: GPS-based Route Discovery Algorithms for On-demand
Routing Protocols. In: MANETs (January 2004)
[25] Nakagawa, H., Ohta, T., Ishida, K., Kakuda, Y.: A Hybrid Routing with Location Infor-
mation for Mobile Ad Hoc Networks. In: Eighth International Symposium on Autono-
mous Decentralized Systems (ISADS 2007). IEEE, Los Alamitos (2007)
[26] Wang, E.N.C., Wang, S.M.: An Efficient Location-Aided Routing Protocol for Mobile
Ad Hoc Networks. In: Proceedings of the 2005 11th International Conference on Parallel
and Distributed Systems (ICPADS 2005). IEEE, Los Alamitos (2005)
[27] Li, J.P., Wu, L.Y., Lin, S.Y., Wu, G.M.: A Novel Routing Protocol Integrate Power Bal-
ance for Ad Hoc Networks. In: Third International Conference on Networking and Ser-
vices (ICNS 2007). IEEE, Los Alamitos (2007)
[28] Ucan, E., Thompson, N., Gupta, I.: A PiggyBacking Approach to Reduce Overhead in
Sensor Network Gossiping. In: MidSens 2007, Newport Beach, USA, November 26-30,
ACM, New York (2007); Copyright 2007, 978-1-59593-929-6/07/11
[29] Singh, S., Mghavendra, M.W.C.S.: Power-Aware Routing in Mobile Ad Hoc Networks.
In: Proceedings of Fourth Annual ACM IEEE International Conference on Mobile Com-
puting and Networking, Dallas, TX, October 1998, pp. 181–190 (1998)
[30] Feng, M., Cheng, S., Zhang, X., Ding, W.: A Self-Healing Routing Scheme Based on
AODV in Ad hoc Networks. In: The Fourth International Conference on Computer and
Information Technology CIT 2004, September 14-16, pp. 616–620 (2004)
[31] Gil, H., Yoo, J., Lee, J.: An On-demand Energy-efficient Routing Algorithm for Wireless
Ad hoc Networks. In: Chung, C.-W., et al. (eds.) HSI 2003. LNCS, vol. 2713, pp. 302–
311. Springer, Heidelberg (2003)
A Review of Routing Protocols for UWB MANETs 245
[32] Chenchen, Z., Zhen, Y.: A New EAAODV Routing Protocol based on Mobile Agent.
In: International Conference on Systems and Networks Communications, ICSNC 2006
(October 2006)
[33] Chang, C.Y., Chang, C.T., Tu, S.C., Hsieh, T.T.: Active Route-Maintenance Protocol for
Signal-Based Communication Path in Ad-Hoc Networks. In: Proceedings of Ninth IEEE
International Conference on Networks, October 10-12, pp. 25–30 (2001)
[34] Bergamo, P., Giovanardi, A., Travasoni, A., Maniezzo, D., Mazzini, G., Zorzi, M.: Dis-
tributed Power Control for Energy Efficient Routing in Ad Hoc Networks. ACMKluwer
WINET 10(1) (January 2004)
[35] Gong, M.X., Midkiff, S.F., Buehrer, R.M.: A New Piconet Formation Algorithm for
UWB Ad Hoc Networks. In: IEEE Conference on Ultra Wideband Systems and Tech-
nologies, November 16-19, pp. 180–184 (2003)
[36] Huang, Y., Zeng, W.: A reliable Routing Protocol for Multi-hop Ad hoc Networks. In:
International Conference on Wireless Communications, Networking and Mobile Com-
puting, WiCom, September 21-25, pp. 1656–1659 (2007)
[37] Nardis, L., Baldi, P., Benedetto, M.G.: UWB Ad-Hoc Networks. In: IEEE Conference on
Ultra Wideband Systems and Technologies, Baltimore, Maryland, USA, May 20-23, pp.
219–223 (2002)
[38] Nardis, L.D., Giancola, G., Benedetto, M.G.D.: A position based routing strategy for
UWB networks. IEEE, Los Alamitos (2003)
[39] Nardis, L.D., Giancola, G., Benedetto, M.G.D.: A power-efficient routing metric for
UWB wireless mobile networks. In: Procs. of Vehicular Technology Conference, IEEE
VTS Fall VTC 2003, October 2003, pp. 3105–3109 (2003)
[40] Horie, W., Sanada, Y.: Novel Routing Schemes Based on Location Information for UWB
Ad-Hoc Networks. Wiley Periodicals, Inc., Chichester (2004)
[41] Home, W., Sanada, Y.: Novel Packet Routing Scheme based on Location Information for
UWB Ad-hoc Network. IEEE, Los Alamitos (2003)
[42] Gong, M.X., Midkiff, S.F., Buehrer, R.M.: A Self-Organized Clustering Algorithm for
UWB Ad Hoc Networks. In: WCNC 2004. IEEE Communications Society, Los Alamitos
(2004)
An Efficient and Reliable Routing Protocol for Wireless
Mesh Networks
Jaydip Sen
1 Introduction
Wireless mesh networking has emerged as a promising concept to meet the challenges
in next-generation wireless networks such as providing flexible, adaptive, and recon-
figurable architecture while offering cost-effective solutions to service providers.
WMNs are multi-hop wireless networks formed by mesh routers (which form a wire-
less mesh backbone) and mesh clients. The mesh routers provide a rich radio mesh
connectivity which significantly reduces the up-front deployment cost of the network.
Mesh routers are typically stationary and do not have power constraints. However, the
clients are mobile and energy-constrained. Some mesh routers are designated as
gateway routers which are connected to the Internet through a wired backbone. A
gateway router provides access to conventional clients and interconnects ad hoc, sen-
sor, cellular, and other networks to the Internet. A mesh network can provide multi-
hop communication paths between wireless clients, thereby serving as a community
network, or can provide multi-hop paths between the client and the gateway router,
therby providing broadband Internet access to the clients.
D. Taniar et al. (Eds.): ICCSA 2010, Part III, LNCS 6018, pp. 246–257, 2010.
© Springer-Verlag Berlin Heidelberg 2010
An Efficient and Reliable Routing Protocol for Wireless Mesh Networks 247
2 Related Work
Although significant work has been done on routing in wireless networks, very little
work has been done for WMNs. Most of the routing protocols for MANETs such as
AODV and DSR use hop-count as the routing metric. However, this is not well-suited
for WMNs. The basic idea in minimizing the hop-count is that it reduces delay and
maximizes the throughput. But the assumption here is that the links in the path are
either perfect or do not work at all, and all links are of equal bandwidth. A routing
scheme that uses the hop-count metric does not take link quality into consideration. A
minimum hop-count path has, on the average, longer links between the nodes present
in the path compared to a higher hop-count path. This reduces the signal strength
received by the nodes in that path and thereby increases the loss ratio at each link [3].
Hence, it is always possible that a two-hop path with a good link quality provides
higher throughput than a one-hop path with a poor link quality. Moreover, wireless
links usually have asymmetric loss rate [4]. Hence, new routing metrics based on link
248 J. Sen
quality are proposed such expected transmission count (ETX), per-hop round-trip
time (RTT), and per-hop packet pair. Different approaches have been suggested by
researchers for designing routing protocols for WMNs. In [5], a QoS routing over
OLSR protocol has been proposed that takes into account metrics such as bandwidth
and delay where the source node proactively changes a flow’s next hop in response to
the change in available bandwidth on its path. In [6], the authors have proposed a link
quality source routing (LQSR) protocol. It is based on DSR and uses ETX as the
routing metric. A new routing protocol called multi-radio link quality source routing
(MR-LQSR) is proposed in [7]. The process of neighbor node discovery and propaga-
tion of link metric are same as those in DSR. However, assignment of link weight and
computation of the path weight is different. A QoS enabling routing algorithm for
mesh-based wireless LAN architecture has been proposed in [8] where the wireless
users form an ad hoc peer-to-peer network. The authors also have proposed a protocol
for MANET called ad hoc QoS on-demand routing (AQOR) [9]. However, both these
algorithms are incapable of guaranteeing QoS [1]. In [10], the authors have shown
that if a weighted cumulative expected transmission time [7] is used in a link state
routing protocol, it does not satisfy the isotonicity property of the routing protocol and
leads to formation of routing loops. To avoid routing loops, an algorithm is proposed
that uses metric of interference and channel switching (MIC) as the routing metric.
In contrast to most of the above approaches, the proposed protocol performs on-
demand route discovery using multiple metrics like bandwidth, delay, and reliability
of the links and provides a routing framework that can support application QoS.
Two major design challenges for design of a mesh routing protocol are detecting and
avoiding unreliable routes, and accurately estimating the end-to-end delay of a flow.
1. Measuring link reliability: It has been observed that in wireless ad hoc networks
neighbor sensing with broadcast messages introduces communication gray zones [11].
In such zones, data messages cannot be exchanged although the Hello messages indi-
cate neighbor reachability. This leads to a systematic mismatch between the route
state and the real world connectivity, resulting in disruptive behavior. Since the rout-
ing protocols such as AODV and WMR [8] rely on control packets like RREQ, these
protocols are highly unreliable for estimating the quality of wireless links. Due to
communication gray zone problem, nodes that are able to send and receive bidirec-
tional RREQ packets may not be able to send/receive high bit-rate data packets. These
fragile links trigger link repairs resulting in high control overhead for routing proto-
cols. Therefore, designing a robust and reliable link quality estimator is crucial for
reducing the control overhead of a routing protocol in WMNs.
2. End-to-end delay estimation: An important issue in a routing protocol is end-to-
end delay estimation. Current protocols estimate end-to-end delay by measuring the
time taken to route RREQ and RREP packets along the given path. However, RREQ
and RREPs are quite different from normal data packets, and therefore are unlikely to
experience the same levels of traffic delay and loss as data packets. It has been ob-
served through simulation that a RREP-based estimator overestimates while a hop-
count-based estimator underestimates the actual delay experienced by the data packets
[1]. The reason for the significant deviation of a RREP-based estimator from the ac-
tual end-to-end delay is interference of signals. The RREQ packets are flooded in the
network resulting in a heavy burst of traffic. This heavy traffic causes inter-flow inter-
ference in the paths. The unicast data packets do not cause such events. Moreover, as
a stream of packets traverse along a route, due to the broadcast nature of wireless
links, different packets in the same flow interfere with each other resulting in per-
packet delays. Since the control packets do not experience per-packet delay, the esti-
mate based on control packet delay deviate widely from the actual delay experienced
by the data packets. Computing a reliable estimate of the end-to-end delay is, there-
fore, another challenge in design of a routing protocol for WMNs.
Every node estimates the reliability of each of its wireless links to its one-hop
neighbor nodes. For computing the reliability of a link, the number of control packets
that a node receives in a given time window is used as a base parameter. An exponen-
tially weighted moving average (EWMA) method is used to update the link reliability
estimate. If the percentage of control packets received by a node over a link in the last
interval of measurement of link reliability is Nt, and if Nt-1 is the historical value of the
link reliability before the last measurement interval, α = 0.5 is the weighting parame-
ter, then the updated link reliability (R) is computed as:
R = α * N t + (1 − α ) * N t −1 (1)
Every node maintains the estimates of the reliability of each of its links with its
neighbors in a link reliability table. The reliability for an end-to-end routing path is
computed by taking the average of the reliability values of all the links on the path.
Computation of the link reliability values are based on the RREQ packets on the re-
verse path and the RREP packets on the forward path. The use of routing path with
the highest reliability reduces the overhead of route repair and makes the routing
process more efficient.
The proposed protocol makes use of the knowledge of network topology by utilizing
selective flooding of control messages in a portion of the network. In this way,
broadcasting of control messages is avoided and thus the chances of network conges-
tion and disruption to the flows in the network are reduced. If both the source and the
destination are under the control of the same mesh router (Fig. 1), the flooding of the
control messages are confined within the portion of the network served by the mesh
router only. However, if the source and the destination are under different mesh
routers, the control traffic is limited to the two mesh groups.
To further reduce the overhead of control message and enhance the reliability in
routing, the nodes accept broadcast control messages from only those neighbors which
have the link reliability value greater than 0.5 (i.e., on average 50% of the control
packets sent from those nodes have been received by the node). This ensures that path
with less reliability values are not discovered and therefore not considered for routing.
packets are sent along the path which is likely to induce less end-to-end delay result-
ing in a better performance of the protocol as observed from the simulation results
presented in Section 5. The probe packets are identical to data packets so far as their
size, priority and flow rates are concerned. The objective of sending probe packets is
to simulate the data flow and observe the delay characteristics in the routing path.
Number of probe packets is kept limited to 2H for a path consisting of H hops to
make a tradeoff between control overhead and measurement accuracy [1].
A destination node sets on a timer after it receives the first probe packet from the
source node. The timer duration is based on the estimated time for receiving all the
probe packets and is computed statistically. The destination computes the average
delay experienced by all the probe packets it has received, and sends the computed
value to the source node piggybacking it on a RREP message. If the computed value
is within the limit of tolerance of the application QoS, the source selects the route and
sends data packets. If the delay exceeds the required limit, the source selects the next
best path (based on the arrival of RREP packets) from its table and tries once again.
Since the routing path is set up based on the probe packets rather than the naïve RREP
packets, the proposed protocol has higher route establishment The proposed algorithm
has higher setup time due to sending of the probe packets and selection of the best
path based on the estimated end-to-end delay. However, since the selected paths have
high end-to-end reliability, the delay and the control packet overhead are reduced
because of minimal subsequent route breaks.
PacketSize
estrat = (2)
X +Y
where,
2Pcongestion (3)
X = RTT
3
3Pcongestion
Y = RTO * min(1,3 * Pcongestion(1 + 32 Pcongestion
2
) (4)
8
In (2), RTT is the average round trip time for a control packet. RTO is the retransmis-
sion time out for a packet, and is computed using (5).
− −−− − −− −−−
RTO = RTT + k * RTT Var (5)
−− −− − −− −−−
where, RTT and RTT Var are the mean and variance respectively of RTTs and k is
set to 4. This bandwidth estimator is employed to dynamically compute the available
bandwidth in the wireless links on a routing path so that the guaranteed minimum
bandwidth for the flow is always maintained throughout the application life-time.
The proposed routing protocol also enforces cooperation among the nodes by identi-
fying the selfish nodes in the network and isolating them. Selfishness is an inherent
problem associated with any capacity-constrained multi-hop wireless networks like
WMNs. A mesh router can behave selfishly owing to various reasons such as: (i) to
obtain more wireless or Internet throughput, or (ii) to avoid path congestion. A selfish
mesh router increases the packet delivery latency, and also increases the packet loss
rate. A selfish node while utilizing the network resources for routing its own packet,
avoids forwarding packets for others to conserve its energy. Identification of selfish
nodes is therefore, a vital issue.
Several schemes proposed in the literature to mitigate the selfish behavior of nodes
in wireless networks such as credit-based schemes, reputation-based schemes and
game theory-based schemes [13]. However, to keep the overhead of computation and
communication at the minimum, the proposed protocol employs a simple mechanism
to discourage selfish behavior and encourage cooperation among nodes. To punish the
selfish nodes, each node forwards packets to its neighbor node for routing only if the
link reliability of that node is greater than a threshold (set at 0.5). Since the link reli-
ability of a selfish node is 0, the packets arriving from this node will not be for-
warded. Therefore, to keep its link reliability higher than the threshold, each node has
to participate and cooperate in routing. The link reliability serves a dual purpose of
enhancing reliability and enforcing node cooperation in the network.
The proposed protocol detects failure to guarantee QoS along a path with the help of
reservation timeouts in flow tables records maintained in the nodes, and by detection
of non-availability of minimum bandwidth as estimated along its outbound wireless
An Efficient and Reliable Routing Protocol for Wireless Mesh Networks 253
link. Failure to guarantee QoS may occur in three different scenarios. In the first case, a
node receives a data packet for which it does not find a corresponding record in its
flow table. This implies that a reservation time-out has happened for that flow. The
node, therefore, sends a route error (RERR) to the source which re-initiates route dis-
covery. In the second scenario, a destination node detects from its flow table records
that the data packets received have exceeded the maximum allowable delay (Tmax). To
restore the path, the destination broadcasts a new RREP back to the source, and the
source starts rerouting the packets via the same path of the RREP. This approach is
similar to the one proposed in [9]. In the third case, an intermediate node on the routing
path may find that the estimated bandwidth (using (2)) in its forwarding link is less
than the guaranteed minimum (Bmin) value. In this case, the intermediate node sends a
route error (RERR) to the source which re-initiates the route discovery process. The
real-time estimation of the bandwidth in the next-hop wireless link at each node on the
routing path makes the proposed protocol more robust and reliable compared to other
existing routing protocols for WMNs. For example, the protocol presented in [1] does
not provide any bandwidth estimation mechanism at intermediate nodes, and therefore
cannot provide delivery of all packets for every admitted flow in the system.
5 Simulation Results
The proposed protocol has been implemented in Qualnet network simulator [14]. The
WMN topology is shown in Fig. 2. The simulated network consists of 50 nodes of
which 5 are mesh routers and remaining nodes are mesh clients. Each mesh router has
9 mesh clients associated with it. To compare the performance of the proposed proto-
col with the protocol presented in [1] by Kone et al, parameters in the simulation are
chosen to be identical as those used in [1]. The simulation parameters are presented in
Table 1. The MAC protocol used is 802.11b, which is based on CSMA/CA protocol.
No wireless fading model is used scheme to keep it simple, yet realistic.
Parameter Values
Simulation area 1300 m * 1300 m
MAC protocol 802.11
PHY-model 802.11b
Propagation channel frequency 2.4 GHz
Raw channel bandwidth 11 Mbps
Simulation duration 15 s
Traffic type CBR UDP
Packet size 512 bytes
Max. no. of packets from a source node 10000
Hello packet broadcast interval 200 ms
Wireless fading model None
Node mobility None
Propagation limit of each node -111.dBm
IP queue scheduler Strict priority
To evaluate the effectiveness of the admission control module of the protocol, the
performance of the protocol in a heavy traffic condition is observed. Due to heavy
traffic, some flows in the network will not have their bandwidth and end-to-end delay
requirement satisfied, and hence will be rejected by the protocol. AODV, on the other
hand being non-aware about the QoS parameters, will try to route packets from all the
sources. The experimental results have shown that the proposed protocol has rejected
admission of new flows after the traffic level in the network has gone above a certain
limit. However, for flows that have been admitted, all packets are routed within the
required end-to-end delay. AODV, on the other hand has admitted all the flows and
routed them with increasing delay as the congestion in the network increased to due
more number of control packets. An important point worth mentioning here is that the
protocol in [1] is not able to achieve 100% packet delivery ratio for all flows. In con-
trast, the proposed protocol has been found to achieve 100% packet delivery. This is
due to bandwidth estimation at each intermediate node as discussed in Section 4.4.
For computing the overhead due to control packets, only the RREQ messages and the
probe packets in the proposed algorithm are considered since these are broadcast
messages and largely contribute to the control overhead of the protocols.
Fig. 3(a) shows the overhead of RREQ packets in AODV and the proposed proto-
col for different data rates. The control overhead of the proposed protocol is first
evaluated only with the RREQ packets and then with RREQ packets and the probe
packets together. This also gives an idea of the additional overhead introduced due to
the probe packets. It can be easily observed that the proposed protocol has a very
limited overhead even with the probe packets compared to naïve AODV protocol. It is
also worth observing that the proposed protocol has about 20% less control overhead
than the protocol proposed in [1] due to its robust bandwidth estimation of links.
An Efficient and Reliable Routing Protocol for Wireless Mesh Networks 255
Fig. 3(b) shows that AODV has a very high overhead of control packets for in-
creasing number of flows in the system. AODV always tries to establish routing paths
based on minimum hop-counts. It does not pay any attention to the link reliability.
This results in frequent selection of unreliable links and consequent link break and re-
discovery of route. This leads to large overhead of control information. In contrast,
the proposed protocol has a very limited control overhead since paths with higher link
reliability are only used for routing. The algorithm proposed in [1] has similar per-
formance as the proposed protocol in this case.
As mentioned in Section 4.4, the proposed protocol has the ability to detect selfish
nodes which try to use network resources without contributing to the cooperative
framework of routing. To evaluate its detection capability of selfish nodes, two types
of flows are distinguished: selfish and honest. A flow is considered selfish if either its
source or destination is a selfish node, otherwise the flow is considered to be honest.
Some mesh clients are selected randomly and configured as selfish nodes. In each of
the 20 run of the simulation, 10 flows of 50 kbps data rate are generated randomly.
It may be observed from Fig. 5 that AODV cannot restrict the traffic along the self-
ish flows. The selfish nodes can fully exploit the routing process to have their packets
routed in the network. However, the proposed protocol reduces the flows along the
selfish paths. In fact, the performance of AODV is not very much affected by pres-
ence of selfish nodes, since it never establishes routing path based on hello packets.
Since the proposed protocol establishes route based on hello packets received form
neighbors, its performance is affected by the presence of selfish nodes. However, its
performance is not substantially affected, since most in most cases, these nodes are
not allowed to participate in the routing, because of the low values of their link reli-
ability. It may also be mentioned that the proposed protocol is able to maintain, on
average, 35% more throughput in presence of selfish nodes when compared with the
protocol proposed in [1]. The large difference is due to its ability to detect selfish
nodes faster by its effective bandwidth estimation in the route where non-forwarding
of packets is treated as packet drops due to congestion.
Fig. 5. Avg. throughput with different protocols for varying percentage of selfish nodes
An Efficient and Reliable Routing Protocol for Wireless Mesh Networks 257
6 Conclusion
WMNs present a promising technology for providing low-cast last mile broadband
Internet connectivity through multi-hop communications. Designing an efficient and
reliable routing protocol for WMNs is a challenging issue. This paper has presented a
QoS-aware routing protocol that is capable of supporting real-time applications. By
robust estimation of available bandwidth in the wireless route and a reliable computa-
tion of end-to-end delay along a path, the protocol is able to sustain a high level of
throughput even in presence of a large proportion of selfish nodes in a WMN. More-
over, it is able to reduce control overhead substantially by exploiting the knowledge
of network topology. Simulation results have shown that the proposed protocol is
more efficient than some of the current routing protocols for WMNs.
References
1. Kone, V., Das, S., Zhao, B.Y., Zheng, H.: Quorum: Quality of Service in Wireless Mesh
Networks. Journal of Mobile Networks and Applications 12(5), 358–369 (2007)
2. Mahajan, R., Rodrig, M., Wetherall, D., Zahorjan, J.: Sustaining Cooperation in Multi-Hop
Wireless Networks. Proc. of NSDI 2, 231–244 (2005)
3. Cuto, D.S.J.D., Aguqayo, D., Bricket, J., Morris, R.: A High-Throughput Path Metric for
Multi-Hop Wireless Routing. In: Proc. of ACM MOBICOM, pp. 134–146 (2003)
4. Aguayo, D., Bicket, J., Biswas, S., Judd, G., Morris, R.: Link-Level Measurements from
an 802.11b Mesh Network. In: Proc. of ACM SIGCOMM, pp. 121–132 (2004)
5. Badis, H., Gawedzki, I., Al Agha, K.: QoS Routing in Ad Hoc Networks using QOLSR
with no need of Explicit Reservation. In: Proc. of VTC (September 2004)
6. Draves, R., Padhya, J., Zill, B.: Comparison of Routing Metrics for Static Multi-Hop
Wireless Networks. In: Proc. of ACM SIGCOMM, pp. 133–144 (2004)
7. Draves, R., Padhye, J., Zill, B.: Routing in Multi-Radio, Multi-Hop Wireless Mesh
Networks. In: Proc. of ACM MOBICOM, pp. 114–128 (2004)
8. Xue, Q., Ganz, A.: QoS Routing for Mesh-Based Wireless LANs. International Journal of
Wireless Information Networks 9(3), 179–190 (2002)
9. Xue, Q., Ganz, A.: Ad Hoc QoS On-Demand Routing (AQOR) in Mobile Ad Hoc
Networks. Journal of Parallel and Distributed Computing 63(2), 154–165 (2003)
10. Yang, Y., Wang, J., Kravets, R.: Interference-Aware Load Balancing for Multi-Hop Wire-
less Networks. Technical Report UIUCDCS-R-2005-2526, Dept. of Computer Science,
University of Illinois at Urbana-Champaign (2005)
11. Lundgren, H., Nordstrom, E., Tschudin, C.: The Gray Zone Problem in IEEE 802.11b
Based Ad Hoc Networks. MC2R 6(3), 104–105 (2002)
12. Yang, F., Zhang, Q., Zhu, W., Zhang, Y.-Q.: End-to-End TCP-Friendly Streaming Proto-
col and Bit Allocation for Scalable Video over Wireless Internet. IEEE Journal on Selected
Areas in Communications 22(22), 777–790 (2004)
13. Santhanam, L., Xie, B., Agrawal, D.: Selfishness in Mesh Networks: Wired Multi-Hop
MANETs. IEEE Journal on Wireless Communication 15(4), 16–23 (2008)
14. Qualnet Simulator, http://www.scalable-networks.com
A Context-Aware Service Model Based on
Workflows for u-Agriculture
1 Introduction
D. Taniar et al. (Eds.): ICCSA 2010, Part III, LNCS 6018, pp. 258–268, 2010.
c Springer-Verlag Berlin Heidelberg 2010
A Context-Aware Service Model Based on Workflows for u-Agriculture 259
2 Related Work
2.1 Context-Aware Workflow Languages
Workflow languages such as XLANG [8], WSFL [9] and WS-BPEL were devel-
oped by international IT companies such as Microsoft, IBM, Bea, Oracle etc.
in order to support service automation in business processes. The BPEL4WS
[10] is a workflow language based on web-service technologies and includes the
advantages of both WSFL and XLANG technologies. XLANG is one of the ini-
tial workflow languages using a directed graph approach and WSFL is a block-
structured workflow language for describing web services. However, They do not
have any element in their language schema or grammar to describe context in-
formation as transition conditions of services. The workflow languages can use
simple return values through XPath, XLink, Xpoint to express current service’s
transition conditions. Therefore, they are not suitable for a workflow language
for context-aware services in ubiquitous environment.
uWDL [2] is a workflow language for ubiquitous computing environment. It
includes <constraints>, <contexts>, and <profiles> elements to describe var-
ious contexts as service transition conditions in its language schema. And, For
260 Y. Cho, J. Moon, and H. Yoe
In Figure 2, the data set that consists of a value and type is an entity. A
few of entities form a constraint, which is based on RDF triplet. And, a context
includes several constraints using rule-based operation. For example, let suppose
that a numeric value 60 is sensed from the current temperature of a section A
in a greenhouse 1. A farmer may wants a service scenario for a heating service,
that when the current temperature of a section A in a greenhouse 1 is less
than Fahrenheit 50, start a heating system. In that case, a context may needs
constraints consisting of 3 entities, < Type:Fahr., value: 60>, <Type:greenhouse,
value: no.1>, and <Type:Position., value: (x, y)>. Therefore, the farmer only
has to describe the context as execution conditions of the heating service into
his workflow scenario. Then, when the workflow processor in Figure 1 finds the
context among numerous entities sensed from a real environment, it will start
the heating service.
Figure 3 describes workflow service patterns [15] added into the uWDL schema.
The suggested service model uses a workflow model as an work automation model
based on uWDL.
For an experiment using the suggested service model, we will take experimental
scenario for context-aware Services in u-agriculture using the expanded uWDL.
To do this, let’s suppose a work scenario in u-agricultural environment. The
scenario is as like this:
There is a farmer to cultivate of mushrooms in his/her greenhouse. Mushrooms
is one of crops that are very sensitive to temperature, humidity, and PH of soil.
From morning to midnight everyday, the farmer checks the cultivation factors
for his mushrooms. And, the farmer has to take proper works according to the
results of the examination. Furthermore, he/she has to incubate the mycelia
of mushrooms and shelter the cultivation soil for the continuous cultivation of
mushrooms. And, the farmer has to control the speed and the amount of the
A Context-Aware Service Model Based on Workflows for u-Agriculture 265
works according the market demands and the production of mushrooms. Clearly,
it is very labor-intensive jobs, and the farmer will want to reduce the amount
of the works which he/she has to do. To do this, let’s suppose a next scenario.
The scenario is based on a workflow which describes the works as the services.
Figure 4 shows the parts of the required services and the workflow scenario.
In Figure 4(a), the services are based on Web services and exist independently.
The service H may requests the current sensed values of humidity, temperature,
PH of soil, a pollution degree of air in the greenhouse, sunshine, and illuminance,
to various sensors positioned around the mushroom greenhouse. In Figure 4(b)
shows the possible service flows that can be connected with the services of Figure
4(a) in order to achieve a specific work. The service H is needed in the basic
cultivation control flow and the pollution control flow. The service C is used in
all of the three flows in Figure 4(b). Of cause, There can be other possible and
reasonable flows in addition to those in Figure 4. And, the procedural orders of
the flows in Figure 4 don’t have to be absolutely like that.
Fig. 4. The parts of the demanded services and the possible service flows
266 Y. Cho, J. Moon, and H. Yoe
Now, let’s make a context-aware workflow service model by using the three
flows with contexts, which may occur in the greenhouse. Figure 5 represents the
context-aware workflow service model. Figure 5(a) shows a possible composition
of the basic cultivation condition control flow and the pollution control flow in
Figure 4. The service H is needed for the basic cultivation condition control flow
and the pollution control flow.
The service H selects a next service flow according to contexts that the service
received from the requested sensors. For example, in case of a context data that
can provoke the basic cultivation services is sensed from the sensors, the service
H will move to the service A. And, when the sensed data is related with the
pollution control services, the service H will move to the service J. Of cause, if
the contexts for the two flows occur all together, the service H will go to the
two flows in a type of parallel. However, the ventilation service or the sunshine
service to purify the air of the greenhouse has to be executed just before the
heating service. So, it may be more reasonable that the service E connects to
the service B just before. Figure 5(b) shows that. Now, the growth control flow
has to be added to the Figure 5(b). First, the sequential flow like Figure 5(c) can
be considered simply. However, because the order check service or the product
checking service can affect the watering service or heating service, it is desired
that the services are started as soon as possible. And, the service I and K may
probably generate air pollution of the greenhouse. Therefore, those service can
reuse the existing service flow, which consists of J, E, and B, as their subflow.
The service J in Figure 5(d) illustrates the process to reuse the section X in
Figure 5(c) as a subflow. The expended uWDL can grammatically support the
method to reuse any partition of the existing service flows as a subflow.
5 Conclusion
A research for service automation and context-aware services is a valuable sub-
ject in u-agricultural environment like in the other fields. However, most existing
context-aware service models are not suitable for work automation and smart
cultivation in u-agriculture, because they have troubles to use or express contexts
to automatically execute a service. This paper introduced a context-aware service
model based on workflows for u-agriculture. The suggested service model offers a
structural context-model to describe contexts in u-agriculture through a entity,
a RDF-based constraint, and rule-based operators. And, it offers flow models of
an expanded uWDL as a service automation model. With the proposed context-
aware service model, developers can easily develope various context-aware service
flows into a service workflow for u-agriculture. And, a developer can reuse the
existing service flows in composition of a new service flow with the flow models.
Therefore, the proposed service model is expected to reduce the time and effort
to develop context-aware applications or the work automation in the fields of
u-agriculture.
References
1. Abowd, G.D., Dey, A.K., Brown, P.J., Davies, N., Smith, M., Steggles, P.: Towards
a Better Understanding of Context and Context-Awareness. In: HUC 1999, pp.
304–307 (1999)
2. Han, J., Cho, Y., Choi, J.: Context-Aware Workflow Language based on Web
Services for Ubiquitous Computing. In: Gervasi, O., Gavrilova, M.L., Kumar, V.,
Laganá, A., Lee, H.P., Mun, Y., Taniar, D., Tan, C.J.K. (eds.) ICCSA 2005. LNCS,
vol. 3481, pp. 1008–1017. Springer, Heidelberg (2005)
3. Aqeel-ur-Rehman, Shaikh, Z.A.: Intelligent Workflow: A Step towards Intelligent
Agent based Academic Workflow System. In: 2nd All Pakistan UTECH and ACM
Technical Conference (UTECH 2007), Karachi, Pakistan, November 23-24 (2007)
4. Tang, F., Guo, M., Dong, M., Li, M., Guan, H.: Towards Context-Aware Work-
flow Management for Ubiquitous Computing. In: Proceedings of ICESS 2008,
pp. 221–228 (2008)
268 Y. Cho, J. Moon, and H. Yoe
5. Lauser, B., Sini, M., Liang, A., Keizer, J., Katz, S.: From AGROVOC to the
Agricultural Ontology Service / Concept Server - An OWL model for creating
ontologies in the agricultural domain. In: Proceedings of the OWLED*06 Workshop
on OWL: Experiences and Directions, Athens, Georgia (USA), pp. 10–11 (2006)
6. Wark, T., Corke, P., Sikka, P., Klingbeil, L.: Transforming Agriculture through
Pervasive Wireless Sensor Networks. IEEE Pervasive Computing, 50–57 (2007)
7. Fileto, R., Liu, L., Pu, C., Assad, E.D., Medeiros, C.B.: POESIA: An Ontologi-
cal Workflow Approach for Composing Web Services in Agriculture. The VLDB
J. 12(4), 352–367 (2003)
8. Satish, T.: XLANG: Web Services for Business Process Design. Microsoft Corp.
(2001)
9. Leymann, F.: Web Services Flow Language (WSFL 1.0), Distinguished Engineer
Member IBM Academy of Technology, IBM Software Group (May 2001)
10. Tony, A., Francisco, C., et al.: Business Process Execution Language for Web Ser-
vices, BEA Systems. Microsoft Corp. IBM Corp., Version 1.1 (2003)
11. W3C: RDF/XML Syntax Specification, W3C Recommendation, 30-39 (2004/2006)
12. Vieira, P., Rito-Silva, A.: Adaptive Workflow Management in WorkSCo. In: 16th
International Workshop on Database and Expert Systems Applications (DEXA
2005), pp. 640–645 (2005)
13. Ardissono, L., Furnari, R., Goy, A., Petrone, G., Segnan, M.: Context-
awareworkflow management. In: Baresi, L., Fraternali, P., Houben, G.-J. (eds.)
ICWE 2007. LNCS, vol. 4607, pp. 47–52. Springer, Heidelberg (2007)
14. Han, J., Cho, Y., Kim, E., Choi, J.: A Ubiquitous Workflow Service Framework.
In: Proceedings of the 2006 International Conference on Computational Science
and its Application (2006)
15. van der Aalst, W.M.P., ter Hofstede, A.H.M., Kiepuszewski, B., Barros, A.P.:
Workflow Patterns. Distributed and Parallel Databases 14(3), 5–51 (2003)
A History-Based Scheduler for Dynamic Load Balancing
on Distributed VOD Server Environments
*
Corresponding Author.
D. Taniar et al. (Eds.): ICCSA 2010, Part III, LNCS 6018, pp. 269–276, 2010.
© Springer-Verlag Berlin Heidelberg 2010
270 J. Moon, H.-j. Moon, and Y. Cho
1 Introduction
Recently, the Internet is growing increasingly due to the development of computer
networks. In addition, there were many attempts to transfer audio and video data
through the Internet, and it can be possible to transfer movies via the World Wide
Web and high-speed Internet. Video-on-Demand (VOD) service is activating due to
these attempts and technologies. VOD allows users to select and watch/listen to video
or audio contents on demand over high speed networks, while sitting at home, and
have almost as good a control over the viewing of the video as when using a conven-
tional VCR [1]. VOD services have some different characteristics from HTTP service
as well access patterns. VOD services have long service time more than HTTP ser-
vices, and request more network bandwidth and hardware disk. Moreover, VOD is
more QoS-sensitive service than HTTP service, so users cancel their requests when
the service is delayed. Therefore, the scheduling algorithms that used in distributed
environments are not suitable for VOD system environments.
There are many researches on scheduling algorithm for distributed system
[2][3][4]. We have to consider dynamic loads information about servers in dynamic
service environments. Moreover, analysis of users request pattern must be considered
to predict system loads and balance loads of VOD system. For examples, when we
know the access pattern of users for a day, we are able to get information about the
time of day that resources are most frequently used. CDN (Content Distribution Net-
work) service uses such access pattern to store contents that are frequently used into
cache memory. If we record histories of user requests and then analyze the history,
the user request pattern is recognized by using some data mining techniques. Such
users request pattern operates as an important part of load balancing for VOD system
environments. In this paper, we suggest a scheduling algorithm for hierarchical VOD
system environment to provide high QoS and load balancing. The proposed schedul-
ing algorithm uses users request pattern, and selects appropriate servers to assign
various requests by using genetic algorithm. We developed a fitness function based on
history of user requests for the genetic algorithm.
2 System Model
VOD systems divided into three categories according to how to manage VOD servers:
central, distributed, and hierarchical server model. The central server model consists of
many VOD servers and a central server which controls users request to the VOD serv-
ers. This model is easy to set up and provides fast response time due to very simple
scheduling algorithm. However, the model results in lower quality when users far from
the central server. Moreover, the front-end server may cause a Single Point of Failure
(SPOF) because the central server does not work when user requests increase rapidly.
Distributed server model distributes VOD servers managing user requests by them-
selves and has no central managing server. This model provides fast response time
due to assigning user requests to the nearest VOD server. When the VOD server gets
high loads, user requests are forwarded to another VOD server nearest. Therefore,
there is no SPOF and provides scalability easily. However, VOD server has to keep
track of other servers’ loads information to forward user requests to another server
A History-Based Scheduler for Dynamic Load Balancing 271
which has low load and located near. The number of distributed servers increase, the
amount of information to keep track of as well as network overhead also increase.
Hierarchical distributed server model combines the advantages of central and dis-
tributed server model. This model distributes VOD servers geographically, and con-
trol servers that located in each site manage user requests and VOD servers. This
model can reduce control servers’ loads by assigning user requests to the nearest
server group. Therefore, there are no SPOF and scalability is provided by adding
VOD server in each site.
In this paper, we construct a hierarchical distributed VOD system model to develop
history based scheduling algorithm. Fig. 1 shows the system model which consists of a
global DNS, VOD servers, control servers, and database server. Each server group can
be composed of many VOD servers and a control server which manage the servers in
the group. And the control server performs load balancing by distributing user requests.
while (1) {
for (i = 1 to n)
{
Ti , Load = α × Li , CPU + β × Li , MEM + γ × Li , NT
First of all, we need to calculate total load of ith control server Si by estimating
CPU, memory, and network utilization rate of each server. A monitoring module
running on the control server collects each server’s CPU, memory, and network loads.
We assign a different weighting value according to the service provided by distributed
VOD system. Some services that transfer large amounts of data such as online broad-
casting or VOD service have more increase in network overhead than CPU and mem-
ory, while some services that request large amount of computing power such as online
game have more increase in CPU and memory utilization than network. Therefore, we
put more weight on the resources that have a big increase according to the type of
service.
When a control server Si is in the global DNS list and total load is larger than
the threshold value TH, the Si is removed from the list. When the Si is not in the list
and total load is less than the threshold value, Si is inserted in the list again. Then the
DNS manage the list with best servers. The threshold value is determined by using
heuristics and is assigned by administrator manually. We provide the following four
scheduling policies based on load information that is collected by monitoring module
running on control servers, and can modify the policy dynamically.
A History-Based Scheduler for Dynamic Load Balancing 273
Local load distribution is a method to distribute load between VOD servers managed
by a control server. We adopt a genetic algorithm to local distribution because there is
no general solution to find best servers under the various types of service and lots of
requests. When the requests come into the queue, we need to sort the requests as the
service type. After sorting, we generate a population of candidate solutions. We create
candidate solutions from 10% of all m× n cases randomly. A fitness function is a
particular type of objective function that prescribes the optimality of a solution in a
genetic algorithm so that that particular chromosome may be ranked against all the
other chromosomes. Optimal chromosomes, or at least chromosomes which are more
optimal, are allowed to breed and mix their datasets by any of several techniques. And
the following is a fitness function and related expressions.
m
F(x) = ∑T
i =1
i
4 Performance Evaluation
We implemented a simulator which is based on OPNET Modeler [6] and conducted
some comparison experiments with GA (genetic algorithm), RR (round-robin), and
Random scheduling algorithm when users increase. The first experiment is about
changes of average response time when users increase from 1 to 3,000, and Fig. 3
shows the result. As you can see from Fig. 3, the average response time is pretty much
the same to some extent which is about 800. However, there is a growing gap among
scheduling algorithms. Random scheduling shows the largest increase because highly
loaded servers are getting user requests continuously. In this experiment, the average
response time of GA, the proposed scheduling algorithm, is under 40 seconds until
users are 3,000.
Fig. 4 shows the cancellation rate when users increase. There is a little probability
that users cancel popular VODs and there is strong probability that users cancel non-
popular VODs. The cancellation rate also increases when the response time is getting
274 J. Moon, H.-j. Moon, and Y. Cho
too long. Compared to response time, it looks like Random algorithm has a high can-
cellation rate. This means that there are many users who cancel the VOD due to long
response time because the response time had a lot of ups and downs. There is no big
change of the average response times of RR and GA. However, the gap is growing
over 2,600 when using RR. In this experiment, Random algorithm shows 77%, RR
shows 13%, and GA shows 2.2% of cancellation rate. The result shows that there is a
little probability that users cancel the service using GA excepting when the user can-
cel the service due to lack of interest.
The next experiment is about average CPU utilization when the users increase.
Fig. 5 shows the result that the CPU utilization rate is growing when users increase.
The CPU utilization rate increase sharply when we use Random scheduling because
the requests may be assigned into highly loaded server. In response to this we can see
the response time increase in Fig. 3. The average CPU utilization of RR and GA
scheduling increases slowly and steadily. GA scheduling shows the very gently in-
crease because lots of requests are evenly distributed and the gap between servers’
loads is getting close.
We also conducted an experiment using user request pattern and history during a
day. The users’ pattern can get from the service history. In this experiment, we as-
sumed the user requests increase dramatically at 6 P.M. and the numbers of users
increase from 1 to 1,000.
Fig. 6 shows the result of the changes of response time as the user request pattern.
The x-axis refers to the time from 10 A.M. to 10 P.M. The response time is getting
longer when the rush hours of about 6 P.M. Random and RR scheduling shows the
response time is getting longer sharply. Random scheduling has a large gap between
response times when the system has not many requests, and moreover the response
time is longer when the requests increase rapidly. However, GA scheduling predicts
the loads of servers and distributes the loads evenly among servers, so the response
time increases slowly. Therefore, we can see that the proposed scheduling algorithm
provides more enhanced QoS.
References
1. Ma, K., Shin, K.G.: Multicast Video-On-Demand Services. ACM SIGCOMM Computer
Communication Review 32(1), 31–43 (2002)
2. Loukopoulos, T., Ahmad, I.: Static and Adaptive Distributed Data Replication Using
Genetic Algorithms. J. of Parallel and Distributed Computing 64(11), 1270–1285 (2004)
3. Rothlauf, F.: Representations for Genetic and Evolutionary Algorithms. In: Studies on
Fuzziness and Soft Computing, vol. 104 (2002)
4. Mundur, P., Simon, R., Sood, A.K.: End-to-End Analysis of Distributed Video-on-Demand
Systems. IEEE Transactions on Multimedia 6(1), 129–141 (2004)
5. Moon, J.-B., Kim, M.-H.: Dynamic Load Balancing Method Based on DNS for Distributed
Web Systems. In: Bauknecht, K., Pröll, B., Werthner, H. (eds.) EC-Web 2005. LNCS,
vol. 3590, pp. 238–247. Springer, Heidelberg (2005)
6. OPNET Technologies Inc., http://www.opnet.com/
A Secure Routing Protocol for Wireless Sensor Networks
1 Introduction
Recent convergence of technological and application trends has resulted in an excep-
tional level of interest in wireless ad hoc networks, and in particular, wireless sensor
networks (WSNs). The push was provided by rapid progress in computation and
communication technology as well as the emerging field of low cost, reliable,
MEMS-based sensors, while the pull was provided by numerous applications that can
be summarized under the umbrella of computational worlds, where the physical world
can be observed and influenced through the Internet and WSN infrastructures. Typi-
cally, WSNs contain hundreds or thousands of sensor nodes that have the ability to
communicate among them and also to a base station (BS) [1]. While the sensor nodes
have limited sensing region, processing power and energy, networking a large number
of such nodes makes a robust, reliable and accurate sensor network covering a wider
region. The WSNs are envisioned to play an important role in a wide variety of appli-
cations ranging from critical military surveillance to forest fire monitoring and the
building security monitoring. Since these networks are mostly deployed in harsh and
often hostile operating environments, security and reliability are two important as-
pects for their operations. Design of a robust and secure routing protocol for WSNs is
D. Taniar et al. (Eds.): ICCSA 2010, Part III, LNCS 6018, pp. 277–290, 2010.
© Springer-Verlag Berlin Heidelberg 2010
278 J. Sen and A. Ukil
2 Related Work
A considerable amount of research effort has been put in design of secure routing
protocols for WSNs. While some of these works have focused on defense against
outsider attacks on the key management protocols being used in the network [8, 9,
10], the others have attempted to secure node-to-node communication [10, 11]. How-
ever, all these schemes fail even if a single node is compromised.
An insider attack on a WSN is even more dangerous, since an attacker in this case
can launch various types of attacks such as dropping of legitimate packets, injecting
of bogus packets and sensing reports, advertising false routing messages, eavesdrop-
ping on the network communication etc. Parno et al have proposed a distributed
detection mechanism [12]. Newsome et al [13] have presented some techniques
for preventing an adversary from launching Sybil attack by arbitrarily creating new
identities.
Deng et al have proposed an intrusion tolerant routing protocol in wireless sensor
networks (INSENS) that adopts a routing-based approach to security in WSNs [14]. It
constructs routing tables in each node, bypassing malicious nodes in the network. The
protocol cannot totally prevent attack on nodes, but it minimizes the damage caused
to the network due to an attack. The scheme has reduced computation, communica-
tion, storage, and bandwidth requirements at the sensor nodes at the cost of greater
computation and communication overhead at the base station. Tanachaiwiwat et al
have proposed a novel secure routing protocol- trust routing for location-aware
A Secure Routing Protocol for Wireless Sensor Networks 279
The goal of the proposed scheme is to defend against packet-dropping attacks that
may be launched by compromised nodes in a WSN. As mentioned in Section 2, the
insider attacks may have devastating effects on the performance of a WSN and suit-
able defense mechanism should be deployed to counter such attacks.
A node in a WSN may drop packets because of two reasons: (i) it may be faulty
due to some problems in its hardware and/or software, or (ii) it may be malicious and
intentionally may not forward any packets. To ensure reliable packet delivery, some
authors [20, 21] have proposed use of acknowledgements (ACK). However, in pres-
ence of malicious nodes, an ACK received from the next-hop node cannot guarantee
that the packet is forwarded by the next-hop node; it only implies that the packet has
been received by the next-hop node. It will not be possible to ensure that the next-hop
node really forwards the packet just by implementing an ACK mechanism.
Due to the broadcast nature of the wireless links in the sensor network, it is possi-
ble for a node to overhear communications to and from its neighbor nodes. Therefore,
a node u after forwarding a packet to its next-hop v can listen on v’s communication
to check whether v really forwards the packet further. The simple monitoring mecha-
nism, however, cannot also guarantee reliable packet delivery either. A malicious
280 J. Sen and A. Ukil
node may forward the packet but it can make its intended next-hop id such that there
is no node in the neighborhood which has the matching id. Therefore, even if the
packet is forwarded by the malicious node it is ultimately lost in the network. This is
called routing disruption attack, and it is impossible to detect such attack by a naïve
neighborhood monitoring scheme.
The objective of the proposed scheme is to address all types of packet-dropping at-
tacks ranging from the most naïve type (e.g., detection of a faulty node) to the most
malicious type (e.g., routing disruption attack).
The cluster key is established by a node after the establishment of the pair-wise key
establishment. A node generates a cluster key and sends it to each of its neighbors
after encrypting it using the pair-wise shared key. The group key is pre-loaded but is
updated once any compromised node is detected in the network. This could be done
either by base station’s sending the new group key to each node using its individual
key, or a hop-by-hop basis using cluster keys.
because they are neighbors of node u, discard the packet. The primary node
knows the secondary monitoring nodes, since every node has the knowledge of
its 1-hop and 2-hop neighbors.
As the packet is routed along a path towards the sink, the above steps of NMS al-
gorithm except the neighbor list checking are executed at each hop so that reliable
packet delivery can happen through a single path. This is in contrast to the previous
schemes proposed in [19, 21, 22]. In these schemes, a node broadcasts a packet with-
out specifying a designated next-hop, and all neighboring nodes with smaller costs
A Secure Routing Protocol for Wireless Sensor Networks 283
(the cost at a node is the minimum energy required to forward a packet from the node
to the base station) or within a specific geographic region continue forwarding the
packet to the base station. If nodes v, w, and y have smaller costs than node u in Fig.
1, then each of them will forward packets received from node u following the existing
approaches. However, in the proposed scheme, nodes w and y only observe the packet
forwarding activities of node v, instead of actively forwarding the packets. In the
event of no packet drop, the routing to the base station happens in a single-path
thereby making the process highly energy-efficient.
Even in the event of a packet drop, the proposed algorithm works in a single-path
mode. This makes it more efficient than the one proposed in [7]. If the node v in Fig.
1 does not forward the packet it has received from node u, then one of the secondary
monitoring nodes w and y would forward the packet to its next-hop nodes. The node
(either w or y) that forwards the packet to its next-hop neighbors will first send a
broadcast message in its neighborhood so that its other neighbors would not forward
the same packet. This single-path routing approach even in the event of packet drop-
ping makes the proposed scheme extremely energy-efficient and very less delay-
inducing. Fig. 2 shows an example of the application of the scheme, where two mali-
cious (or faulty) nodes are bypassed as the packet is routed to the base station in a
single-path.
For the scheme to work, each packet should be encrypted with a cluster key of the
forwarding node so that all the neighbors of the forwarding node can decrypt and
overhear it. If a link-level encryption was applied between each pair of nodes in the
routing path, the scheme would have been more robust, since a compromised node
could decrypt only the packets which were destined to it. However, it would have
made the scheme less resilient to packet dropping attack. Since encryption with a
cluster key provides a reasonable level of robustness to a node compromise, [11] and
also supports local broadcast (i.e. resiliency against packet-dropping) it makes the
algorithm optimum in its performance [11].
To make the scheme robust to routing disruption attack, where a node intentionally
forwards the packets to a spurious address of the next-hop so that the packet is lost in
routing, it is necessary that each node should prove that it really has the claimed
neighbors. Apparently, a node has the knowledge of its direct neighbors by neighbor
discovery and pair-wise key establishment phases discussed earlier. However, in the
284 J. Sen and A. Ukil
case of two-hop neighbors, a malicious node v can inform its neighbor u that it also
has neighbor node x (any possible id in the network) which in fact is not a neighbor of
node v (Fig. 1). Apparently, there is no way node u can detect these false claim of v
since x is not in the neighborhood of u. Therefore, it is absolutely necessary to devise
a scheme by which a node can verify the neighbors of each of its neighboring nodes.
This scheme is explained in the next subsection.
To verify the 2-hop neighbors, i.e., the neighbors of the neighbors of each node, the
proposed scheme utilizes a concept of neighbor list checking (NLC). NLC uses the
pair-wise keys established in LEAP. During neighbor discovery, NLC uses three-way
handshaking in order to identify not only the communicating nodes but also their
respective neighbors. There are two distinct types of neighbor discovery in NLC. In
one case, both the nodes involved in the neighbor discovery process are still within
the initial Tmin – both the nodes are referred to as new nodes. In the other case, the
neighbor discovery happens between a newly-deployed node within the initial Tmin
and an existing node over the initial Tmin – the latter node is referred to as an old node.
(i) Neighbor discovery between new nodes: If a new node broadcasts its neighbor
list before the expiry of its initial Tmin, it is verifiable. Therefore, it is essential that
both nodes broadcast their respective neighbor lists before their respective Tmin. The
three-way handshaking neighbor discovery between pure node u and v may be repre-
sented as follows [7]:
u Æ * : u, Ru
v Æ u : v, Tv, Rv, MAC (Kv, Ru | Ku | (u, Tv, Rv))
u Æ v : u, Tu, MAC (Kuv, Rv | (u, Tu))
Tv and Tu are the time remaining to reach Tmin of v and Tmin of u, respectively. Ru is
a random number generated by node u. MAC (K, M1 | M2) denotes message authenti-
cation code of concatenated message of M1 and M2 using key K. Node u broadcasts its
id, and waits for each neighbor v’s ACK. The ACK from every neighbor v is authenti-
cated using the master key Kv of node v. Since node u knows KI, it can derive Kv = fKI
(v). The ACK from node v contains Tv. If Tv is non-zero, node v claims to be a new
node. Ku in MAC proves node v to be a new node, since new node v should know KI
and derive Ku = fKI (u). Node u records T^v (Tv added to the current time of node u) in
the entry for node v in the neighbor information table (NIT). Node u computes its
pair-wise key with v, Kuv = fKv (u). Node u also generates MAC (Kv, v | u) (which
means that v certifies u as an immediate neighbor), and stores it as a certificate.
The ACK from node u also contains Tu. This ACK is authenticated using their pair-
wise key Kuv, which proves node u a new node and u’s identity. Node v then records
T^u (Tu added to the current time of v) in the entry for u in its NIT. It also generates
MAC (Ku, u | v) and stores it as a certificate. This completes the three-way handshak-
ing process.
A Secure Routing Protocol for Wireless Sensor Networks 285
Every new node u broadcasts its neighbor list just prior to Tmin of u. Each receiving
neighbor v checks whether the receiving time at v is prior to T^u in the NIT. If yes, the
neighbor list of u is now certified by each neighbor v.
(ii) Discovery between a new and an old node: After the bootstrapping phase, new
nodes can join the network. This situation is depicted in Fig. 3, where the new node x
has joined the network. The nodes u, v and w are old nodes. In this scenario, the mes-
sages exchanged during the three-way handshaking are different. The message ex-
change between the new node x and the old node u is as follows [7]:
x Æ * : x, Rx
u Æ x : u, Tu, Ru, v, MAC (Kv, v | u), w, MAC (Kw, w | u), MAC (Ku, Rx | (u, Tu, v,
MAC (Kv, v | u), w, MAC (Kw, w | u))
x Æ u : x, Tx, MAC (Kx, x | u), v, MAC (Kv, x | u), w, MAC (Kw, x | u), MAC (Kxu,
Ru | (x, Tx, MAC (Kx, x | u), v, MAC (Kv, x | u), w, MAC (Kw, x | u))
The node x broadcasts its id, and waits for the ACKs from its neighbors. The ACK
from neighbor u is authenticated using the master key Ku of u. Since x knows K1, it
computes Ku = fKI (u). The ACK from u contains Tu, the time remaining to reach Tmin
of u. As u is an old node in Fig. 3, Tu is zero. Node u sends its neighbor list (v and w)
to x including their respective certificates in the ACK. Node x can verify the list since
it knows KI. If the verification is positive, x computes pair-wise key Kxu = fKu (x), and
also generates MAC (Ku, u | x) and stores it as a certificate.
The ACK from x also contains Tx, the time remaining to reach Tmin of x. This ACK
is authenticated using Kxu, to prove x’s identity as a new node. Node u computes T^x
(Tx added to the current time of u) and makes an entry for x in the neighbor table.
Node x also sends the certificate for u, MAC (Kx, x | u) in the ACK. It also provides
one-time certificates for each of u’s neighbors (v and w) and completes the three-way
handshaking process.
Node u broadcasts the one-time certificates received from x to its neighbors w and
y so that they are able to verify the authenticity of x. The contents of the one-time
certificates are as follows [7]:
u Æ * : u, x, v, MAC (Kv, x | u), w, MAC (Kw, x | u), KAu, MAC (Kcu, (u, x, v, MAC
(Kv, x | u), w, MAC (Kw, x | u), KAu))
286 J. Sen and A. Ukil
KAu is a local broadcast authentication key in u’s one-way key chain, Kcu is the
cluster key of u. Each neighbor v of u verifies x by checking its one-time certificate,
MAC (Kv, x | u) and x is certified by each neighbor v of u.
Before the expiry of Tmin of x, it broadcasts its neighbor list. Each neighbor u of x
checks whether the receiving time at u is before T^x in the neighbor table. If yes, the
neighbor list of x is certified by u.
Through the above neighbor list checking process, every node knows the certified
neighbors of each of its neighbors. With this knowledge, the monitoring nodes check
the validity of the claims of the next-hop of the target forwarding node.
The information obtained through neighbor list checking is stored in the neighbor
information table (NIT) of each node. Table 1 shows the NIT of node u as depicted in
Fig. 3.
The NIT is used by the monitoring nodes. For example, if node u overhears a
packet sent from w to v, it first checks its NIT and finds that both w and v are its
neighbors. Node u starts listening on v’s traffic as a secondary monitoring node, and if
v does not forward the packet to one of its verified neighbors, u will first broadcast a
message in its neighborhood and then forward the packet to its next-hop other than v.
Node u then starts acting as a primary monitoring node.
5 Simulation Results
To evaluate the performance of the proposed protocol, simulations have been carried
out on network simulator ns-2. To compare the performance of the scheme with that
proposed in [7] identical parameters are chosen. The simulation parameters are pre-
sented in Table 2.
The malicious nodes are configured in such a way that they do not forward any
packets. These nodes are strategically distributed in a square area of 200 m side (from
150m to 350m of each side of the 500 m × 500 m target area), so that they are half-
way between the source and the base station. TinyOs beaconing [23] has been used as
the base routing protocol in the simulation to maintain uniformity with [7]. Each
simulation is repeated with 100 different topologies, and the average value is taken as
the final result.
Fig. 4 and Fig. 5 show the packet delivery ratio (the ratio of the number of packets
successfully reaching the base station to the total number packets sent by the source)
in presence of varying number of malicious nodes (x) with N = 300 and N = 600
A Secure Routing Protocol for Wireless Sensor Networks 287
Parameter Values
No. of sensor nodes (N) 300, 600
Area of simulation 500 m * 500 m
Transmission range of each sensor 30 m
Degree of each node 10 (N=300), 20 (N=600)
Position of source (50, 50)
Position of BS (450, 450)
Avg. no of hops from source to BS 18
Fig. 4. Packet delivery ratio vs. no. of malicious nodes (N = 300, x = 0 ~ 50)
Fig. 5. Packet delivery ratio vs. no. of malicious nodes (N = 600, x = 0 ~ 100)
respectively. It is observed that with N = 300, the packet delivery ratio falls with the
increase in number of malicious nodes. The performance of the proposed protocol is
found to be almost identical to that of NWS proposed in [7]. With both these proto-
cols, the packet delivery ratio is on the average more than 80%. Since in NWS some
packets are routed in multi-path, it has a marginally higher of packet delivery ratio in
288 J. Sen and A. Ukil
Fig. 4. As can be observed from Fig. 5, with N = 600, the proposed protocol has same
performance as that of NWS. Even when x = 100, the delivery ratio is as high as 99%.
As has been pointed out in [7], this is because of the higher degree of sensor nodes (d
= 20), more number of malicious nodes can be bypassed.
Fig. 6 and Fig. 7 show the number of nodes that has participated in packet forward-
ing for N = 300 and N = 600 respectively. Since the source is 18 hops away from the
base station, the number of relaying nodes in single-path routing case remains con-
stant at 18. The number of forwarding nodes was found to be increasing with x, since
more number of nodes needs to be bypassed. However, in the proposed protocol, the
number of forwarding nodes was found to be much smaller than in the NWS. In
NWS, with N = 600, the number of forwarding nodes increases very fast with the
increase in x, since more number of secondary monitoring nodes participate in rout-
ing. While this ensures reliability, it has an adverse effect on the energy-efficiency of
the network. As the same packet is being routed in multi-path, NWS protocol has
large energy consumption on the average. The proposed protocol while maintaining
A Secure Routing Protocol for Wireless Sensor Networks 289
the same level of packet delivery ratio (i.e. reliability), involves less number of
forwarding nodes. It, therefore, consumes much less energy than NWS and thereby
enhances the network life-time.
6 Conclusion
This paper has presented a reliable and energy-efficient single-path routing protocol
for WSNs. The proposed protocol works reliably and efficiently even in presence of
malicious nodes on the routing path that drop packets without forwarding them. The
scheme is designed for hop-by-hop reliable forwarding of packets from the source to
the base station while routing every packets in single-path. The simulation results
have shown that the protocol is more energy-efficient than some of the existing rout-
ing protocols for WSNs while maintaining the same level of reliability.
References
1. Akyildiz, I.F., Su, W., Sankarasubramaniam, Y., Cayirci, E.: A Survey on Sensor Net-
works. IEEE Communications Magazine 40(8), 102–114 (2002)
2. Karlof, C., Wagner, D.: Secure Routing in Wireless Sensor Networks: Attacks and Coun-
termeasures. In: Proc. of 1st IEEE International Workshop on Sensor Network Protocols
and Applications, pp. 113–127 (2005)
3. Perrig, A., Stankovic, J., Wagner, D.: Security in Wireless Sensor Networks. Communica-
tions of the ACM 47(6), 53–57 (2004)
4. Wood, A., Stankovic, J.: Denial of Service in Sensor Networks. IEEE Computer 35, 54–62
(2002)
5. Intanagonwiwat, C., Govindan, R., Estrin, D., Heidemann, J., Silva, F.: Directed Diffusion
for Wireless Sensor Networking. ACM/IEEE Transactions on Networking 11(1), 2–16
(2003)
6. Gupta, G., Younis, M.: Fault-Tolerant Clustering of Wireless Sensor Networks. In: Proc.
of IEEE Wireless Communications and Network Conference (WCNC), pp. 1579–1584
(2003)
7. Lee, S.-B., Choi, Y.-H.: A Resilient Packet-Forwarding Scheme against Maliciously
Packet-Dropping Nodes in Sensor Networks. In: Proc. of the 4th ACM Workshop on Secu-
rity of Ad Hoc and Sensor Networks, pp. 59–70 (2006)
8. Chan, H., Perrig, A., Song, D.: Random Key Pre-Distribution Schemes for Sensor Net-
works. In: Proc. of IEEE Symposium on Security and Privacy, pp. 197–213 (2003)
9. Eschenauer, L., Gligor, V.D.: A Key Management Scheme for Distributed Sensor Net-
works. In: Proc. of the 9th ACM Conference on Computer and Communication Security
(CCS), pp. 41–47 (2002)
10. Perrig, A., Szewczyk, R., Wen, V., Culler, D., Tygar, J.: SPINS: Security Protocols for
Sensor Networks. In: Proc. of ACM MobiCom, pp. 189–199 (2001)
11. Karlof, C., Sastry, N., Wagner, D.: TinySec: A Link Layer Security Architecture for Wire-
less Sensor Networks. In: Proc. of ACM SensSys, pp. 162–175 (2004)
12. Parno, B., Perrig, A., Gligor, V.D.: Distributed Detection of Node Replication Attacks in
Sensor Networks. In: Proc. of IEEE Symposium on Security and Privacy, pp. 49–63
(2005)
290 J. Sen and A. Ukil
13. Newsome, J., Shi, E., Song, D., Perrig, A.: The Sybil Attack in Sensor Networks: Analysis
and Defenses. In: Proc. of IEEE IPSN, pp. 259–268 (2004)
14. Deng, J., Han, R., Mishra, S.: INSENS: Intrusion-Tolerant Routing in Wireless Sensor
Networks. Technical Report CU-CS-939-02, Dept. of Computer Science, University of
Colorado at Boulder (2002)
15. Tanachaiwiwat, S., Dave, P., Bhindwale, R., Helmy, A.: Routing on Trust and Isolating
Compromised Sensors in Location-Aware Sensor Networks. In: Proc. of 1st International
Conference on Embedded Networked Sensor Systems, pp. 324–325 (2003)
16. Zhu, S., Setia, S., Jajodia, S.: LEAP: Efficient Security Mechanisms for Large-Scale Dis-
tributed Sensor Networks. In: Proc. of the 10th ACM Conference on Computer and Com-
munications Security (CCS), pp. 62–72 (2003)
17. Ganesan, D., Govindan, R., Shenker, S., Estrin, D.: Highly Resilient, Energy-Efficient
Multipath Routing in Wireless Sensor Networks. Computing and Communications Review
(MC2R) 1, 11–25 (2002)
18. Deng, J., Han, R., Mishra, S.: A Performance Evaluation of Intrusion-Tolerant Routing in
Wireless Sensor Networks. In: Proc. of the 2nd International Workshop on Information
Processing in Sensor Networks (IPSN), pp. 349–364 (2003)
19. Ye, F., Zhong, G., Lu, S., Zhang, L.: GRAdient Broadcast: A Robust Data Delivery Proto-
col for Large Scale Sensor Networks. ACM Wireless Networks, WINET (2005)
20. Carbunar, B., Ioannidis, I., Nita-Rotaru, C.: JANUS: Towards Robust and Malicious Resil-
ient Routing in Hybrid Wireless Networks. In: Proc. of ACM Workshop on Wireless Secu-
rity, WiSe (2004)
21. Morcos, H., Matta, I., Bestavros, A.: M2RC: Multiplicative-Increase/Additive-Decrease
Multipath Routing Control for Wireless Sensor Networks. ACM SIGBED Review 2
(2005)
22. Yang, H., Ye, F., Yuan, Y., Lu, S., Arbough, W.: Towards Resilient Security in Wireless
Sensor Networks. In: Proc. of ACM MobiHoc, pp. 34–45 (2005)
23. Hill, J., Szewczyk, R., Woo, A., Hollar, S., Culler, D., Pister, K.: Systems Architecture Di-
rections for Networked Sensors. In: ACU ASPLOS IX (2000)
24. Ssu, K.-F., Wu, T.-T., Huang, S.-K., Jiau, H.C.: A Dependable Routing Protocol in Wire-
less Sensor Networks. In: Proc. of ICWMC, pp. 1–6 (2006)
Efficient Pairwise Key Establishment Scheme Based on
Random Pre-distribution Keys in WSN∗
Abstract. Key establishment and management are the cores of many security
mechanisms and services in wireless sensor networks which are commonly de-
ployed in a hostile environment. Pre-distributing keys into sensor nodes is a
practical option for key management in wireless sensor environment. The nego-
tiation and establishment of pairwise keys are much more important after the
nodes are deployed, so that the whole network will work in expectation and
safely. In this paper, we proposed an efficient pairwise key establishment
scheme used in cluster-based random key pre-distribution protocol, which can
provide perfect connectivity and security for large scale wireless sensor net-
works. Simulation results and performance analysis indicates that it also has
other strongpoint, lower communication and computational overhead, much
more scalable and flexible for the network size, a better balance between the
node storage capacity and communication overhead.
1 Introduction
Wireless Sensor Networks (WSN), the main task is sending data which is collected
from the sensor nodes to the base station in time, consists of a large number of sensor
nodes. And the sensor node, which is powered by the built-in battery and equipped
with a tiny sensor device, has limited capability of data processing and storage, and
can carry out a certain distance wireless communication. WSN is a new research
direction in information technology due to its large-scale of applications, such as in
environmental monitoring, military, homeland security, traffic control, community
security, forest fire prevention, targeting areas and so on. As the sensor nodes are
mostly deployed in unattended or hostile areas, once the sensor network deployed in
enemy controlled areas, the enemy will tap the network communication, palm off the
right nodes ,send malicious information to the network, so that they can destroy the
network, and make it loss of normal function.
∗
The work reported in this paper is supported in part by the national science and technology
major project No. 2009ZX03006-001.
D. Taniar et al. (Eds.): ICCSA 2010, Part III, LNCS 6018, pp. 291–304, 2010.
© Springer-Verlag Berlin Heidelberg 2010
292 H. Wang et al.
2 Related Work
To ensure secure communications, the sensor nodes need to consult keys; according
to the characteristics of sensor networks and the practical application, secret keys are
pre-loaded into each sensor node by using random deployment method before they
are deployed. There are several key pre-distribution models, the simplest way is that
all nodes share a master key, but this method has poor security; or any two of nodes
share the pairwise key, the security of such programs is good, but poor network ex-
pansion is not suitable for large-scale node network.
E-G program [9] , the basic scheme, with the consideration of security and the space
of storage in nodes, gives the basic random key pre-distribution scheme. This program
is suit for sensor networks and complete the establishment of the keys in three phases,
but did not give a clear way how to determine the pairwise key, or how to establish the
pairwise key by the search path. Chan et al. [3] add the minimum requirements that any
two of nodes share at least q common keys based the scheme [9]. This way increased
security, but with the increase of the compromised nodes, the performance become
relatively poor, meanwhile, it has the same deficiency with scheme [9].
Ren et al. [4] improve the structure of key ring in the basic scheme, give the
method of sub-key pool allocation; Du et al. [5] build a multi-key space based on
matrix theory; Liu et al. [6] use the polynomial instead of the key pool and construct
the application of the safety threshold in the polynomial; Du et al. [7] and Liu et al.
Efficient Pairwise Key Establishment Scheme 293
[8] further extended the basic scheme using network pre-deployment knowledge. All
of these have a clear way how to determine the pairwise key.
Blom et al. [11] proposed the symmetric matrix operations; in order to the
network has the λ-secure property. Blundo et al. [12] proposed a polynomial-based
key pre-distribution scheme. Scheme [13] [14] also use the symmetric polynomial
f ( x, y ) = f ( y, x) to establish the pairwise keys, either for single base-station or mul-
tiple base-station. All these need much computing and have a poor scalability.
Our goal is to develop and improve the current key pre-distribution schemes. We
give the detailed description how to establish the pairwise keys which provide a per-
fect security and connectivity. The scheme we proposed is not only reduced the keys
stored in nodes but also improve the network’s security.
3.1 Notation
keys within the same cluster of nodes, but also can reduce the number of keys stored
in nodes. Through the overlapping sub-key pool, the nodes in adjacent cluster can be
up to a certain probability, then, when using cluster head mechanism, inter-clusters
communication can be completed through the cluster head. In this model, we use the
head management mechanisms, cluster head products after achieving connectivity in
the cluster; the specific method is not our focus of this paper. After sensor network is
divided into several clusters, when the base station sends control commands, we can
broadcast news according to the requirements of different regions, this has more per-
tinence. When the whole network has requires or control commands to, it broadcasts
to the whole network. If this, it can reduce the numbers of broadcasting information.
The entire network topology model is shown in Fig. 1. Key server (base station) es-
tablishes key pool P by generating random numbers as keys, the size of key pool is |P|.
Suppose that the total number of nodes in the network is n. It has the node clustering
when deploying the networks. Assuming that it is divided into t, the number of nodes
in each cluster is r. Before the deployment of the network, the sub-key pool selected
randomly by the key server for each cluster is M, which has a number of |M|; in each
cluster, for each node we take randomly select m keys from its sub-key pool M.
When the network deployed, it has a certain probability that any two or more of
nodes share the same key within a cluster. When the two nodes have the same key, we
can choose one or more as the pairwise key. As the network takes the cluster head as
manager for each region, each cluster can consider a cluster head using some methods
and establish the pairwise keys with the adjacent clusters. Obviously the different
cluster still has a certain probability to get the same key.
According to this model of network topology, node connectivity has two kinds of
situation, connectivity between two nodes in one cluster, namely, the probability of
Efficient Pairwise Key Establishment Scheme 295
existence of the same key in the different clusters. For the same cluster, in the net-
work of Fig. 1 , take C1 for example, the probability that Ni and Nj share at least one
key is p. Ni selects m keys from | M | has a total of C|mM | ways, Nj has also C|mM |
ways. Suppose that nodes Ni, Nj shared w keys, selecting w shared keys has C|wM |
m − w)
methods, Ni and Nj still have 2(m-w) distinct keys, which has a kind of C|2(
M |− w
selection ways. And to partition 2(m-w) different keys to the two nodes equally, we
have C2(m −mw− w) ways. So the probability that two nodes share w keys is
m−w) m− w
C|wM |C|2(
M |− w C2( m − w )
p( w) = (1)
(C|mM | ) 2
The probability that two nodes share at least one key is
w
p = ∑ p (i ) ( w ≤ m) (2)
i =1
When the deployment of a whole network is completed, each cluster enters bootstrap-
ping phase. For each cluster, the node find the same key by broadcasting consultation
296 H. Wang et al.
mechanism, and select one of the common keys (if there were multiple) as the pairwise
key through the rules scheduled, the process of looking for shared key is as follows.
The node u broadcasts its own message that contains the node’s ID u, random
number N u and m keys’ IDs,
u − > *: Bu = {u, N u }
U {k _ ID1, k _ ID 2,L , k _ IDm}
When other nodes of the cluster, such as the node v, received the message Bu , it will
compare all the ID with its own stored in the keys’ ring, then reply to u with the same
ID which is in order from small to large. Due to the rules predetermined by the node u
and node v: they take the key of the smallest identifier as their pairwise key, and when
this key is compromised or leaks, the key by followed will be as pairwise key. Node v
sends a message to the node u as follows,
v − > u : M = {v, N u }
U {k _ ID _ uv1, k _ ID _ uv 2,
L , k _ ID _ uvw}
When u received all the nodes’ messages of common key in the cluster, the pairwise
keys will be stored in a form just as Table 1. (1 ≤ wi ≤ wi +1 ≤ m (1 ≤ i ≤ n′) )
The other nodes broadcast their own key messages similarly, and build the pair-
wise key tables with neighbor nodes. If network needs, such as unified management
in cluster, data integration and so on, we can use hierarchical, that’s to say, the estab-
lishment of cluster head for each cluster. The communication between clusters will be
achieved by the cluster head who already had the pairwise keys with the cluster head
of the adjacent clusters. The method establishing the pairwise key between the cluster
heads is the same as the above.
Neighbor
Pairwise key common keys
node
v K _ uv1 K _ uv 2,L , K _ uvw1
a K _ ua1 K _ ua 2,L , K _ uaw2
b K _ ub1 K _ ub 2,L , K _ ubw3
c K _ uc1 K _ uc 2,L, K _ ucw4
…… …… ……
In section 4.1, the pairwise key is established when there exist the common keys be-
tween two nodes, so it just needs one session. When the direct establishment of the
Efficient Pairwise Key Establishment Scheme 297
pairwise key is completed, the node-to-node within the cluster not only has the pair-
wise key, but also forms a table of shared keys. So does with the cluster head-to-head
between the clusters. The table is composed of the neighbor nodes and their security
links which have established their pairwise keys. The physically adjacent nodes which
are not able to find common keys can create it via intermediate nodes. This section
will give an approach about indirectly establishing the pairwise key.
When there is no common key between two nodes, such as the node u and v, but it
has formed the table of the pairwise key and the shared keys with other neighboring
nodes. Now the node u and v have to establish the pairwise key, and it must be safe,
not known to other nodes in the establishing process. The principle of establishment
is: we choose the node which exists the least shared keys with the sending node as the
forward transmitting node. Shown in Fig. 2, let u stored the shared key with a, b, c, d,
e, it needs to choose two nodes from them, making the both nodes and v have the
maximum probability of the common keys. Assume that a, c have the least numbers
of shared keys in u’s pairwise keys’ table, then u chooses a, c as the forwarding node.
The reason is that there is no shared key between u and v, yet a, c has the least shared
keys with u, then the probability of existing shared keys with v will be the largest.
The specific description of establishing process is as follows.
When node v receives the broadcast message of node u, found that there is no
common key, and then returns the following message,
M = {v, N u , N v } .
Node u receives the message M, stores N v , then selects a, c from the pairwise
keys table to forwarding the key’s materials in two paths according to above principle.
u − > a : Μ = {u , Ekua {v | N u1}}
u − > c : Μ = {u , Ekuc {v | N u 2 }}
When node a, c receive their message M, if they have the pairwise keys with v,
transmit them to v. As the node v receives the message from node a and c, calculates
the pairwise key:
kuv = N v ⊕ N u1 ⊕ N u 2 ,
and node u generates pairwise key using the same method.
298 H. Wang et al.
When either a or c doesn’t have pairwise key with v, u receives the response; it
will choose other nodes in the pairwise key table in the same principle except a and c,
repeats the above steps until successfully establish pairwise key with v. Other nodes
use the same method to establish pairwise keys with the nodes they do not have
shared keys in the cluster, and add the pairwise keys into their own pairwise keys
table. If there is no shared key between the adjacent cluster-heads, they establish the
pairwise key through the base station. In the whole network, each node established
pairwise key among the nodes in the cluster and between the cluster-heads, we
achieved the whole network connectivity.
5 Performance Analysis
In this section, we analysis and evaluate the security and performance of our scheme
in connectivity, security, scalability and node storage. Connectivity: it is on behalf of
the capability of the network security connectivity. It defined as the probability of
establishing secure link when two nodes are within the scope of communication; for
this article, it means that the probability of the two nodes in a cluster storing at least
one common key, as well as the two nodes in different clusters. When nodes store
fewer keys, the probability should be as high as possible. Security: the nodes in the
network should ensure that expected security, which should store the least key and
cost a minimum communication overhead. Node storage: since the node’s limited
storage resource, when the network reaches certain connectivity, the nodes should
store fewer keys.
In the section 3.2, we have got the probability of at least one shared key existed in two
nodes. Suppose the number of key pool is |P|=10000, when |M|=(0.01,0.02,
0.03,0.04)|P|, this probability is equation (3), the relationship between the probability
and the keys stored in node is shown in Fig. 3.
ye
k
en
ot
sa
el
ta
gn
ir
ah
sf
o
yit
ilb
ab
or
P
p
0.92~0.95 0.95~0.98 0.98~0.99
|M|
100 15 16,17 18,19
200 22,23 24-26 27,28
300 27,28 29-32 33,34
400 31-33 34-37 38-40
The Fig. 3 shows, when the sub-key pool is constant, the probability increases as
the number of keys stored in nodes increases; nodes store the same amount of keys,
the smaller the sub-key pool, the greater the probability; that’s to say the possibility of
finding pairwise keys will be larger. Even if the two nodes does not exist any same
key, we can establish pairwise keys which is described in section 4.2. Use the method
in section 4.2, we can guarantee the network wide connectivity. Table 2 shows when
asked a certain probability, a single node, should store the number of keys. Compared
to E-G scheme, the new scheme significantly reduces the number of keys.
For the adjacent two clusters, the probability of two nodes, equation (4), simplified
as,
p2 = (1 − f1 ( M , P))(1 − f 2 (m, M ))
( P − M )( P − M − 1)L ( P − 2M + 1)
f1 ( M , P) =
P( P − 1)L ( P − M + 1)
( M − m)( M − m − 1)L ( M − 2m + 1)
f 2 (m, M ) =
M ( M − 1)L ( M − m + 1)
Fig. 4 shows, when sub-key pool is constant, the relationship between the probabil-
ity and the number of keys stored in node.
Fig. 4 and Table 3 shows that sub-key pool is small, such as 100,200, the probabil-
ity of finding common keys between clusters will be very small; when the sub-key
pool increases, the probability significantly increases while the node storage has not
increased much; such as when sub-key pool is 400, the amount of keys stored in dif-
ferent clusters is approximately as 1.4 times as the number of inner cluster. According
to practical application, when we need a large amount of inner-cluster communica-
tion, we may properly increase both the number of sub-key pool and key ring stored
in nodes. Then any node who will be the cluster head has a big probability with oth-
ers’. Even if there is no shared key, we can establish pairwise key through the meth-
ods in 4.2. In actual application, the communication between the nodes in different
clusters is relatively small, if need, we can transmit through the cluster head.
In all, we can see that the scheme has a good performance. When ensure connec-
tivity, we largely reduce the number of key stored in node; and we can pre-configured
key flexibility according to the specific application requirement. When cannot find
shared keys between nodes, we can create pairwise key through a small amount of
communications.
300 H. Wang et al.
Table 3. In a certain probability, the number of keys stored in node in different clusters
p
0.92~0.95 0.95~0.98 0.98~0.99 Maximum p
|M|
100 -- -- -- 0.3895
200 -- -- -- 0.8696
300 38-41 42-49 50-72 0.9900
400 43-46 47-52 53-57 0.9997
1000 69-74 75-85 86-92 1.0000
2000 99-106 107-122 123-132 1.0000
3000 121-131 132-150 151-162 1.0000
4000 140-152 153-173 174-188 1.0000
1.1
1
0.9
0.5
0.4
0.3
0.2
0.1
0
0 20 40 60 80 100 120 140 160 180 200
size of m
Probability of sharing at least one key
Fig. 4. Probability of sharing at least one key in two nodes in different clusters
Efficient Pairwise Key Establishment Scheme 301
For the random key pre-distribution scheme, the network must establish all the com-
munication keys in shortest time after the distribution of the network nodes. If there is
neither leak of any key nor captured of nodes by attacker, the network is relatively
supposed to be secure and all keys are successfully achieved by each node in the fu-
ture communication. Otherwise, the message encrypted by the pairwise key will lose
confidentiality, and one node compromised by adversary will pose a threat to others
even the network.
The goal of this scheme is to allow sensor nodes to find a common secret key
safely with each of their neighbors after deployment. So every node broadcasts its
key’s identifiers at the network’s initial time to avoid disclose keys. We’ll analyze the
network’s resiliency when one or more nodes are captured and disclose all their keys.
Let the number of captured nodes be x. Since each node contains m keys, the prob-
ability that a given key has not been compromised is
x
⎛ m ⎞
⎜1 − | | ⎟ .
⎝ S ⎠
Therefore, the expected fraction [6,16] of total keys being compromised can be esti-
mated as:
x
⎛ m ⎞
p3 = 1 − ⎜1 − ⎟ (5)
⎝ | S |⎠
When |M|=400, |P|=10000, m=35 and m=55, the results and comparison with existing
key pre-distribution schemes are depicted in Fig. 5, which contains the in-cluster and
global resilience of our scheme, E-G scheme[9] that is same as our global resilience
and Chan-Perrig-Song scheme[3].The Fig. 5 show that our scheme substantially low-
ers the fraction of compromised communication after x nodes are compromised com-
pared with the Chan-Perrig-Song scheme in network-wide. We get that if the x nodes
were compromised in one cluster, the resilience of the network in that local area (in
the cluster) is lower than that of the entire network. We also realize that in reality,
these x nodes might be randomly distributed in the entire region, so the network will
have a higher probability of resilience.
The result is that our scheme is secure and efficient. Even the adversary eaves-
dropped the message, it only get the key’s ID instead of the keys themselves. For
another, other pairwise keys are calculated by the nodes themselves. Because the key
materials are transmitted through different intermediate nodes, the forwarding nodes
only know one of the materials, random value in this method; they can’t evaluate the
pairwise key. So, the key established in this method is absolutely secure compared
to using the secure channels in previous scheme; it is only known the two nodes
who’ll communicate with each other in future instead of any path-node or other third
node.
302 H. Wang et al.
After the deployment of all the nodes in the network, to discover the common keys,
each node of the schemes [9] [10] broadcast all of the keys or the key identifiers to
their neighbor nodes. In particular, when the network size is large, the node will store
a large number of keys in its subset. Then key information exchange process, not only
consumes a lot of energy but also wastes considerable radio bandwidth, resulting in
the link load overhead, collision opportunities to grow. Our method, according to
specific requirements, has the strongpoint that nodes only need to store the appropri-
ate number of keys in cluster, broadcast key identifier, to establish a certain number
of pairwise key. The remaining nodes that do not share common keys only take a
small amount of communication and calculation to establish the pairwise keys, which
can reduce the network traffic load and energy costs as much as possible. With our
method, the node transmits its key materials by using its neighbors who have the least
common keys to the destination, which means he/she will find the right intermediate
nodes in the largest probability. This path-materials computation and pairwire key
establishment approach extremely has less CPU power and time consumption than
previous two schemes. We can almost guarantee any two neighbor nodes to establish
a pairwise key and have a safe communication.
In our scheme, they use the very key whose identity is the smallest as pairwise key
in all key’s IDs when the two nodes find their common keys. Otherwise, they only
need four communications to get their key materials to calculate the key via XOR. This
extremely reduces the computation overload for resource constrained sensor nodes
especially their limited computing capability. Meanwhile, if they leak their pairwise
key, they can easily delete it and take the key of the second smallest ID in their key
tables as their later pairwire key while the two nodes have more than one common key.
This is much flexible and secure for updating the pairwise key. Therefore its commu-
nication and computational overhead are much lower than previous schemes.
6 Conclusion
In this paper, we describe our pre-distribution key management model in WSN and
propose a cluster-based key management scheme depending on the application
Efficient Pairwise Key Establishment Scheme 303
References
1. Wang, Y., Attebury, G., Ramamurthy, B.: A survey of security issues in wireless sensor
networks. IEEE Communications Surveys and Tutorials 8(2) (2006)
2. Su, Z., Lin, C., Feng, F.-J., Ren, F.-Y.: Key Management Schemes and Protocols for Wire-
less Sensor Networks. Journal of Software 18(5), 1218–1231 (2007)
3. Chan, H., Perrig, A., Song, D.: Random key pre-distribution schemes for sensor networks.
In: Proceedings of Symposium on Research in Security and Privacy, Oakland, USA, pp.
197–213 (2003)
4. Ren, K., Zeng, K., Lou, W.: A new approach for random key pre-distribution in large-scale
wireless sensor networks. Journal of Wireless Communication and Mobile Computing
(Special Issue on wireless Networks Security) 3(6), 307–318 (2006)
5. Du, W., Deng, J., Han, Y.: A pairwise key pre-distribution scheme for wireless sensor
networks. In: Proceedings of the 10th ACM Conference on Computer and Communica-
tions Security, Washington, DC, USA, pp. 42–51 (2003)
6. Liu, D., Ning, P.: Establishing pairwise keys in distributed sensor networks. In: Proceed-
ings of the 10th ACM Conference on Computer and Communications Security, Washing-
ton, DC, USA, pp. 52–61 (2003)
7. Liu, D., Ning, P.: Location-based pairwise key establishments for static sensor networks.
In: Proceedings of SASN 2003, Fairfax, USA, pp. 72–82 (2003)
8. Du, W., Deng, J., Han, Y.: A key management scheme for wireless sensor networks using
deployment knowledge. In: Proceedings of IEEE INFOCOM 2004, Hong Kong, China,
pp. 586–597 (2004)
9. Eschenauer, L., Gligor, V.D.: A key-management scheme for distributed sensor networks.
In: Proc. CCS 2002, pp. 41–47. ACM Press, New York (2002)
10. Bollobas, B., Fulton, W., Katok, A., et al.: Rand Graphs, 2nd edn., pp. 160–200. Cambridge
University Press, Cambridge (2001)
11. Blom, R.: An optimal class of symmetric key generation systems. In: Beth, T., Cot, N.,
Ingemarsson, I. (eds.) EUROCRYPT 1984. LNCS, vol. 209, pp. 335–338. Springer,
Heidelberg (1985)
12. Blundo, C., Santis, A.D., Herzberg, A., Kutten, S., Vaccaro, U., Yung, M.: Perfectly-
secure key distribution for dynamic conferences. In: Brickell, E.F. (ed.) CRYPTO 1992.
LNCS, vol. 740, pp. 471–486. Springer, Heidelberg (1993)
304 H. Wang et al.
13. Wang, Y., Ramamurthy, B., Xue, Y.: A Key Management Protocol for Wireless Sensor
Networks with Multiple Base Stations. In: IEEE International Conference on In Communi-
cations, ICC 2008, pp. 1625–1629 (2008)
14. Wang, Y., Ramamurthy, B., Zou, X.: KeyRev: An efficient key revocation scheme for
wireless sensor networks. In: Proc. ICC 2007, Glasgow, Scotland, UK (June 2007)
15. Traynor, P., Choi, H., Cao, G., Zhu, S., Porta, T.L.: Establishing pairwise keys in hetero-
geneous sensor networks. In: Proc. INFOCOM 2006 (2006)
16. Ma, C.-g., Shang, Z.-g., Wang, H.-q.: Domain-based key management for heterogeneous
wireless sensor networks. Journal on Communications 0l.30(5) (2009)
17. Chan, A.C.-F.: Distributed symmetric key management for mobile Ad Hoc networks. In:
Proceedings of the INFOCOM 2004, Twenty-third Annual Joint Conference of the IEEE
Computer and Communications Societies, vol. 4, pp. 2414–2424 (2004)
18. Liu, Z.-H., Ma, J.-F., Huang, Q.-P.: Domain-Based Key Management for Wireless Sensor
Networks. Chinese Journal of Computers 29(9) (September 2006)
Agent Based Approach of Routing Protocol Minimizing
the Number of Hops and Maintaining Connectivity of
Mobile Terminals Which Move One Area to the Other
1 Introduction
Wireless networks in today meaning are usually set up with a centralized access point
for each area of connectivity. The access point has knowledge of all devices in its area
and routing to them is done in a table driven manner. The future trend of mobility, it
becomes important to give users the possibility of finding service they need anywhere
in the network. Users would like to obtain access to services automatically, without
reconfiguring their system. Especially with mobile devices dynamic discovery of
services in other than home or corporate network and automatic network
configuration will be very useful.
An ad-hoc network is a special kind of network, where all of the nodes move in
time. So the topology of the network changes as the nodes are in the proximity of
each other. Ad-hoc networks are generally self-configuring, because of the concept of
no stable infrastructure takes a place. Each node in the network must be able to take
D. Taniar et al. (Eds.): ICCSA 2010, Part III, LNCS 6018, pp. 305–320, 2010.
© Springer-Verlag Berlin Heidelberg 2010
306 K. Arai and L. Sugiyanta
care of routing of the data. So the concept of centralized network with pre-defined
central routing tables could be applied only with the limited nodes changing and
coverage area. The coverage and connectivity of wireless mesh networks decreases as
the end-user becomes further from the other nodes. Multi-hop communications among
mesh nodes can enable long distance communication with acceptable performance
degradation. Multi-hop wireless network is network where communication between
two nodes may go through multiple consecutive wireless links without relying on any
existing, pre-configured network infrastructure or centralized control. These networks
are constrained by the interference of wireless communications, the finite battery
energy and the mobility of nodes.
However, wireless network throughput and multi-hop coverage extension are two
contradictory goals. On one hand, the multi-hop communications can extend the
coverage area. On the other hand, as the number of hops increases, the repeatedly
relayed traffic will exhaust the radio resource. The throughput will sharply degrade
due to the increase of collisions from a large number of users. Therefore, it becomes
an important issue to design a wireless network, so that the coverage of a wireless
network can be extended without sacrificing the overall throughput.
Beside throughput, the ability to transport data from a source to a destination is a
fundamental ability of wireless network. Delay is characterized by their lack of
connectivity, resulting in a lack of instantaneous end-to-end paths. This is due to
protocol routing trying to first establish a complete route and then, after the route has
been established, forward the actual data. Also, wireless clients are mobile nodes and
commonly run on batteries. Thus, power usage of mesh clients should be limited.
These natures of wireless ad hoc networks made them require many improvements
compared to wireless managed networks, before it ready to be implemented.
can connect to other networks by the mesh portals (MPPs), but this scheme is not
covered in this research. Multiple wireless meshes can also be connected by the MPP.
There will be several wireless areas, each with the nodes inside. Nodes are always
active moving with direction, speed, and distance change randomly. It is possible that
nodes able to move inside wireless area and leaving its home area. During this
activity, nodes always reconfigure their routing path allowing dynamically changing
of route selection. Interaction among nodes became rapidly increased.
When moving, nodes also allowed entering new wireless area, leaving home/
original area. In order to maintain connectivity along this movement, individual
components will need to discover and maintain awareness of surroundings nodes and
also to configure and adapt themselves in response to changing node’s condition.
Fig. 1. The network architecture for the IEEE 802.11s WLAN mesh network
There are special nodes called MPP. A MPP provides the ability to bridge wireless
networks over a mesh. MPP has capacity to find optimum path with minimum hops to
other nodes. Therefore MPP should support traffic propagation to both wireless areas
at where it located. Nodes at both areas able to sent and receive data outside its
home/original mesh and vice versa by using this MPP. It has consequences that nodes
must aware about both their neighbors and also MPP. As long as communication path
between source node and destination node is built inside wireless area, then MPP is
not involved. In this paper, this MPP node is called Agent. The agent node
reorganizes the network considering the available nodes.
Agents are located in such place to facilitate communication between wireless
areas and minimize the number of hops to achieve optimum throughput of mesh
clients which communicate each other. Through minimizing the number of hops
between source and destination, it is hoped that the optimum throughput between
source node and destination can be reached. The optimal tradeoff between network
capacity and throughput for wireless mesh networks need to be investigated.
Simulation programming approach is built to determine the optimal deployment
parameters. This simulation runs with large number of nodes, with high dynamic
parameters and considering mobile devices resources (battery power, bandwidth) and
wireless environment constraints.
Simulation is split into two steps. First has to cover movement among nodes inside
area, not move to another area. Second, simulation covers node’s activities that
include communication among nodes which also move to other areas. During the first
step, route discovery play important role. Source nodes that do route discovery flood
the network with a route request. Its neighbors propagate this route request in the
same manner until it reaches the destination node or a node that already has a route to
the destination. The data analyzed in this paper is intended to this first step.
3 Algorithm
3.1 Establishing the Initial Topology
Initialization topology. Nodes discover themselves and use their maximum trans-
mission power to build the initial topology. We consider the case where all nodes in
the network are similar, i.e., assuming an homogeneous infrastructure. Let us assume a
square area of side l, in which n nodes are deployed uniformly, distributed at random
position. This deployment produces a connected topology under some assumptions;
sometimes a completely connected topology is built. From the simulation, a large
connected component is built quickly using a communications radius considerably
smaller than the radius needed to have the entire network connected.
Fig. 3. 200 nodes distributed randomly in the area of l2, some nodes become backbone for
others. Captured at initial time (t= 0).
3.4 Energy
Energy is power kept in each node. It was assumed that the radio dissipates Eelec =
50 nJ/bit to run the transmitter or receiver circuitry and ϵamp = 100 pJ/bit/m2 for the
transmit amplifier. Thus, to transmit a k-bit message a distance d using this radio
model, the radio expends:
, (1)
and to receive this message, the radio expends:
(2)
The energy model included in simulation was based on the following formulas, taken
from [9]:
E E ϵ . πr (3)
E E (4)
The energy behaviors of nodes are defined as:
• During the idle time, a node does not spend energy. Even though this
assumption has been proven untrue because being idle might be as costly as
receiving data, this is still an assumption that can be done in most experiments,
since the most important factor is the overhead in terms of message exchange
and its associated cost.
Agent Based Approach of Routing Protocol Minimizing the Number of Hops 311
• The nodes are assumed to have one radio for general messages. The main
radio is used in all operations when the node is in active mode, and to send and
receive control packets. When this radio is turned off, then no messages will
be received and no energy will be used.
• Energy distribution among nodes can either be constant value, normally
distributed, Poissonly distributed, or uniformly distributed.
• The communication range of the nodes is a perfect symmetric unit disk. If dx,y
≤ rx → x and y can see each other.
• A constant bit error rate (BER) is defined for the whole network. It is a simply
implementation of an error model. Whenever a packet is going to be sent, a
random number is generated and compared to the message error rate (that
depends on the size of the message). If the random number is greater, the
message is received, otherwise it is lost. The default value for the BER is 0,
which means there is no packet loss.
• To increase the randomization of the simulation process, simulation introduces
some delay on some common processes in the network, like message
transmission delay, processing delay, time out, etc. so each instance of a
simulation, would produce different results.
• Simulation manages homogeneous networks, in which all nodes have the same
characteristics. The parameters that can be defined in a homogeneous family
of nodes are number of nodes, communication radius, and size of the area of
deployment, position distribution, and energy distribution.
9.799 9.623
10
9.575 8.161
9.323 9.214
8.933 7.551
8
6.800
7.807 7.814
6.145
7.016
6 5.510
5.092
4.417
5.360
4 4.748 4.686 3.516
3.874
3.235
2 2.553
2.340
10 20 30 40 50 60 70 80 90 100
-1.119
-2
# Nodes
16
14
12
10
0
106
113
120
127
134
141
148
155
162
169
176
183
190
197
204
211
218
225
232
239
246
253
260
267
274
281
288
295
302
309
316
323
330
337
344
351
358
365
372
379
386
393
400
1
8
15
22
29
36
43
50
57
64
71
78
85
92
99
(a)
30
25
20
15
10
0
100
109
118
127
136
145
154
163
172
181
190
199
208
217
226
235
244
253
262
271
280
289
298
307
316
325
334
343
352
361
370
379
388
397
1
10
19
28
37
46
55
64
73
82
91
(b)
30
25
20
15
10
0
1
100
109
118
127
136
145
154
163
172
181
190
199
208
217
226
235
244
253
262
271
280
289
298
307
316
325
334
343
352
361
370
379
388
397
10
19
28
37
46
55
64
73
82
91
(c)
Fig. 5. The faster node is moving, the more Energy is be used
Agent Based Approach of Routing Protocol Minimizing the Number of Hops 313
Each node was given energy of 1000 at the beginning of simulation. Simulation
executed repeatly with number of initiator nodes increment by 10. As number of
initiator node increase, the total energy remained on nodes is decreased. Simulation
with 4 initiator nodes gave result negative average value at # Nodes= 80. It was to
show that some nodes were already dead/OFF.
With node always moving, simulation shows that energy used during cycles is
affected with its speed. More energy is used by node when node moves faster. Fig. 5
below illustrated the effect of speed to energy used by initiator node.
Used Energy
20
18
16
Energy used *100
14
12
10
0
1 13 25 37 49 61 73 85 97 109 121 133 145 157 169 181 193 205 217 229 241 253 265 277 289 301 313 325 337 349 361 373 385 397
Simulation cycle
Energy used (2 In) Energy used (1 In) Energy used (1 In) Energy used (1 In) Energy used (1 In)
# Overheads Packets
20,000
Hundreds
14,000
10,000
9,642 9,592
9,057 9,002
8,512 8,462
8,000
7,520 7,476
7,029 6,956
10 20 30 40 50 60 70 80 90 100
# Nodes
It is clear from Fig. 6, the increasing number of initiator nodes caused nodes to
route a large number of messages. Thus these nodes will die out quickly, causing the
energy required to get the remaining messages to increase and more nodes to die. This
will create a cascading effect that will shorten system lifetime. In addition, as nodes
close to the initiator die, then the environment is no longer being monitored. To prove
this point, we ran simulations using the random 10 – 100 node shown in Fig. 5. After
completed cycle, several nodes were considered dead.
Using Equations 3 and 4 and the random node from 10 until 100, we simulated
transmission of first data from initiator to every node (communication range 100 m).
Fig. 4 showed the total energy remained on nodes. Topology was changed all the time
cycle. With transmission energy is on the same order as receive energy, it can be seen
that the energy-efficient to use depends on the network topology.
25
20
15
10
0
10
19
28
37
46
55
64
73
82
91
100
109
118
127
136
145
154
163
172
181
190
199
208
217
226
235
244
253
262
271
280
289
298
307
316
325
334
343
352
361
370
379
388
397
1
Simulation cycle
(a)
20
18
16
14
12
10
8
6
4
2
0
10
19
28
37
46
55
64
73
82
91
1
100
109
118
127
136
145
154
163
172
181
190
199
208
217
226
235
244
253
262
271
280
289
298
307
316
325
334
343
352
361
370
379
388
397
Simulation cycle
(b)
16
14
12
10
0
10
19
28
37
46
55
64
73
82
91
1
100
109
118
127
136
145
154
163
172
181
190
199
208
217
226
235
244
253
262
271
280
289
298
307
316
325
334
343
352
361
370
379
388
397
Simulation cycle
(c)
The accumulated transmission delay and increasing number of hops among nodes that
lowered the throughput can be seen from the previous research results. In [7],
variables of evaluation are the number of hops, length of data transmission, and round
trip time between nodes. The data obtained is shown in Fig. 8 below.
ms RTT delay
#HOP affected the #TCP response time #HOP affected TCP throughput
Fig. 9. The effect of multi-hop transmission in Wireless Mesh Network [7]
5 Related Works
Many researchers have evaluated general problems of performance of multi-hop
wireless mesh network.
One branch of related work is concerned with the evaluation of wireless mesh
product considering their applicability into the enterprise network. A second branch of
work considers the evaluation of node behavior that composes multi-hop radio network.
Nemoto Nozomu [7] considered to introduce the technical review of wireless
mesh network products IEEE802.11 standard with fixed wireless mesh network
node. In terms of performance, the research will evaluate the overwhelming use of
outdoors Muni-WiFi with a view to applying the same technology in the corporate
network.
There are many products and technology for wireless LAN. The difference
technology among products includes route, control technology and architecture.
Wireless LAN is not competing with legacy network, otherwise it can be considered as
complement. In this research, wireless LAN and wireless mesh network is compared to
determine the level of performance. Table 1 shows the main contents of evaluation.
Agent Based Approach of Routing Protocol Minimizing the Number of Hops 317
6.1 Agents
The agent nodes are special nodes that are included to receive the information from
all active nodes in both neighborhood wireless areas. They served as bridges and used
to transport the sensed data to its final destination somewhere between the wireless
network areas. They are also in charge of initiating, executing, and/or controlling the
topology construction and maintenance protocols, routing protocols, etc.
Despite the fact that in real life the hardware configuration of the agent nodes may
different compared to a regular wireless device, simulation uses the same data
structure to model both nodes, with the main difference that agent nodes are
considered to have more amount of energy. Agents in the simulation have two main
functions: as service agents and delivery agents.
1. Service agents: takes a place relay between areas which is advertised with service
request from source nodes. They could create paths and provide optimum hop/path
to destination node.
2. Delivery agents: collectors of information received from nodes in both areas.
7 Conclusion
In this paper, we evaluate topology of wireless network with multi-hop routing. One
of the main processes for building hop tree is energy-efficient. The algorithm is done
Agent Based Approach of Routing Protocol Minimizing the Number of Hops 319
Acknowledgments. The authors would like to thank the anonymous reviewers for the
helpful comments and suggestions.
References
1. Ettus, M.: System Capacity, Latency, and Power Consumption in Multihop-routed SS-
CDMA Wireless Networks. In: Radio and Wireless Conference (RAWCON 1998), August
1998, pp. 55–58 (1998)
2. Rodoplu, V., Meng, T.H.: Minimum energy mobile wireless networks. IEEE Journal on
Selected Areas in Communications 17(8), 1333–1344 (1999)
3. Lin, X., Stojmenovic, I.: Power-Aware Routing in Ad Hoc Wireless Networks. In: SITE,
University of Ottawa, TR-98-11 (Dec. 1998)
4. Meng, T., Volkan, R.: Distributed Network Protocols for Wireless Communication. In:
Proc. IEEEE ISCAS (May 1998)
5. Manfredi, V., Hancock, R., Kurose, J.: Robust routing for dynamicmanets. In: ACITA
2008 (2008)
6. Ye, Z., Krishnamurthy, S., Tripathi, S.: A framework for reliable routing inmobile ad hoc
networks. In: Proceedings of IEEE INFOCOM (2003)
7. Nozomu, N.: Consideration and Evaluation of Wireless Mesh Network. Nomura Research
Institute (NRI) Pacific Advanced Technologies Eng., 70–85 (2006)
8. Jarábek, B.: Advanced Network Approaches for Wireless Environment. In: Bieliková, M.
(ed.) IIT. SRC 2005, April 27, pp. 108–115 (2005)
9. Heinzelman, W., Chandrakasan, A., Balakrishnan, H.: Energy-efficient communication
protocol for wireless microsensor networks. In: Proceedings of the 33rd International
Conference on System Sciences (HICSS), pp. 1–10 (2000)
320 K. Arai and L. Sugiyanta
10. Oh, S.Y., Gerla, M.: Robust MANET Routing using Adaptive Path Redundancy and
Coding. In: COMSNET 2009, Bangalore, India, January 10 (2009)
11. The Dynamic Source Routing Protocol for Mobile Ad Hoc Networks (DSR), RFC Draft
(July 2004),
http://www.ietf.org/internet-drafts/draft-ietf-manet-dsr-10.txt
12. Kyushu University, Kyushu Institute of Technology.: Developing Multi-Hop Wireless
Communication Technology (WiRem), Press Release Project, March 17 (2008)
Ensuring Basic Security and Preventing Replay Attack
in a Query Processing Application Domain in WSN
Amrita Ghosal1, Subir Halder1, Sanjib Sur2, Avishek Dan2, and Sipra DasBit2
1
Dept. of Comp. Sc. & Engg, Dr. B. C. Roy Engineering College, Durgapur, India
[email protected], [email protected]
2
Dept. of Comp. Sc. & Tech., Bengal Engineering and Science University, Shibpur, India
{sanjib.sur11,avishekdan}@gmail.com, [email protected]
1 Introduction
The advent of efficient short range radio communication and advances in miniaturiza-
tion of computing devices have given rise to strong interest in wireless sensor net-
works (WSNs) [1]. A wireless sensor network (WSN) consists of a large number of
small, battery-powered [2], [3] wireless sensor nodes. The nodes are distributed in ad-
hoc fashion and process sensing tasks. A sensor network may be deployed in hostile
environments where there can be malicious attackers. On the other hand, the WSNs
are used in critical application domain e.g. defense, medical instrument monitoring.
Therefore, securing the activities of WSNs is of utmost importance.
These WSNs are prone to attacks [4], [5] due to their nature of deployment in an
unattended environment and also due to the broadcast nature of the medium. One
such attack is Denial of service (DoS) attack [6] that possesses a great threat to wire-
less sensor network environment. One form of DoS attack attempts to disrupt the
network service, may be by blocking communication between nodes. The other form
of DoS attack is the path-based DoS (PDoS) attack where the network is flooded with
bogus packets along a multi-hop data delivery path [7]. The packets are continuously
D. Taniar et al. (Eds.): ICCSA 2010, Part III, LNCS 6018, pp. 321–335, 2010.
© Springer-Verlag Berlin Heidelberg 2010
322 A. Ghosal et al.
replayed back to the sink node. This attack is known as replay attack. However, the
objective of both types of such attacks is to eliminate or diminish the network per-
formance and thereby hamper the working of the whole system. In replay attack,
because of the broadcast nature of sensor networks, an adversary eavesdrops on the
traffic, injects new messages and replays/changes old messages.
Many works are so far reported towards the solution of various forms of DoS at-
tack. Deng et al. [7] have considered a type of DoS attack along a multihop data de-
livery path. As WSNs are generally tree structured, so an attack on the nodes of a path
also affects the branches connected to that path. A one way hash chain mechanism has
been proposed by the authors to prevent such path based DoS attacks and protect end
to end communication along a multihop data delivery path. Here an OHC (One way
Hash Chain) number is added with each message packet leading to an extra overhead
of 8 bytes per packet. These 8 bytes of additional overhead is a major constraint for a
resource constrained sensor mode.
Authors in [8] have proposed an authentication protocol capable of resistance
against replay attacks and other DoS attacks. But the scheme uses symmetric keys
where keys are shared by sensor nodes and therefore a compromised node can send
forged messages which may possess a great security threat to the network.
Perrig et al. [9] have used RC5 algorithm for encrypting messages which means a
lot of computations have to be done. Here the µ TESLA protocol is used for secure
broadcast of messages and an 8-byte Message Authentication Code (MAC) is used by
each node for verification along the communication path. However, exchange of huge
authentication information is a real bottleneck for the resource-constrained WSN.
Moreover a node must have prior knowledge of all the nodes along its path.
The authors in [10] have devised a mechanism to prevent replay attacks. Here it is
considered that the packets used for time-synchronization of any two nodes are re-
played. A receiver-receiver model for time synchronization has been considered
where a reference node broadcasts beacon message to all its neighbours. Based on the
arrival time of the beacon message the receiving nodes adjust their clocks. A time
offset is calculated based on the difference between the recording times of the beacon
message by these receiving nodes. The time offsets are exchanged between the receiv-
ing nodes to calculate a threshold value which is the difference between the two time
offsets of two nodes. But the arrival of a beacon message at a node can be delayed by
certain factors leading to gross errors while deriving the time required for synchroni-
zation between the nodes. Moreover, by not using global time synchronization model
a large overhead has been introduced as huge number of time offsets need to be com-
puted by the nodes.
Dong et al. [11] have proposed the use of hash chains in their work where each
node combines the hash value with its own node-id and forwards this combined value
to its next higher hop count node. The receiving node is able to detect replay attacks
by observing the combined value of node-id and hash value. But the computation of
all these values by nodes takes significant amount of time.
Soroush et al. [12] have developed a scheme to defend replay attacks where a
monotonically increasing counter is maintained to keep track of old replayed mes-
sages. But here each node maintains a counter to store the timing information of all
other nodes which requires a large amount of memory leading to a major bottleneck
for memory- constrained sensor nodes.
Ensuring Basic Security and Preventing Replay Attack 323
In this paper we propose a secured query processing scheme in WSN with a target
to build a security mechanism from within a query-driven application environment.
The proposed security mechanism gives a solution of replay attack while ensuring
authentication, data integrity and data freshness.
The rest of the paper is organized as follows. In section 2 system model along with a
brief description of the present work is given. A detailed description of the scheme is
given in section 3. Section 4 gives the performance evaluation of the proposed scheme
and also the comparative study between the proposed scheme and another scheme.
Concluding remarks and future scope has been stated in section 5.
2 System Model
An attack [6] is defined as an attempt to gain unauthorized access to a service, re-
source, or information, or the attempt to compromise integrity, availability, or confi-
dentiality. DoS is one type of attack which hinders communication among the sensor
nodes. The present work considers replay attack, which is one form of path based
DoS (PDoS) attack. If there is a PDoS attack, the network gets flooded with bogus
packets along a multi-hop delivery path [7]. The occurrence of bogus packets in the
network is due to the replaying of packets by adversaries leading to replay attack. So
PDoS attacks in the network may result in replay attacks.
The system model in the present work considers clustered network architecture
based on LEACH protocol [13]. In this architecture nodes organize themselves into
local clusters with one node acting as the cluster head (CH). LEACH performs local
data fusion to reduce the amount of data sent from the clusters to the base station/
sink. Once all the nodes are organized into clusters, each CH assigns a time slot
(TDMA schedule) for the member nodes in its cluster. The member nodes sense data
and transmit data to the cluster head nodes that are located at one hop distance away
from them. The cluster head nodes transmit the same to the base station after receiv-
ing data from their member nodes.
The present work considers a query-driven application platform where base station
generates query messages and broadcasts the query. Cluster heads receive the broad-
casts query and start the registration process to authenticate their respective member
nodes. Once the registration phase is over, the CHs forward the query messages to
their members. Depending on the nature of queries, specific member nodes send re-
sponse-packets.
The objective of the proposed scheme named as Secured Query Processing Scheme
(SQPS) is to ensure basic security in general and prevent replay attack in particular
while working for a query-driven application. The essential properties [14] of a WSN
required for maintaining basic security within the network are:
• Authentication- A sensor network must be able to differentiate between data
coming from reliable sources and those coming from adversaries.
• Data Integrity- The received data should not be tampered during communica-
tion between the sender and the receiver.
• Data Freshness- The data received at a particular time should be current data
and not old data which may be replayed by adversaries.
324 A. Ghosal et al.
3 The Scheme
In this secured query processing scheme as shown in figure 1, periodically queries are
broadcast from base station. The process begins as soon as the cluster heads receive
queries from base station. The scheme has two phases-
• Registration Phase
• Query Response Phase
This phase is used for registering member nodes by the respective cluster head nodes.
The objective of this phase is to register only the authenticated nodes. Moreover the
phase provides a mechanism that will help a CH to ensure data integrity and data
freshness during query processing phase. Upon receiving a query from the base sta-
tion, cluster heads initially broadcast registration packet.
The registration packet contains 8 bits including ‘0’ as MSB. This MSB differenti-
ates between a registration packet and a query packet sent by the cluster heads. As
mentioned in section 2, each member node is allotted a particular time slot which is
used for sending a registration response packet to the cluster head corresponding to
the registration packet and also for receiving identification (node-id) for the member
node from the cluster head for continuing the query processing session. On receiving
the registration packet, the respective member nodes decide the number of bits to be
shifted and accordingly left shift the bits of registration packet. Then the nodes gener-
ate registration response packet including left-shifted registration packet (8 bit), num-
ber of bits left shifted (3 bit), and present time-stamp (12 bit). The key used here is the
number of left shifted bits. Symmetric key cryptography is used as the same key is
used for encryption and decryption of the data packets. Symmetric key cryptography
is beneficial for sensor networks as less computation has to be done and memory
requirement is also minimized. Here, though the key is being sent along with the
registration packet it will not be possible for any adversary who captures or eaves-
drops the contents of the packet to decipher which bits in the packet refer to the key.
Moreover in traditional networks the adversary performs some computations to obtain
the key used in the network; this is not possible in case of sensor networks as the
adversary nodes are also equipped with less computational power. The present time-
stamp indicates the time when a member node sends a reply packet in response to the
registration packet. The time-stamp is represented in minutes and seconds.
Immediately after receiving a registration response packet from one of its member
node, the CH right shifts the reply packet designated number of times. The number of
bits to be right shifted is same as the number of bits the registration packet is left
shifted and it is provided as control data in the registration response packet itself. If
the CH can retrieve the original registration packet after right shifting the received
registration response packet, the member node is considered as an authenticated and
registered node. Upon authenticating a member node, the CH generates a node-id and
sends it to the member node at the same time slot of authenticating and registering the
member node. Once the registration of a member node is successful, the CH stores
this information.
Ensuring Basic Security and Preventing Replay Attack 325
Node authentication is one of the major parameters of network security. The registra-
tion phase of the present scheme is able to ensure this part of security.
In the event of query processing, the base station broadcasts query. Upon receiving
the query, a CH starts registration phase. Once the registration phase is over (section
3.1), the CH broadcasts the query for its member nodes. Depending on the query,
specific member nodes send response and cluster head nodes forward the responses to
base station. For example, if the query is related to temperature, then member nodes
responsible for sensing temperature provide response to the query.
Query packet is of 8 bits. The MSB bit of the query packet is always 1, to distin-
guish between registration packet and query packet. So, 128 different queries can be
generated with the remaining 7 bits of the query packet.
present time stamps when the member node is sending the response packet for the
query and previous time stamp of last communication (packet exchange) between
member node and CH respectively. To start with, TMN _ present of the 3-tuple data
stored by the member node at registration phase replaces TMN _ previous.
The entire scheme is illustrated with the help of an activity diagram shown in Figure
1.The diagram shows how a cluster with n number of nodes and one cluster head
works. To be more specific, it shows communication among different components of
a wireless sensor network and control flow among various computational modules
within the components of the network. Broadcast communication from cluster head is
shown by lightning symbol whereas unicast communication between cluster head and
member nodes and control flows are shown by solid lines with arrowhead. All com-
munication symbols are labeled to make understand the content of the packet along
with the timestamp at which it is being sent. For example, one broadcast communica-
tion labeled by RBt means registration packet has been broadcast at time t. Time-
stamp helps to know the steps of operations visibly.
4 Performance Evaluation
The effectiveness of the proposed security scheme reported in the earlier section is
evaluated through simulation.
Simulation is performed using MATLAB (version 7.1). We consider 500 nodes in the
network and number of malicious nodes is varied from 25 to 100.
Performance of the scheme is evaluated based on the following two metrics:
Authentication Rate – Number of authenticated nodes / total number of nodes in the
network.
Data freshness (%) – (Number of received packets containing current data / total num-
ber of packets sent) x 100.
The relevant parameters and their associated values are listed below in Table 1-
Parameters Value
Initial Energy ( Einitial ) 2J
Network area (user input) 25m × 25m to 100 m × 100 m
Communication range of sensor ( R c ) 160 m
Sensing range of sensor ( R s ) 80 m
In presence of replay attack, data freshness is affected. In other words, if replay at-
tack can be considered as a cause, data freshness is an effect. Therefore, the metric,
data freshness is measured to cover analysis on both the security parameters replay
attack and data freshness. Authentication rate is computed to cover another security
parameter i.e. authentication. No separate experiment is carried out to measure data
integrity as violation of data integrity due to presence of malicious node does not
arise.
Authentication rate is measured and plotted with time in Figure 2 for varying number
of malicious nodes. Two sets of experiments are conducted to compute authentication
rate.
Ensuring Basic Security and Preventing Replay Attack 331
1
Authentication rate
0.98
0.96
0.94
Registration without attack
0.92
Registration with attck
0.9
0 50 100 150 200 250 300
TDMA time slots (sec)
1
Authentication rate
0.95
0.9
0.85
0.8
Registration without attack
0.75
Registration with attack
0.7
0 50 100 150 200 250 300
TDMA time slots (sec)
1
Authentication rate
0.8
0.6
0.4
Registration without attack
0.2
Registration with attack
0
0 50 100 150 200 250 300
TDMA time slots (sec)
In one set of experiment, we consider that malicious nodes are present and attempt to
participate for sending data but stop participation once it is refused to do so. Results are
plotted in Figure 2 for varying number of malicious nodes and designated as ‘without
attack’. In the other set of experiment, malicious nodes attempt to send data repeatedly
resulting in replay attack. Results of the same are plotted and designated as ‘with attack’.
We observe that in all the cases of Figure 2 ((a), (b), (c)) results ‘with attack’ show
a fall of authentication rate compared to the results of ‘without attack’. This signifies
presence of replay attack from malicious nodes. Further, if we compare results of all
the plots ((a), (b), (c)), it is observed that authentication rate decreases with increase
in number of malicious nodes.
Figure 3 shows percentage of data freshness over a period of time in presence of
malicious nodes. Results in presence of 25 malicious nodes show that average data
freshness is near about 96% whereas its values are 88% and 81% for 50 and 100 ma-
licious nodes respectively. The results indicate that data freshness is inversely related
to the number of malicious nodes present in the network.
1
Data Freshness (%)
0.8
0.6
0.4
Due to unavailability of an existing suitable work so that our scheme can be compared
based on all the security parameters considered here, we have chosen a work [11] on
secured routing scheme to compare packet loss and data packet rejection ratio as com-
parison metrics as defined below.
Packet loss (%) – Total number of received packets by a CH / total number of packets
sent by the member nodes.
Packet rejection ratio (%) – Total number of received packets by a CH containing
tampered data / total number of packets sent by the member nodes.
Packet loss and packet rejection ratio (%) are computed and plotted for varying num-
ber of malicious nodes in Figure 4 and Figure 5 respectively.
Ensuring Basic Security and Preventing Replay Attack 333
60
50
Packet loss (%)
40
30
20
10 M VCS
SQPS
0
0 20 40 60 80 100 120
Number of Malicious nodes
The figures also show the results for secured routing scheme named as MVCS
(mitigating attack against virtual coordinate system) [11]. We observe that packet loss
(Figure 4) increases with the increase of malicious nodes for both the schemes. How-
ever, in our scheme (SQPS), packet loss is about 16% less than MVCS for up to 25
numbers of malicious nodes. For 25 to 85 numbers of malicious nodes packet loss in
SQPS is about 20% less than MVCS and this onwards packet loss is about 16% less
in SQPS. Summarily, it can be said that packet loss in SQPS is less than MVCS for all
sets of values of malicious nodes.
60
Packet Rejection Ratio (%)
50
40
30
20
10 M VCS
SQPS
0
0 20 40 60 80 100 120
Number of Malicious nodes
Further we observe that packet rejection ratio (Figure 5) increases with the increase
of malicious nodes for both the schemes. But in our scheme (SQPS), packet rejection
is about 10% less than MVCS for up to 25 malicious nodes. For 25 to 100 malicious
nodes packet rejection ratio in SQPS is 20% less than MVCS and this trend continues.
So it can be inferred that packet rejection in SQPS is less than MVCS for all sets of
values of malicious nodes. As all the member nodes’ identities are verified through
registration phase, there is very little chance that a malicious node is able to steal the
identity of a legitimate node and passes through registration phase. That is why SQPS
packet rejection ratio is lower than MVCS.
5 Conclusion
In this paper, we have proposed a scheme to defend replay attacks on nodes of WSN
as well as preserve the essential basic security properties such as authentication, data
integrity and data freshness of such a network. The scheme is designed in such a
manner that no malicious node can take part in actual query processing thereby ensur-
ing authentication.
Moreover, as there is no participating malicious node, violation of data integrity
due to attack has been eliminated. However, a malicious node can attempt to partici-
pate stealing some time slots due to which there may be some packet loss. The merit
of the scheme lies on the fact that simple symmetric key cryptography has been used
to maintain security making the solution very lightweight. We have substantiated our
claims by simulating the scheme in presence of attacks. Finally the scheme is com-
pared with one of the existing routing schemes considering packet loss and data
packet rejection ratio as comparison metrics. Results show our scheme outperforms
the existing one.
As a future extension, the scheme may be made more realistic considering cluster
head nodes are also vulnerable to attack. Further enhancement may be done to make it
applicable for continuous data-flow application domain as well.
References
1. Akyildiz, I.F., Su, W., Sankarasubramaniam, Y., Cayirci, E.: A survey on sensor networks.
IEEE Communications Magazine 40(8), 102–114 (2002)
2. DasBit, S., Ragupathy, R.: Routing in MANET and Sensor Network- A 3D position based
approach. Journal of Foundation of Computing and Decision Sciences 33(3), 211–239
(2008)
3. Halder, S., Ghosal, A., Sur, S., Dan, A., DasBit, S.: A Lifetime Enhancing Node Deploy-
ment Strategy in WSN. In: Lee, Y.-h., et al. (eds.) FGIT 2009. LNCS, vol. 5899, pp. 296–
308. Springer, Heidelberg (2009)
4. Ghosal, A., Halder, S., DasBit, S.: A Scheme to tolerate Jamming in multiple nodes in
Wireless Sensor Networks. In: Proceedings of Wireless VITAE, pp. 948–951. IEEE Press,
Los Alamitos (2009)
5. Ghosal, A., Halder, S., Chatterjee, S., Sen, J., DasBit, S.: Estimating delay in a data for-
warding scheme for defending jamming attack in wireless sensor network. In: Proceedings
of 3rd International Conference NGMAST, pp. 351–356. IEEE CS Press, Los Alamitos
(2009)
Ensuring Basic Security and Preventing Replay Attack 335
6. Wood, A.D., Stankovic, J.A.: Denial of Service in Sensor Networks. IEEE Com-
puter 35(10), 54–62 (2002)
7. Deng, J., Han, R., Mishra, S.: Limiting DoS attacks during multihop data delivery in wire-
less sensor networks. Journal of Security and Networks 1(3/4), 167–178 (2006)
8. Liu, D., Ning, P.: Efficient distribution of key chain commitments for broadcast authenti-
cation in distributed sensor networks. In: Proceedings of 10th annual Network and Distrib-
uted System Security Symposium, pp. 263–276 (2003)
9. Perrig, A., Szewczyk, R., Wen, V., Culler, D., Tygar, J.: SPINS: Security Protocols for
Sensor Networks. Wireless Networks Journal (WINET) 8(5), 521–534 (2002)
10. Song, H., Zhu, S., Cao, G.: Attack-Resilient Time Synchronization for Wireless Sensor
Network. In: Proceedings of IEEE International Conference on Mobile Adhoc and Sensor
Systems Conference, vol. (7-7), pp. 772–779 (2005)
11. Dong, J., Ackermann, K.E., Bavar, B., Nita-Rotaru, C.: Mitigating Attacks against Virtual
Coordinate Based Routing in Wireless Sensor Networks. In: Proceedings of 1st ACM con-
ference on Wireless Network Security, pp. 89–99 (2008)
12. Soroush, H., Salajegheh, M., Dimitriou, T.: Providing Transparent Security Services to
Sensor Networks. In: Proceedings of IEEE International Conference on Communications,
pp. 3431–3436 (2007)
13. Heinzelman, W.R., Chandrakasan, A., Balakrishnan, H.: Energy Efficient Communication
protocol for Wireless Microsensor Networks. In: Proceedings of the 33rd Hawaii Interna-
tional Conference on System Sciences, vol. 2, pp. 8020–8029 (2000)
14. Zia, T., Zomaya, A.: Security Issues in Wireless Sensor Networks. In: Proceedings of In-
ternational Conference on Systems and Network Communications, p. 40 (2006)
A Review of Redundancy Elimination Protocols for
Wireless Sensor Networks
1 Introduction
Coverage is one of the most important metrics for evaluating the quality of service
(QoS) in wireless sensor networks. A region in a sensor field is considered as a cov-
ered region if every point in this area is in the sensing radius of one or more sensors.
However, it is important to consider how efficiently the sensor field is covered by the
deployed sensors.
Two different aspects of the coverage problem have been discussed in the litera-
ture. The first aspect considers the efficiency of coverage of a deployment region in
terms of the percentage of the total area covered. The second aspect of the coverage
problem is to explore how to reduce redundancy in the coverage. The same area of a
sensor field may be covered by many sensor nodes in a dense deployment. The aim is
to ensure that no more than a few nodes cover a particular area at any time while the
sensor field is fully covered, increasing sensing efficiency. In this paper, we review
the proposed protocols from this sensing efficiency point of view.
We consider three main fields for redundancy elimination protocols. These fields
are node scheduling schemes [1, 2, 3, 4, 5, 6, 7, 8, 9], approaches for solving the cov-
erage problem (redundancy point of view) [10, 11, 12] and solutions for efficient
routing [13, 14]. Although routing protocols proposed in [13, 14] are targeted for
wireless ad-hoc networks. They have the potential to be deployed in wireless sensor
networks.
Redundancy elimination protocols can also be classified into two major groups:
location aware or location free schemes. The former utilizes the precise or relative
D. Taniar et al. (Eds.): ICCSA 2010, Part III, LNCS 6018, pp. 336–351, 2010.
© Springer-Verlag Berlin Heidelberg 2010
A Review of Redundancy Elimination Protocols for Wireless Sensor Networks 337
location information of sensor nodes. The latter does not consider any geographical
position information of nodes.
In the following, these protocols are reviewed and a table is presented in the dis-
cussion section to summarize the attributes of each protocol.
High redundancy in wireless networks results in several paths between source and
sink. GAF [14] discovers a set of nodes that are equivalent in terms of routing. It then
minimizes the number of paths by turning off the redundant intermediate nodes. A
GAF simulation study confirms that in the presence of representative active nodes, a
stable level of routing fidelity is preserved [14].
The first phase in GAF is to organize nodes in virtual grids. The monitoring area is
divided into a number of grids. Each node is associated with one particular grid based
on its location information. Nodes that are equivalent from the routing point of view
are linked to the same grid. Two grids are adjacent when each node in one grid can
communicate with every node in another gird. Grid size is determined based on the
communication range of nodes.
After assigning a grid to each node, the main phase of GAF begins which determines
and maintains the active node of each grid. In fact, each node has three different states.
These are sleeping, discovery and active. Initially, all nodes are in the discovery mode.
When the discovery timer of a node expires, it broadcasts a discovery message that
338 B. Pazand, A. Datta, and R. Cardell-Oliver
contains the grid id and node id of this node (with this approach equivalent nodes find
each other). If a discovery node does not receive any response from the active node of
its grid, it becomes the active node of this grid. Otherwise, it goes into the sleep state.
The timer of sleeping nodes is set to a random number from the range of [enlt/2, enlt].
Here, enlt is the expected lifetime of the active node. Upon the expiration of this timer, a
sleeping node changes its state to discovery to detect the possible lack of an active node
in its grid. GAF also provides fair energy consumption among sensor nodes. When a
node becomes active, it sets a timer to a pre-defined active duration. When this timer
fires, the active node returns to the discovery state. Now, discovery nodes that recently
have become awake, have the chance to replace the active node of their grid.
GAF has some advantages. First, it sits over any underlying ad hoc routing proto-
col. Second, it dynamically changes the set of active nodes which provides load bal-
ancing. Third, it supports mobility of nodes. However, there are some drawbacks as
well. Although it has been shown in [14] that the percentage of packet loss and route
latency is as good as AODV and DSR routing protocols, it has been investigated un-
der the relatively small number of traffic nodes. GAF does not propose any solution to
preserve the capacity of network after eliminating the redundancy. Therefore, the
performance of GAF can decrease significantly under high data traffic.
is its assumption of having the transmission range at least twice the sensing range. This
assumption is necessary to determine the connectivity tier successfully [3].
Authors in [10] propose a solution for the coverage problem. It determines whether
the target area is adequately k-covered or not. Every point in the sensor field should
be covered by at least k sensors. They further propose some applications such as node
scheduling and finding insufficiently covered areas based on their approach.
The main idea of the solution is to determine whether the perimeter of each sensor is
k-covered or not. First, they define that a sensor is k-perimeter-covered if every point on
the perimeter of this sensor is covered by at least k sensors [10]. Then a low-cost algo-
rithm is presented that discovers whether a particular sensor node is k-perimeter-
covered or not. Also, another algorithm is proposed which works with sensor nodes
with different sensing ranges. Finally, the perimeter-coverage of sensor nodes is related
to the coverage property of the monitoring area. It has been proved that when all the
sensor nodes are k-perimeter-covered, then the entire target area is k-covered.
In [10], several applications have been presented based on the proposed approach
for the coverage problem. Determining insufficiently covered areas in the sensor field
is one of the applications. Suppose k is the required value of coverage, first the base
station broadcasts this value to all sensors. Then each sensor communicates with its
neighbours to discover whether it has the segments in its perimeter which are less
than k-covered or not. The base station collects the information about the insuffi-
ciently covered segments to find out the areas that are less than k-covered.
A basic node scheduling protocol is also proposed in [10]. In this approach, each
sensor node S i needs to communicate with its neighbours and ask them to reassess
their perimeter coverage without node. If all the neighbours respond that still they are
k-perimeter-covered, then node S i can turn its radio off.
The proposed algorithm for coverage problem in this research work is quite effi-
cient considering its cost which is Ο(nd log d ) , where n is the number of sensors and
d indicates the highest number of nodes that might intersect a node. An interesting
point about this scheme is its ability to work with irregular sensing ranges based on
polygon approximation. However, its major drawback is its dependency on knowing
the exact coordinates of sensor nodes.
The basic aim in [5] is to maintain both sensing coverage and network connectivity as
the result of its coverage maintenance protocol. A geometric analysis is presented to
find out the relationship between coverage and connectivity [5]. It has been discov-
ered that full network connectivity is preserved when sensing range is equal or less
than half of the transmission range. As a result, a Coverage Configuration Protocol
(CCP) is devised to eliminate redundancy based on a pre-defined coverage degree. If
the coverage degree requested by the application is k, then the eligibility rule in CCP
is as follows. If all the intersection points inside the sensing range of a particular sen-
sor node are at least k-covered, then the sensor node is not entitled to be an active
node. In order to calculate the intersection points inside the sensing circle, each node
340 B. Pazand, A. Datta, and R. Cardell-Oliver
should maintain a list of its active neighbours located within a distance of twice its
sensing range. Three states have been considered for sensor nodes. These are sleep,
active and listen. Initially all the nodes are in active mode. Then each node checks the
eligibility rule to determine its next state. If the node becomes ineligible it goes to
sleep mode. Then it wakes up periodically and enters into the listen state to re-
evaluate its eligibility. Another solution is also considered in [5] to address the case
when a node’s sensing range is higher than half of the transmission range. SPAN [13]
which is a redundancy eliminator protocol (reviewed later) is integrated with CCP.
The former preserves the full connectivity while the latter maintains the requested
coverage degree for the sensor field. The result of this integration is a new eligibility
algorithm. With this new approach an inactive node executes the eligibility rules of
both SPAN and CCP. If the result of one of them is positive then the inactive node is
eligible to become active. Also, an active node will turn its radio off in case that it
does not comply with the eligibility rule of either SPAN or CCP.
CCP provides very good flexibility in both preserving connectivity and addressing
different levels of coverage. However, there are some shortcomings in this approach.
The computation complexity of CCP is Ο(n 3 ) , where n is the number of active nodes
within twice the sensing range of a node. This complexity is high for wireless sensor
networks and it is not as efficient as other approaches. Moreover, this technique is
infeasible for dense deployment of sensor nodes. Location error can also affect the
performance of CCP. More importantly, the CCP eligibility rule assumes considering
a perfect circular sensing model which is not true in the real world.
Slijepcevic et al. [12] focus on the set k-cover problem as a general form of the cover-
age problem. A heuristic is proposed to solve this problem by determining mutually
exclusive sets of sensor nodes. These sets are called covers that monitor the entire
area of the sensor field. The heuristic has been targeted to maximize the number of
these covers. The proposed solution is based on dividing the deployment area into a
number of fields. Each field is an area of the region of interest in which every point is
covered by the same set of sensor nodes. Suppose A denotes the set of calculated
fields. Then the set k-cover is formulated as follows. There are K disjoint covers, C1,
C2, ….., Ck, in which every member of A is covered by at least one element of every
cover Ci [12]. The first part of the solution organizes points of the deployment area
into a number of fields. A field is a unique area of the sensor field in which points
in this area are covered by the same set of sensors. The second part of the solution
utilizes a greedy algorithm to detect k covers. First, a field with lowest number of
associated sensors is selected. An objective function is then utilized to calculate a
probability value for every node in this field. This function measures the likelihood of
redundant coverage for every sensor node if it is selected. Therefore, a node is se-
lected which has the lowest likelihood. The selected field and its sensors are then
ignored for the next repetition of the heuristic. This process is repeated until the
current cover set attains full coverage. Then the members of calculated cover are
deleted from the fields and the above process is repeated for the next cover set. The
K+1th pass of the above heuristic which is the exit point, occurs when in the begin-
ning of the process there is a field that has not been covered by any of the previously
A Review of Redundancy Elimination Protocols for Wireless Sensor Networks 341
calculated covers. Calculated covers are responsible to monitor the sensor field. At
every time slice one cover is active while sensors belonging to the other covers are in
sleep mode. The activity period is considered the same for all the covers. However,
the authors have not implemented any scheduling scheme for covers.
The main advantage of this solution is its approach to organize the sensor network
into mutually exclusive sets of nodes. However, the assumption of circular sensing
area is not applicable in the real world. Therefore, the proposed scheme fails to de-
liver the promised full network coverage in real implementations.
Research in [11] specifically manages the coverage problem in wireless sensor net-
works. Two general aspects of the coverage problem are considered. The worst-case
coverage problem identifies the regions in the deployment area that have the lowest
level of coverage compared to other areas. The best-case coverage problem identifies
the regions that have the highest level of coverage by sensor nodes. For each case an
algorithm is presented based on geometric computation and graph theory.
Worst-case coverage is identified by a path with pre-determined starting and end
points in the sensor field. This path is called the Maximal Breach Path. The aim of the
proposed algorithm is to find this path so that for every point on the path the distance
of the point to the closest sensor node is maximized [11]. First, the Voronoi diagram
[24] of the sensor field with its sensor nodes is generated. Then from this diagram a
graph is created so that its nodes are the vertices of the diagram and its edges are the
line segments of the diagram. A weight is also assigned to each edge that is the lowest
distance from the closest sensor in the field. Finally, some graph search algorithms
[11] are utilized to discover the path.
The best-case coverage is identified by a path whose starting point and end point are
initially specified. This path is called the Maximal Support Path. The path is discovered
in a way that the distance of every point on this path to the closest sensor is minimized.
For this case another geometric structure is utilized that is Delaunay triangulation [25].
First, the Delaunay triangulation is computed based on the deployment area and its
sensor nodes. Then a graph is formed such that its vertices are the sensor nodes and its
edges are the lines of the triangles. The weight that is assigned to each edge is the length
of its corresponding edge in the triangle. Finally, a search algorithm with specific search
criteria [11] is applied to the graph to discover the maximal support path.
The most obvious advantages of the proposed scheme are the applications that can
be derived from it. For example, the network administrator is able to discover the
paths in sensor field which suffer from poor coverage. Then, coverage can be easily
improved by deployment of new nodes in those areas. Additionally, by discovering
maximal support paths, it is possible to schedule the redundant nodes of these paths
to sleep. Therefore, the discovery of breach and support paths leads to a balanced
network coverage. There are also some drawbacks in the proposed scheme. It only
considers homogeneous sensor nodes. Moreover, the accuracy of the discovered paths
is strongly related to the precision of location information of sensor nodes. Therefore,
the localization techniques that rely on a few GPS devices can affect the quality of
maximal breach and support paths.
342 B. Pazand, A. Datta, and R. Cardell-Oliver
In [4] a node scheduling scheme is proposed in which every node is autonomously re-
sponsible for checking an off-duty eligibility rule. This rule considers the neighbours of
a node to determine whether they can sponsor the node by covering its sensing area. The
proposed solution comprises of two rounds. Round one is a self-scheduling process that
is followed by a typical sensing round. These rounds are repeated periodically.
In the beginning of the self-scheduling round, each node is responsible for discov-
ering its neighbours within its sensing range. This discovery includes location infor-
mation for these nodes. Then, the node determines its eligibility to be an inactive node
by performing some geometric calculations. These calculations include the measure-
ment of the central angles of the sectors that are formed based on intersection of this
node by each of its neighbours.
Every node is responsible for calculating all the central angles that exist between
the node and its neighbours. If the summation of all these angles covers 360°, the
sensing area of the node is fully sponsored by its neighbours and it is eligible to turn
off its radio. In order to avoid any conflict that may occur if nodes sponsor each other
concurrently, a standard back-off technique is utilized. Also, the connectivity between
on-duty nodes is provided by LEACH [26], a cluster-based protocol.
It has been proved in [4] that the requirement of location information can be replaced
by direction information in order to determine sponsor nodes. Additionally, the off-duty
eligibility rule has been further extended to respond to the case that nodes have different
sensing ranges. Nevertheless, there are some major shortcomings. It has been assumed
that the sensing round should be long enough in order to compensate for the overhead of
the self-scheduling round. This assumption can lead to large sensing and communica-
tion gaps in case of node failures. Moreover, in case of dense deployments, there is no
approach to tackle the packet collisions that might occur during neighbourhood discov-
ery. Loss of accuracy of neighbourhood information leads to higher numbers of on-duty
nodes. The last disadvantage is the dependency of the off-duty eligibility rule on the
existence of at least three neighbours for every sensor node.
Detecting and eliminating redundant sensor nodes while preserving complete network
coverage is the main target of the solution presented in [1, 2]. Moreover, it presents an
approach to discover the coverage-boundary of a sensor field. The proposed methods
are based on Voronoi diagrams generated locally at each sensor node.
A redundant sensor is defined as a node whose sensing disk is completely under cov-
erage of other nodes. This approach utilizes two different Voronoi diagrams. One is the
Voronoi diagram of the network. The second one, constructed for every sensor node, is
the Voronoi diagram of the Voronoi neighbours of that node, called the 2-Voronoi dia-
gram. The vertices of this diagram are called 2-Voronoi Vertices (2-VV). In addition,
the intersection point between the coverage circumcircle of the sensor node and an edge
of its 2-Voronoi diagram is calculated. It is called a 2-Voronoi Intersection Point (2-
VIP). Then, every sensor node is responsible for discovering whether it is redundant or
not by utilizing this information. It has been proved in [1, 2] that a sensor node is
redundant if its 2-VVs and 2-VIPs are under the coverage of its Voronoi neighbours.
Every sensor node is responsible for checking this rule to determine whether it is a
A Review of Redundancy Elimination Protocols for Wireless Sensor Networks 343
redundant node or not. Redundant nodes that are Voronoi neighbours of each other
should not decide to turn their radio off simultaneously. Since that could cause blind
spots in the sensor field. In order to prevent it, an approach based on discovering a
maximal independent set [27] of redundant nodes is implemented.
The second contribution of the research work presented in [1, 2] is dedicated to the
detection of the coverage-boundary of the deployment area. It is formulated that the
coverage-boundary includes the sensing area of those sensor nodes whose coverage
areas are not fully covered by sensing disks of other nodes. It has been proved that a
node is on the boundary area if its Voronoi cell is not covered by its sensing disk [1,2].
The main advantage of the proposed scheme is its ability to address the heteroge-
neous sensor networks. It is achieved by a modified version of Voronoi diagrams.
Additionally, the proposed scheme utilizes a distributed approach compared to the
centralized Voronoi calculations in [11]. It is capable of recomputing the Voronoi
information locally in order to address node failures and deployment of new nodes.
As a result, it is much more robust in this regard compared to [11]. Nevertheless, this
protocol imposes a large amount of overhead for two reasons. The first issue is the
regular maintenance of Voronoi cells. The second factor is the initial overhead im-
posed when every sensor node is supposed to find its Voronoi neighbours and com-
pute its Voronoi cell. Also, similar to [11], the performance of the scheme is sensitive
to the accuracy of the location information of each node. This is because of the
dependency of these approaches to geometrical calculations.
Chen et al. [13] propose a protocol to save energy in ad hoc wireless networks by
keeping only the necessary nodes active to forward routing traffic. In SPAN each
node periodically decides whether it is eligible to become a coordinator node and join
the forwarding backbone or it should turn its radio off.
344 B. Pazand, A. Datta, and R. Cardell-Oliver
Younis et al. [9] propose a solution for node scheduling that relies on four different
conditions. Two tests (RTest-D1 and RTest-D2) are geometrical. They determine that
a node is redundant if its sensing area is covered geometrically by other active nodes.
Two other tests (RTest-H1 and RTest-H2) specify the redundancy of a node if some
conditions based on node density are met [9].
RTest-D1: A node v is redundant if these conditions are satisfied. 1) There are
nodes vj, 1 ≤ j ≤ 3 that lie in the sensing range of node v. 2) vj nodes are active. 3) vj
nodes are pairwise neighbours insides the sensing range. 4) Node v is placed inside
the triangle formed by vj nodes. 5) Boundary of node v is covered by the boundary of
vj nodes.
RTest-D2: A node v is redundant if these conditions are satisfied. 1) There are
nodes vj, 1 ≤ j ≤ 3 that lie in the range 0.618*Rs. Rs is the sensing range of node v. (for
more details please refer to [9]). 2) vj nodes are active. 3) vj nodes are pairwise non-
neighbours within the sensing range of node v.
A Review of Redundancy Elimination Protocols for Wireless Sensor Networks 345
RTest-H1: A node v is redundant if these conditions are satisfied. 1) There are ac-
tive neighbours vj, j ≥ 4, within distance Rs. 2) Every neighbour of node v is a
neighbour of one or more of vj nodes.
RTest-H2: First these conditions are considered. 1) Node v has at least y active or
undecided neighbours inside its sensing range. 2) Its other neighbours have one or
more neighbours among those y neighbours. Node v is considered redundant if it has
the least remaining energy among those y active neighbours. In [9] the authors as-
sumed that y is 6.
Each node has three states: active, sleep and undecided. A Location-Unaware Cov-
erage (LUC) algorithm utilizes rules D1, D2, H1 and H2 to determine the state of
each node in network. The first phase is started by a neighbourhood discovery when
each node learns its one-hop and two-hop neighbours as well as its distance to its one-
hop neighbours. Then the operational phase begins that determines the active and
asleep nodes. Whenever a node notices that one of its neighbours has become an ac-
tive node, it performs a redundancy check procedure to decide whether it is redundant
or not. This procedure performs the tests in a special order. The order is the same as
we listed above. If the result of any test determines that the node is redundant, it will
quit the procedure and will change its state to sleep. Otherwise, the procedure is trig-
gered by the next active neighbour. However, if a specific timer expires and the node
still remains in an undecided state, the last round of the redundancy checking proce-
dure is performed. Nodes become active if all the tests have a negative result.
There are some shortcomings of this scheme. First and foremost, it has been shown
via the simulations in [9] that the total success ratio of the first two tests is almost
60%. Since the next two tests are exploited for dense networks, the proposed ap-
proach can not detect all the redundant nodes for low density networks. Moreover, the
algorithm is limited by considering the assumption that the transmission range should
be at least two times the sensing range. Last but not least, nodes should be equipped
with special hardware to measure the distance between neighbours.
The proposed solution in [8] comprises of two parts. First, the authors present some
theoretical analysis to estimate the coverage level of the monitoring area based on
some parameters of the wireless sensor network. Second, they devise a location-free
node scheduling scheme.
In order to formulate the coverage expectancy, three factors are considered. These
are sensing range, number of deployed nodes and area of the sensor field. First the
expected coverage of each sensor node is predicted and then the expected coverage
ratio of the entire deployment area is discovered as follows: (more details in [8])
⎡ ⎛ 1 4 4 3 4 3 ⎞ ⎤
n
⎢ ⎜ r − lr − mr + πr 2 ml ⎟ ⎥
E[C n ] = ⎢1 − ⎜1 − 2 3 3 ⎟ ⎥lm (1)
⎢ ⎜ m 2l 2 ⎟ ⎥
⎢ ⎜⎝ ⎟ ⎥
⎠ ⎦
⎣
Here, E[Cn] is the expected coverage level, r is the sensing range, n is the number of
deployed nodes and the deployment area is represented by an l by m rectangle. Also,
346 B. Pazand, A. Datta, and R. Cardell-Oliver
this expression can be utilized to determine the minimum number of sensor nodes
needed to maintain a pre-defined coverage ratio requested by the application.
This analysis for coverage estimation is further utilized to present a node scheduling
scheme. The proposed solution is an uncoordinated scheme in which each node stochas-
tically decides to become active or asleep. This decision is repeated independently. The
duration of time that nodes remain in active or sleep state is based on an exponential
distribution with parameters λ a and λ s , respectively [8]. If we consider that a particular
node has changed its state m times. Then the probability that the node would be active is:
mλ a
Pa = (2)
mλ a + mλ s
Then it can be estimated that the number of active nodes at any given time is nPa.
Now if we put the estimated number of active nodes in equation 1, the expected cov-
erage of the network can be calculated. This calculation can help us to adjust λ a
and λ s to attain the coverage ratio requested by an application.
Although, there are some basic analyses [35, 36] in the literature for estimating the
coverage, this research takes into account the border effect to present a more accurate
approximation. Border effect refers to nodes that are located close to the border of the
monitoring area. Moreover, the proposed node scheduling scheme is an uncoordinated
approach with very low overhead. Nevertheless, the scheduling process is completely
stochastic which might not possess a stable behaviour for some issues such as pre-
serving the network capacity. In addition, the proposed solution for coverage expec-
tancy neither addresses sensor nodes with different sensing ranges nor real-world
irregularity of sensing area.
λd
λnew = λ (3)
λ
Subsequently, a new sleeping time is generated according to the equation 4. Where, ts
is the previous sleeping time.
f (t s ) = λ new e − λ
new
ts
(4)
It has been proven in [6, 7] that with this kind of adjustment of probe rate as well as
exponentially distributed sleeping times, the probe rate that every working node re-
ceives is equal to the desired probe rate.
PEAS is a location-free solution for the coverage problem that does not rely on dis-
tance information. Moreover, by simply varying the probe range it can provide differ-
ent coverage levels based on application requirements. Additionally, its proposed
approach of adaptive sleeping can address node failures quite robustly while keeping
the overhead of this action low. Due to the high efficiency of PEAS in terms of han-
dling node failures, we have utilized it as the fault-tolerant module of our node sched-
uling scheme. However, it needs to be mentioned that the main disadvantage of PEAS
is its quality in terms of temporal coverage. Temporal coverage indicates the cover-
age ratio of the sensor field throughout the network lifetime.
Wi et al. [29] propose a location-free approach for selecting a set of active sensor
nodes. The main contribution of this research is that in the absence of node coordi-
nates, it can achieve 100% network coverage. However, it assumes that nodes are
equipped with omni-directional antennas to measure the distance between nodes by
received signal strength. The proposed solution comprises of two phases. In phase one
the initial set of active nodes is determined. The second phase is dedicated to detect-
ing blind spots in the sensor field.
First, all the nodes are inactive and each of them runs a back-off timer. Upon the
expiration of this timer, a node wakes up and checks whether there is a working node
within its sensing range or not. If the discovery is not successful, the node becomes a
new working node. Otherwise, the node remains awake for a period of time to build a
Neighbour Table based on some control messages. This table includes node id and
distances of all the one-hop and two-hop neighbours. It is utilized by the second phase
for detection of sensing holes. When this phase is finished an initial set of active
nodes has been discovered.
In the second phase, a triangular self-test is performed by each inactive node to
eliminate any possible blind spots. For every inactive node a triangle is formed based
on the three closest active neighbours. Then relative coordinates are assigned to the
endpoints of the triangle followed by the calculation of the relative coordinates of the
circumcentre of the triangle. Based on calculating the distance from circumcentre to
the endpoints of the triangle and comparing it with the sensing range, the existence of
a blind spot is determined. If the result of this self checking approach is positive the
inactive node becomes active to fill the sensing holes.
348 B. Pazand, A. Datta, and R. Cardell-Oliver
There are two main advantages of this solution. Firstly, a relative coordinate is
assigned to each node based on virtual coordinates of its neighbours. This position
information is not used for other location calculations. Therefore, the scheme is not
affected by propagation of location errors. Secondly, the simulation results presented
in [29] indicate that the proposed scheme is quite scalable so that the packet overhead
does not change when node density increases. However, there are some drawbacks as
well. Although the simulation results show that the set of active nodes covers the
deployment area completely, the scheme achieves that by eliminating the margins of
the sensor field. Also, there are some real-world factors that affect the efficiency of
distance measurement based on received signal strength. This issue decreases the
performance of the triangular self-test mechanism. Moreover, the proposed solution
assumes that the transmission range should be at least three times the sensing range
which limits the applicability of the solution.
4 Discussion
The first and foremost characteristic of a redundancy elimination protocol is its ap-
proach regarding location information. In sections 2 and 3 we discussed the advan-
tages and disadvantages of location-free and location-aware solutions. However, an
important point that needs to be mentioned is that some location-free approaches [29,
9] consider distance information. There are three main drawbacks in this regard. First,
we need special hardware to measure distance. This increases the cost of the wireless
sensor network. Second, environmental conditions affect the accuracy of distance
estimation which might decrease the quality of coverage. The third drawback is re-
lated to the techniques that are used for measuring the distance between nodes. The
RF signal strength approaches might have low performance when the transmission
range is short. Fading, multipath and interferences are the factors that affect the signal
strength when the radio range is short [30]. Moreover, time of flight approaches [31,
32, 33] might deliver inaccurate results mostly because of clock skew and drift be-
tween nodes.
Protocols [10, 12, 4, 5, 29, 9] determine redundant nodes based on some calcula-
tions on either the area of the sensing circle or the perimeter of these circles. These
approaches might not deliver what they have promised to achieve. This is due to the
assumption of perfect circular sensing areas. In the real-world, sensing is not only
irregular but it might vary over time. Although other protocols consider circular sens-
ing areas, they use it only for calculation of the coverage ratio and not as a building
block of their schemes. Therefore, the only shortcoming of these approaches is some
degrees of error in measuring their coverage ratio metric.
All the solutions that we have reviewed in this chapter utilize the perfect circular
radio propagation model. This model does not reflect the real-life behaviour of the
wireless channel. Therefore, the performance of these protocols when they are im-
plemented can be different from what we have seen through simulation results. Al-
though GAF [14] has investigated the effect of the shadowing radio model, this proto-
col is originally proposed for unrealistic radio models.
Elimination of redundant nodes might result in an unconnected set of representa-
tive nodes. In order to tackle this issue, some approaches [3, 29, 9] consider a special
assumption that the transmission range should be at least twice the sensing range.
A Review of Redundancy Elimination Protocols for Wireless Sensor Networks 349
Therefore, they can not address particular applications where the communication
range might be even lower than sensing range.
Finally, Table 1 compares the main characteristics of the protocols that we have
reviewed in this paper.
Redundancy Use of
Location Location Distance
Eliminator Distributed Centralized Coordinated Uncoordinated Unrealistic
Free Aware Information
Protocol Radio Model
GAF [14]
- 9 - 9 - 9 - -1
TTS [3]
- 9 - - 9 - 9 9
K-Coverage
based on
Perimeter-
Coverage of - 9 - 9 - 9 - 9
Sensor
Nodes[10]
Integrated
Coverage and
Connectivity - 9 - 9 - 9 - 9
[5]
SET
K-COVER
- 9 - - 9 9 9
[12]
Worst-Case &
Best-Case
- 9 - - 9 N/A - 9
Coverage [11]
Node
Scheduling by
Sponsoring - 9 - 9 - 9 - 9
Nodes [4]
Coverage
Preserving
Redundancy
- 9 - 9 - 9 - 9
Elimination
[1, 2]
SPAN [13] 9 - - 9 - 9 - 9
LUC [9] 9 - 9 9 - 9 - 9
Expectancy of
Coverage [8] 9 - - 9 - 9 9
PEAS [6,7] 9 - - 9 - 9 - 9
Triangular
[29] 9 - 9 9 - 9 - 9
1
GAF is originally proposed for unrealistic radio models. However it has also been investi-
gated under shadowing model.
350 B. Pazand, A. Datta, and R. Cardell-Oliver
References
1. Carbunar, B., Grama, A., VIitek, J., Carbunar, O.: Redundancy and Coverage Detection in
Sensor Networks. ACM Transactions on Sensor Networks 2(1), 94–128 (2006)
2. Carbunar, B., Grama, A., Vitek, J., Carbunar, O.: Coverage Preserving Redundancy Elimina-
tion in Sensor Networks. In: Proceedings of the First IEEE Communications Society Confer-
ence on Sensor and Ad Hoc Communications and Networks (SECON 2004) (October 2004)
3. Tezcan, N., Wang, W.: TTS: A Two-Tiered Scheduling Algorithm for Wireless Sensor
Networks. In: Proceedings of IEEE International Conference on Communications (ICC
2006) (June 2006)
4. Tian, D., Georganas, N.D.: A Coverage-Preserving Node Scheduling Scheme for Large
Wireless Sensor Networks. In: Proceedings of the 1st ACM International Workshop on
Wireless Sensor Networks and Applications (WSNA 2002) (September 2002)
5. Wang, X., Xing, G., Zhang, Y., Lu, C., Pless, R., Gill, C.: Integrated Coverage and Connec-
tivity Configuration in Wireless Sensor Networks. In: Proceedings of the 1st International
Conference on Embedded Networked Sensor Systems (SenSys 2003) (November 2003)
6. Ye, F., Zhang, H., Lu, S., Zhang, L., Hou, J.: A Randomized Energy-Conservation Proto-
col for Resilient Sensor Networks. Wireless Networks 12(5), 637–652 (2006)
7. Ye, F., Zhong, G., Cheng, J., Lu, S., Zhang, L.: PEAS: A Robust Energy Conserving Pro-
tocol for Long-lived Sensor Networks. In: Proceedings of the 23rd International Confer-
ence on Distributed Computing Systems, ICDCS 2003 (May 2003)
8. Yen, L.-H., Yu, C.W., Cheng, Y.-M.: Expected K-Coverage in Wireless Sensor Networks.
Ad Hoc Networks 4(5), 636–650 (2005)
9. Younis, O., Krunz, M., Ramasubramanian, S.: Coverage Without Location Information.
In: Proceedings of the IEEE International Conference on Network Protocols (ICNP 2007)
(October 2007)
10. Huang, C.-F., Tseng, Y.-C.: The Coverage Problem in a Wireless Sensor Network. In:
Proceedings of the 2nd ACM International Conference on Wireless Sensor Networks and
Applications (WSNA 2003) (September 2003)
11. Meguerdichian, S., Koushanfar, F., Potkonjak, M., Srivastava, M.: Coverage Problems in
Wireless Ad-Hoc Sensor Networks. In: Proceedings of the twentieth Annual Joint Confer-
ence of the IEEE Computer and Communications Societies (INFOCOM 2001) (April 2001)
12. Slijepcevic, S., Potkonjak, M.: Power Efficient Organization of Wireless Sensor Networks.
In: Proceedings of the IEEE International Conference on Communications (ICC 2001)
(June 2001)
13. Chen, B., Jamieson, K., Balakrishnan, H., Morris, R.: Span: An Energy-Efficient Coordi-
nation Algorithm for Topology Maintenance in Ad Hoc Wireless Networks. Wireless
Networks 8(5), 481–494 (2002)
14. Xu, Y., Heidemann, J., Estrin, D.: Geography-Informed Energy Conservation for Ad Hoc
Routing. In: Proceedings of the 7th Annual International Conference on Mobile Comput-
ing and Networking (MobiCom 2001) (July 2001)
15. Bulusu, N., Estrin, D., Girod, L., Heidemann, J.: Scalable Coordination for Wireless Sen-
sor Networks: Self-Configuring Localization Systems. In: Proceedings of the 6th Interna-
tional Symposium on Communication Theory and Applications (ISCTA 2001) (July 2001)
16. Priyantha, N.B., Chakraborty, A., Balakrishnan, H.: The Cricket Location-Support System.
In: Proceedings of the 6th Annual International Conference on Mobile Computing and
Networking (MobiCom 2000) (August 2000)
17. Savvides, A., Han, C.-C., Strivastava, M.B.: Dynamic Fine-Grained Localization in Ad-
Hoc Networks of Sensors. In: Proceedings of the 7th Annual International Conference on
Mobile Computing and Networking (MobiCom 2001) (July 2001)
A Review of Redundancy Elimination Protocols for Wireless Sensor Networks 351
18. Whitehouse, C.D.: The Design of Calamari: An Ad-hoc Localization System for Sensor
Networks. Master Thesis, University of California at Berkeley (2002)
19. He, T., Huang, C., Blum, B.M., Stankovic, J.A., Abdelzaher, T.: Range-Free Localization
Schemes for Large Scale Sensor Networks. In: Proceedings of the 9th Annual International
Conference on Mobile Computing and Networking (MobiCom 2003) (September 2003)
20. Moore, D., Leonard, J., Rus, D., Teller, S.: Robust Distributed Network Localization with
Noisy Range Measurements. In: Proceedings of the 2nd International Conference on Em-
bedded Networked Sensor Systems (SenSys 2004) (November 2004)
21. Niculescu, D., Nath, B.: Ad Hoc Positioning System (APS). In: Proceedings of Global
Telecommunications Conference (GLOBECOM 2001) (November 2001)
22. Zhong, Z., He, T.: MSP: Multi-Sequence Positioning of Wireless Sensor Nodes. In: Pro-
ceedings of the 5th International Conference on Embedded Networked Sensor Systems
(SenSys 2007) (November 2007)
23. Cheng, X., Thaeler, A., Xue, G., Chen, D.: TPS: A Time-Based Positioning Scheme for
Outdoor Wireless Sensor Networks. In: Proceedings of the Twenty-third Annual Joint
Conference of the IEEE Computer and Communications Societies (INFOCOM 2004)
(March 2004)
24. Aurenhammer, F.: Voronoi Diagrams - A Survey of a Fundamental Geometric Data Struc-
ture. ACM Computing Surveys 23(3), 345–405 (1991)
25. Preparata, F.P., Shamos, M.I.: Computational Geometry: An Introduction. Springer-Verlag
New York, Inc., Heidelberg (1985)
26. Heinzelman, W.R., Chandrakasan, A., Balakrishnan, H.: Energy-Efficient Communication
Protocol for Wireless Microsensor Networks. In: Proceedings of the 33rd Hawaii Interna-
tional Conference on System Sciences (HICSS 2000) (January 2000)
27. Luby, M.: A Simple Parallel Algorithm for the Maximal Independent Set Problem. In:
Proceedings of the 7th Annual ACM Symposium on Theory of Computing (STOC 1985)
(May 1985)
28. Wireless LAN Medium Access Control and Physical Layer Specifications. IEEE 802.11
Standard (IEEE Computer Society LAN MAN Standards Committee) (August 1999)
29. Wi, T.-T., Ssu, K.-F.: Determining Active Sensor Nodes for Complete Coverage without
Location Information. International Journal of Ad Hoc and Ubiquitous Computing 1(1-2),
38–46 (2005)
30. Bulusu, N., Heidemann, J., Estrin, D.: GPS-less Low Cost Outdoor Localization for Very
Small Devices. IEEE Personal Communications 7(5), 28–34 (2000)
31. Girod, L., Lukac, M., Trifa, V., Estrin, D.: The Design and Implementation of a Self-
Calibrating Distributed Acoustic Sensing Platform. In: Proceedings of the 4th International
Conference on Embedded Networked Sensor Systems (SenSys 2006) (November 2006)
32. Whitehouse, K., Culler, D.: Calibration as Parameter Estimation in Sensor Networks. In:
Proceedings of the 1st ACM International Workshop on Wireless Sensor Networks and
Applications (WSNA 2002) (September 2002)
33. Youssef, M., Youssef, A., Rieger, C., Shankar, U., Agrawala, A.: PinPoint: An Asynchro-
nous Time-Based Location Determination System. In: Proceedings of the 4th International
Conference on Mobile Systems, Applications and Services (June 2006)
34. Lazos, L., Poovendran, R.: SeRLoc: Secure Range-Independent Localization for Wireless
Sensor Networks. In: Proceedings of the 3rd ACM Workshop on Wireless Security (WiSe
2004) (October 2004)
35. Hall, P.: Introduction to the Theory of Coverage Processes. John Wiley and Sons, Chiches-
ter (1988)
36. Philips, T.K., Panwar, S.S., Tantawi, A.N.: Connectivity Properties of a Packet Radio
Network Model. IEEE Transactions on Information Theory 35(5), 1044–1047 (1989)
A Light-Weighted Misused Key Detection
in Wireless Sensor Networks
1 Introduction
Recent advances in wireless communication, micro system and low-power tech-
nologies have enabled the wireless sensor networks(WSNs). Generally, WSNs
consist of many, low-cost, low-power and small sensor nodes and the base sta-
tion(BS). Once the sensor nodes are deployed at a target area, they collect some
information from the area and report it to the BS. The BS takes the role of
gateway to the traditional networks(:Internet), so we can use the collected in-
formation for specific applications. For example, some applications of WSNs
D. Taniar et al. (Eds.): ICCSA 2010, Part III, LNCS 6018, pp. 352–367, 2010.
c Springer-Verlag Berlin Heidelberg 2010
A Light-Weighted Misused Key Detection in WSNs 353
are habitat monitoring, object tracking, fire detection, traffic monitoring and so
on[1,2,7].
The security services in WSNs are very important because the sensor nodes
are often deployed in hostile environments. An adversary can easily overhear
the radio channels, insert or modify some messages, or capture physically some
sensor nodes. To guarantee security service such as the integrity, authentication
and confidentiality in the WSNs, cryptographic methods are essential and the
key management is the most fundamental one. However, the traditional key
management methods such as KDC(Key Distribution Center) and public key
mechanism are the impractical solutions due to the limited resources of sensor
nodes[14].
Many key management techniques in WSNs have been introduced[4,8,9,10,13].
Eschenauer and Gligor proposed the basic random key distribution scheme based
on the probability model to overcome the storage limitation of the sensor node
for storing keys[8]. The main idea is that k keys are randomly selected by each
node out of a large pool of P keys. However, it has the critical drawback that
the attacker can easily threat the network survivability by capturing some nodes.
Chan suggested the q-composite random key distribution scheme[6]. This scheme
makes the resilience against node capture more strong but makes the key sharing
probability between any pair of nodes decreased. Liu and Ning proposed more
enhanced key management scheme using t-degree bivariate polynomials and two-
dimensional grid[3]. It perfectly provides the network resilience until t + 1 nodes
are captured for any specific keys and reduces the overhead on establishing
pairwise keys. An attacker can get the keys between any two non-compromised
nodes because the all of the key management schemes mentioned above are
commonly based on the probability model. If An adversary gets some keys using
physical node capture attack, the attacker might infer some keys between non-
compromised nodes by using the collected keys from compromised nodes, and
then he can attempt some kinds of attacks such as eavesdropping, insertion
and modification of messages. The insertion and modification attacks are more
harmful than the eavesdropping because they can lead the abnormal operations
of the applications in WSNs. Figure 1 describes this situation. Node 3, 6 and
7 were compromised by the attacker and the attacker obtained the shared key,
K4,5 , between node 4 and 5. He can insert or modify some messages and finally
threaten the overall functionality of an application.
Recently, Liu and Dong has proposed a misused key detection scheme in WSNs
to prevent these kinds of attacks[5]. According to them, the misused key can be
defined under the following three conditions.
– It should be the key shared between non-compromised nodes.
– The key is exposed to an adversary.
– The key must be used for insertion and modification attacks.
To detect a misused key efficiently in distributed manner, they has introduced
committing values and detecting nodes. Whenever a sender node transfers a mes-
sage to a receiver node, the sender node has to attach committing values to the
message and the receiver node performs the detecting procedure by generating
a report message using the committing values under the predefined probability.
Finally, the detecting nodes and the BS check the report message to determine
whether the key was misused.
However, Liu and Dong’s scheme has four limitations. First, the unused com-
mitting values at the receiver node are big overhead. Second, the reporting
procedure can detect only the unidirectional key misuse from a sender to a
receiver per one report message. Third, it is impractical that the detecting node
lists must be exchanged between sensor nodes because the pairwise key estab-
lishment depends on the routing protocol deployed for the purpose of efficient
operation of specific applications in WSNs. Additionally, the detecting proce-
dure in Liu and Dong’s scheme occurs unbalanced traffic overhead among sensor
nodes.
In this paper, we introduce a light-weighted misused key detection mechanism
based on Liu and Dong’s scheme. Our scheme has better two properties, which
are the reduction of the detecting overhead and the high detecting probability,
than Liu and Dong’s scheme by performing the bidirectional misused key de-
tection, removing of unnecessary message and using the reasonable probability
method for the balanced misused detection among the sensor nodes.
The rest of this paper is organized as follows. The next section briefly re-
views the Liu and Dong’s scheme and discusses about its limitations. Section
3 introduces our scheme: a light-weighted misused key detection mechanism. In
section 4, we will analyze our mechanism from the viewpoints of security and
energy efficiency compared with the Liu and Dong’s method. Finally, section 5
summarizes our paper and discusses about future research.
messages make the application choice wrong decisions. To prevent these kinds of
attacks, Liu and Dong suggests the additional protection for non-compromised
sensor nodes even if an attacker has learned the shared keys between them.
– The adversary can eavesdrop, modify, replay, forge or block any network
traffic.
– It is computationally impossible to break the cryptographic primitives.
– The adversary can compromise a few nodes and obtain some shared keys
between non-compromised nodes.
(a) Initialization
When the detecting node i receives the report message, i verifies whether V1
equals H(Cu,v , H(Ku , i)). If the verification fails, i simply forwards
u||v||Cu,v ||V2 ||M AC3 to the BS. Finally, the BS checks if V2 equals H(Cu,v , Ku ).
If not, the BS judges the pairwise key between u and v to be misused and then
distributes the new pairwise key to u and v using the unique secret key Ku and
Kv .
Exchanging of the detecting node lists. The selection of detecting node de-
pends on the sequence number of the message and the size of the detecting node
list. The sender node must previously know the detecting node list of receiver
node in secure manner. However, this may be infeasible since the establishment
of pairwise key between the sender node and the receiver node is not predictable.
The reason is that the establishment of a pairwise key not only depends on the
routing or the application of the network but also frequently changes as the
environment varies.
our scheme will make each sensor node perform the detecting procedure one
per 1/p message transmitting rounds on average for randomized rotation of the
detection procedure among the sensor nodes.
probability model for the randomized cluster leader election of clustering scheme
in WSNs. Whenever u sends a message M to v, u generates a random number δ
between 0 and 1. If the δ < Ptr (u, r), u performs the detecting procedure which
will be described in section 3.2. Equation (1) defines the Ptr (u, r), where p is
the optimal reporting probability for detecting a misused key of a sensor node,
r is the current message transmission round and G is the set of sensor nodes
that did not perform the reporting procedure during last rmax = 1/p message
transmitting rounds. According to Eq. (1), every sensor node must perform the
reporting procedure only once during rmax message transmission rounds. In this
way, the sensor nodes which did not perform the reporting procedure recently
have a more chance for detecting the misused key than the nodes that have
already performed during last rmax rounds. As a result, our scheme ensures that
each sensor node performs the detecting procedure only one per 1/p message
transmitting rounds on average and it achieves the balanced and fair detection
of a misused key among the sensor nodes.
⎧ p
⎪
⎨ ,if n ∈ G
Ptr (u, r) = 1 − p × rmod 1
p (1)
⎪
⎩
0 ,otherwise
Initializing Phase. Initialization phase is exactly same with the Liu and
Dong’s scheme. Each regular sensor node pre-loads the unique secret key shared
with only the BS. The detecting node i stores H(Ku , i) for every sensor node u as
described in Figure 2(a). After deployment, each regular sensor node u discovers
the detecting nodes around itself and makes the detecting node List (u).
Tracking Phase. In this phase, each sensor node only tracks the history of
messages. Whenever u sends a new message M to v, u updates the Cu,v =
H(Cu,v , M ) where the initial value of Cu,v is zero. v updates the Cu,v with
H(Cu,v , M ) if it receives the message M from u. The Cv,u also is tracked by u
and v for the communication from v to u.
updates Cv,u = H(Cv,u , M ) and then immediately sends back the committing
message, (M )||V1 ||V2 with M AC2 as the message ② in Fig. 3. (M ) denotes the
piggybacked message heading for u if there is a message M to send. If not, only
the V1 ||V2 is transferred. The V1 is the hashed value with H(Cu,v , Cv,u , H(Kv , i))
and V2 is H(Cu,v , Cv,u , Kv ). The M AC2 is generated by using the pairwise key
Ku,v between u and v.
The sensor node u waits the message ② in Fig. 3 until the T1 expires. If T1 ex-
pires, u reports it to the BS that the message was blocked. Otherwise, u generates
a report message as ③ in Fig. 3, where the M AC3 and M AC4 are generated by
using H(Ku , i) and Ku in order. The first piece of information, v||Cu,v ||Cv,u ||V1 ,
is for the detecting node i and the second one, u||v||Cu,v ||Cv,u ||V2 , is for the BS.
Detecting Phase. When the detecting node i receives the report message ③ in
Fig. 3, it verifies whether the V1 equals H(Cu,v , Cv,u , H(Kv , i)). If it is not same,
it means that the key Ku,v was misused so the detecting node i simply forwards
the second piece of information, u||v||Cu,v ||Cv,u ||V2 ||M AC4 as the message ④
in Fig. 3. Otherwise, the detecting node i sets a timer T2 for detection of the
blocking the report message.
The BS checks whether the V2 equals H(Cu,v , Cv,u , Kv ). If it is not equal,
the BS generates a new pairwise key Ku,v and transmits it to the both u and v
securely using Ku and Kv . The BS concludes a correct verification with a high
probability if the detection procedure at the detecting node was successfully
performed.
A Light-Weighted Misused Key Detection in WSNs 361
4 Evaluation
To enhance the Liu and Dong’s scheme, we have Lintroduced our ideas in the
previous section. This section compares our scheme with Liu and Dong’s scheme
in the aspects of security and energy efficiency.
4.1 Security
False Key Revocation. Any non-compromised key must not be revoked if the
key is shared between non-compromised regular sensor nodes. There are three
cases that we have to consider.
1. If an adversary sends a forged report message to a detecting node i to revoke
a non-compromised key Ku,v , i will never report a false alarm to the BS.
The adversary cannot generate the M AC2 to authenticate his forged report
message to i since he cannot know H(Ku , i).
2. If a detecting node i is compromised and the adversary tries to revoke a
non-compromised key Ku,v through i, the adversary will never revoke the
key. He cannot create the M AC4 because the attacker does not know the
shared key between u and the BS, Ku . As a result, the forged message by
the compromised detecting node will be always rejected at the BS.
3. If a report message is authenticated by a detecting node i or the BS and the
detecting node i is not compromised, the detection of a misused key always
is guaranteed because the V1 and V2 are protected by the shared key, Kv ,
with only the BS.
Liu and Dong’s scheme has proved that their scheme satisfies these three cases.
Our scheme also satisfies the security requirements as described above.
1. we can define the probability of not performing the detecting procedure until
r messages have been transmitted within rmax = 1/p rounds as Eq. (3).
r
Pnotr (u, r) = (1 − Ptr (u, i)) (3)
i=1
2. If we assume that u did not performed the detecting procedure until r mes-
sages have been transmitted within rmax rounds, the probability that u does
not perform the reporting procedure until the additional x messages have
been transmitted from u to v can be calculated from Eq. (4).
⎧
⎪
⎪
r+x
⎪
⎪ {1−Pnotr (u, i−1) × Ptr (u, i) × (1−fc)}
⎪
⎪
⎪
⎪
⎪
⎪
i=r+1
⎪
⎪ ,if x + r ≤ rmax
⎪
⎨
Pnotdet|notreport (u, x, r) =
⎪
⎪
⎪
⎪ (x+r)modrmax
⎪
⎪
⎪
⎪ f (x+r)/rmax
× {1−Pnotr (u, i−1)
⎪
⎪ c
⎪
⎪
⎪
⎩
i=1
×Ptr (u, i) × (1 − fc )}] ,otherwise
(4)
3. If we assume that u has performed the detecting procedure when r messages
have been transmitted within rmax rounds, the probability that u does not
perform the reporting procedure until the additional x messages have been
transmitted from from u to v can be calculated from Eq. (5).
⎧
⎪
⎪ 1 ,if x + r ≤ rmax
⎪
⎪
⎪
⎪
⎪
⎨
(x+r)modrmax
Pnotdet|reported (u, x, r) =
⎪
⎪ f (x+r)/rmax −1
× {1−Pnotr (u, i−1)
⎪
⎪ c
⎪
⎪
⎪
⎩
i=1
×Ptr (u, i)×(1 − fc )} ,otherwise
(5)
4. Since our scheme performs the bidirectional misused key detection, node v
also can perform the detecting procedure for Cu,v and Cv,u . Equation (6) thus
denotes the probability that both u and v does not perform the reporting
procedure until the additional x and y messages have been transmitted.
Figure 4 depicts the probability of detecting a misused key under assuming that
the average number of neighbors is b = 50, the total number of detecting nodes is
n
m = 10 and y = x. In Fig. 4(a), the detecting probability of our scheme is higher
1
than that of the Liu and Dong’s scheme under the same probability p(= rmax ).
Moreover, the detecting probability of our scheme with p = 0.1 is almost same
with that of the Liu and Dong’s scheme with p = 0.2. Figure 4(b) shows that
the probability of detecting a misused key is slowly decreased as the fraction of
the compromised detecting nodes is increased. Our performance is more slowly
decreased than the Liu and Dong’s scheme under the same probability p. This is
because our scheme performs the bidirectional misused key detection and uses
the probability model of LEACH.
The p is very important factor since it controls a tradeoff between the traffic
overhead and the detecting probability. The higher p is, the higher the detecting
probability is but the more the traffic overhead is increased. Our scheme increases
the probability of a misused key detection while consuming the lower traffic
overhead compared to the Liu and Dong’s scheme.
Parameter Value
Eelec 50nJ
Ef s 10pJ/bit/m2
Emp 0.0013pJ/bit/m4
Size of CV 8 bytes
Size of Cu,v 8 bytes
Size of message 80 bytes
Size of node ID 4 bytes
Size of MAC 8 bytes
Average of d 6m
simplistic, we have assumed that u and v always have a message to send and
transmit to each other by turns.
Ef s
l × (Eelec + Ef s × d2 ) , if d ≤ Emp
Etrans = { (9)
Ef s
l × (Eelec + Emp × d4 ) , if d > Emp
Our scheme has better two properties, which are the reduction of the detecting
overhead and the high detecting probability, than Liu and Dong’s scheme by
performing the bidirectional misused key detection, and removing of unnecessary
message and using the reasonable probability method for the balanced misused
detection among the sensor nodes. We can finally achieve an efficient mechanism
with a high detecting probability and low energy consumption. On the average,
our scheme is 36% more energy efficient than the Liu and Dong’s scheme and
the detecting probability of our scheme is higher than that of Liu and Dong’s
under the same reporting probability. The analysis shows that our scheme is
good enough for a misused key detection in the WSNs.
A further direction of this study will be applying our scheme into the existing
sensor networks with some specific routing protocols, key management mecha-
nisms, applications and so on. By applying, we will analyze the effectiveness of
our scheme.
References
1. Cerpa, A., Elson, J., Estrin, D., Girod, L., Hamilton, M., Zhao, J.: Habitat Monitor-
ing: Application Driver for Wireless Communications Technology. In: Proceedings
of the ACM SIGCOMMWorkshop on Data Communications in Latin America and
the Caribbean, San Jose, Costa Rica (2001)
2. Shen, C., Srisathapornphat, C., Jaikaeo, C.: Sensor Information Networking Ar-
chitecture and applications. In: IEEE Personal Communications, August 2001, pp.
52–59 (2001)
3. Liu, D., Ning, P.: Establishing pairwise keys in distributed sensor networks. In: Pro-
ceedings of the 10th ACM conference on Computer and communications security,
pp. 52–61. ACM Press, New York (2003)
4. Liu, D., Ning, P., Du, W.: Group-Based Key Pre-Distribution in Wireless Sensor
Networks. In: Proc. 2005 ACM Wksp. Wireless Security (WiSe 2005), September
2005, pp. 11–20 (2005)
5. Lui, D., Dong, Q.: Detecting Misused Keys in Wireless Sensor Networks. In: Perfor-
mance, Computing, and Communications Conference (IPCCC 2007), April 2007,
pp. 272–280 (2007)
6. Chan, H., Perrig, A., Song, D.: Random Key Predistribution Schemes for Sensor
Networks. In: Proc. IEEE Sec. and Privacy Symp., pp. 197–213 (2003)
7. Akyildiz, I.F., Su, W., Sankasubramaniam, Y., Cayirci, E.: Wireless Sensor Net-
works: A Survey. Computer Networks, 393–422 (2002)
8. Eschenauer, L., Gligor, V.: A Key Management Scheme for Distributed Sensor
Networks. In: Proc. 9th ACM Conf. Comp. and Commun. Sec., November 2002,
pp. 41–47 (2002)
9. Eltoweissy, M., Moharrum, M., Mukkamala, R.: Dynamic Key Management in
Sensor Networks. IEEE Communications Magazine 44, 122–130 (2006)
10. Camtepe, S.A., Yener, B.: Key distribution Mechanisms for Wireless Sensor Net-
works: A Survey, Technical Report TR-05-07, Rensselaer Polytechnic Institute
(March 2005)
11. Mhatre, V., Rosenberg, C.: Design Guidelines for Wireless Sensor Networks: Com-
munication, Clustering and Aggregation. Ad. Hoc. Networks, 45–63 (2004)
A Light-Weighted Misused Key Detection in WSNs 367
1 Introduction
Typhoon has a large low-pressure center and numerous thunderstorms that produce
strong winds and heavy rains. The strong winds and heavy rains usually cause land-
slides and mudslides. In Taiwan, there are sometimes over a dozen of typhoons vis-
ited in summer. In fact, the name Typhoon means “the deadly storm from Taiwan.”
On August 8th, 2009, Typhoon Morakot hit southern Taiwan. It is the deadliest ty-
phoon that sweeps Taiwan in recorded history. The storm brought a record-setting
rainfall, nearly 3000mm (almost 10 feet) rainfall accumulated in 72 hours. The rain-
fall spawned mudslide made a devastating damage to several villages and buried
hundreds of lives. There was not any warming mechanism worked when the mudslide
embodied as an advent devil. Threatened by such nature disaster, how good a disaster
warming system performs decides life and death.
A sensor system is one of the key tools to predict a disaster. Several previous stud-
ies proposed some detection techniques. One of the techniques uses the operating
theorem of a mechanical mouse as basic. Another uses the image processing to get the
D. Taniar et al. (Eds.): ICCSA 2010, Part III, LNCS 6018, pp. 368–379, 2010.
© Springer-Verlag Berlin Heidelberg 2010
Identifying Mudslide Area and Obtaining Forewarned Time 369
differences between two images. A key issue of a disaster detection system is the
communication network. With the rapidly emerging Advanced Metering Infrastruc-
ture (AMI), AMI communication network is a substantial candidate for constructing
the mudslide detection network. AMI Meter Data Management System (MDMS)
records since the location and status of each electricity pole. With such data, a mud-
slide warming system performs the detection and prediction processes. This study
also quantifies the degree of possible damage as the severity indicator. The mudslide
warming system triggers different alarms according to the given severity indicator. To
obtain a forewarned time, AMI MDMS needs to record the distances information for
each electricity pole and its related residence areas. This study then uses the distance
information to calculate the remaining forewarned time when an anomaly happens.
2 Background
Previous studies developed many detection and identification techniques for the land-
slides and mudslides, including Image enhancement [1, 2, 3], Image differencing [4,
5, 6], Vegetation index differencing [7], Image classification [3, 8] and Image regis-
tration [9]. Image enhancement emphasizes the landslides or mudslides in the satel-
lites images. This technique requires human experience and knowledge of the ob-
served areas for visual interpretation. Image differencing is a straightforward method
and easy to interpret the results. However, this method does not provide a detail
changed matrix. The selection of thresholds affects the correctness of the result.
Vegetation index differencing emphasizes the differences in spectral response. This
method reduces the impacts of topographic effects and illumination. However, this
technique also enhances the random noise or coherent noise. Image classification
minimizes the impact of atmospheric, sensor and environmental differences between
multi-temporal images. However, this method requires a sufficient set of the training
samples for classification. Image registration detects landslide/mudslide movements
with sub-pixel accuracy. The main concern is its requirements of the high computa-
tional cost in terms of CPU time. A simple method of change detection and identifica-
tion uses a local similarity measure based on the mutual information and the image
thresholding [11]. This method uses the mutual information to measure the similarity.
It then detects the landslides and mudslides from different images. However, this
method is not suitable for detecting a small area landslide. This method fails, if the
size of the used imagery is too large in compared with the landslide areas. Table 1
shows the ratio of the approximate size of the landslide over the size of the imagery.
Some studies use the wireless sensor networks (WSN) as the communication infra-
structure of their disaster detection system, such as: Slip Surface Localization in
Wireless Sensor Networks [10]. This method uses a sensor network consisting of a lot
sensor columns. This detection system deploys the sensor columns on hills to find the
early signals preceding a mudslide. This sensor network consists of a collection of
sensor columns placed inside the vertical holes drilled during the network deployment
phase. Installers arrange these sensor columns on a semi-regular grid over the moni-
tored area. The size of the grid depends on the characteristics of the site where the
370 C.-J. Tang and M.R. Dai
Table 1. Comparison on the ratio of the approximation of the landslide area with size of
imagery used. (Reprinted from [11])
sensors are deployed. Each sensor column has two components: the sensing compo-
nent that is buried underground and contains all the instruments, and the computing
component that stays above ground and contains the processor and radio module.
(a) (b)
Fig. 1. (a) Landslide Warning System (b) Sensor Column (Reprinted from [10])
The major problem of this technique is the deployment cost. To obtain a “good”
prediction, the density of the sensors must be high enough. In other words, if there is a
broad area requiring monitoring, the investment must be sky-high. Ironically, the
mudslide usually happens in an area that draws little economical interests. Therefore,
to protect people living in these areas, the detection devices must be inexpensive.
Furthermore, the mounting of these devices must be associated with some public
utilities system to reach where people are living.
3 System Architecture
A disaster warning system conducts the environment sensing, data analysis, and
communication. A database stores the periodically collected sensor data. The pro-
posed system analyses the gathered sensor data from the database that is a part of
AMI system. This disaster warning system uses the tilting degree of an electric pole
to define the severity of the location that the pole is set. The system stores every
Identifying Mudslide Area and Obtaining Forewarned Time 371
analyzed data in the same database. Using a specialized suffix-tree algorithm, the
warning system sends the forewarning signals according the information. Both data
gathering and message broadcasting use AMI-associated communication network.
Recently, many advanced areas such as Europe, America, and Japan etc. commit
themselves to develop the Smart Grid, SG. SG is an integration of electricity usage
monitoring, generation and distribution automation, meter data management and effi-
ciency improving. SG achieves the energy saving and carbon reduction. SG has func-
tions of energy transmission, energy management and two-way communication,
which involve energy generation, transmission, distribution and the customer site.
Advanced Metering Infrastructure (AMI) is regarded as the fundamental and key
technology of enabling SG.
AMI has a communication network which consists of many advanced metering and
sensing devices, including Smart meters. Smart meters provide interval usage (at least
hourly) and collect at least daily. According to the analysis from Gartner, AMI system
consists of following characteristics: data acquisition, data transform, data cleansing,
data processing, information storage/persistency.
AMI architecture is shown in Figure 2. The blue lines represent the power lines that
connect the transforming stations, utility devices, and customer sites. Each customer
site has a meter, a meter interface unit (MIU), and a communication module. The in-
formation of power consumption is sent to the data collector through the communica-
tion module. The data collectors also use the same communication channel to send
back messages. Each smart meter is connected to an energy management device. The
smart meter communicates with the control center to manage the energy consumption
at its location. The communication lines at the customer sites are presented in green.
For storing the collected data, there is a meter data management system (MDMS)
that communicates with the customer information portal, the customer information
management system, the dispatch automation system (DAS), and the data collector.
The status of each device within an AMI installation is sent to the data collector,
and then forward to the MDMS.
372 C.-J. Tang and M.R. Dai
This study proposes an inexpensive detection device that is mounted on each elec-
tricity pole. The movement or tilting of an electricity pole triggers the mounted detec-
tion device to send a message back to the data collector through the AMI communica-
tion network. This displacement message then forwards to the MDMS that activates
the mudslide analysis module for further calculations.
(a) (b)
Fig. 3. (a) Camera photographs neighbor one; (b) Camera adding the laser light
The weather affects the performance of this simple detection method. For example,
the heavy fog hampers the image identification. The unclear images increase the de-
tecting error. Adding a laser light projector in the upside of camera enhances this
situation. Laser light is usually spatially coherent, which means that the light either is
emitted in a narrow or low-divergence beam. Coherent light typically means the
source produces light waves that are in step.
Identifying Mudslide Area and Obtaining Forewarned Time 373
Another detection method, instead of using an expensive sensor device, the pro-
posed detection technique just uses a portion of a mechanical mouse as the detection
device. Everybody knows that when a mouse is moved, the ball inside rolls. This
motion turns two axles, and each axle spins a slotted wheel. A light-emitting diode
(LED) sends a path of light on one side of each wheel through the slots to a receiving
photo-transistor on the other side. The pattern of light to dark is translated to an elec-
trical signal, which reports the position and the moving speed of the mouse. In a
word, the operation of a mechanical mouse has five steps as the following: i) moving
the mouse turns the ball. ii) X and Y rollers grasp the ball and transfer movement. iii)
Optical encoding disks include light holes. iv) Infrared LEDs shine through the disks.
v) Sensors gather light pulses to convert to X and Y velocities.
To make this device, one needs to attach a small piece of iron or other material that
makes one side of the ball always facing down, and then mounts this “tilt-aware
mouse” on an electricity pole. If the pole is moved or tilted, the ball rolls and the
displacement messages are sent.
According how the rollers turns, the tilting direction is obtained. If the RollerA
turned to the left, the ball is moving forward, as shown in the Figure 4.
To obtain the tilted angle, the ball separates into x-z plane and y-z plane. From the
x-z plane, the radian represents the angle of turning left/right, and the vector is OD .
From the y-z plane, the radian represents the angle of turning forward or backward,
and the vector is OC . The two vectors OD and OC construct a plane ODCP with the
vector sum OP . The angle between OP and z-axis is the tilted angle of the electricity
pole.
4 Data Analysis
The displacement information of the tilted electricity pole is sent to an AMI data
collector periodically. The data collector then forwards the information to the MDMS.
MDMS then inserts a displacement record in the Pole_Displacement table. A dis-
placement record has the following fields: pole id, time, tilted direction, tilted angle,
difference of tilted angle, checking bit, and the tilted severity of an electricity pole.
{ t1 , t2 , t3 ,..., tn } denotes the set of recording time of each displacement. an denotes the
tilted angle. The differences between the continued tilted angles are obtained by
an − an −1 = d n and an −1 − an − 2 = d n −1 . d k denotes a difference between two continuous tilted
Identifying Mudslide Area and Obtaining Forewarned Time 375
angles of a pole. When d k is not zero, the checking bit ck is set and the tilted severity
is also calculated according to d k . If d k ≤ 2m , the tilted severity sk is m. If ck-1 is set, ck
is also set. If ck is set and dk is zero, sk is set to sk-1.
The pairs of checking bit and tilted severity of all electricity poles constructs the
alphabets in the proposed mudslide analysis system. The neighboring poles are as-
signed with adjacent numbers. The pair ci , si of time ti of all numbered electricity
poles constructs the Si according to the pole numbers. The mudslide analysis system
then builds a suffix tree STi of two adjacent Si and Si+1. The longest repeating pattern
in the STi of highest tilted severity indicates the most dangerous area that needs to put
most attentions.
Suffix tree is one of the data structure which uses to process a stream data. For ex-
ample, there is a string “mississippi”.
The main reason of using Suffix Tree as the analysis model is for its well-known
computation complexity of Ο(m) .
Several substrings have the same suffix such as “i”, “m”, “p”, and “s”. The sub-
strings use ‘is’ as the suffix that also use ‘issi’ as the suffix. In this situation, these
substrings share a node in the suffix tree. In the proposed mudslide analysis, this
study assumes that the repeating patterns with ci been set indicate where mudslide
will probably happen.
A table Pole_Displacement in the MDMS stores the status of all poles. The status
of a pole includes its id, time, tilt direction, tilt angle, the difference of tile angle,
checking bit, and the degree.
Field “pole_id” is a unique value that indicates an electricity pole in the system.
Field “time” uses the UNIX timestamp format, and denotes the time to store data.
Figure 4 shows the detection of tilted direction. Through the detection of two rollers
A and B, a detector finds the tilted direction. The roller A stays at X-axle for detecting
376 C.-J. Tang and M.R. Dai
a back-forth movement. The roller B stays at Y-axle for detecting the left-right
movement. Field “t_dir” is a 2-bit value that is for recording the back-forth, and left-
right movement. Field “t_angle” stores the tilted angle. Field “diff_angle” records the
differences of “t_angle”. The field “checkbit” denotes whether a pole is tilted or not.
A tilted pole marks “1”, otherwise “0”. Field “degree” denotes the severity degree of
a pole. diff _ angle ≤ 2 m , and the value m is the degree of severity. An example is
shown in Figure 7.
The MDMS stores the distances between each pole and the related residences to
obtain a forewarning time. Table Distance_Ri_Di stores the distances. There are two
fields in this table. Field “pole_id” is the unique id of a pole. The other one is
“Pairs_of_Ri_and_Di”. A pair (Ri, Di) denotes the distance Di between a pole Ek and a
residence Ri. An example is shown in Figure 8.
Each affecting scope of an electricity pole than contains a set of residence ID. This
system associates each electricity pole and every residence ID of this pole with an
offset that indicates the distance between the residence area and the pole.
{E1 , E 2 ,..., E k } denotes all poles. {R1 , R2 ,..., Ri } denotes the residence IDs of an
electricity pole k. Dik denotes the distance between Rik and Eik .Assume the probability
of landslide is P(x ) .
⎧1 , landslide happened
P( x) = ⎨
⎩< 1 , otherwise
At any given time after Ek has tilted, the first recorded time is t1, and t0 is subtracting
Δt from t1. Δt is the period time for retrieving data. The tilted displacement for pole Ek
378 C.-J. Tang and M.R. Dai
is m that is hk sin Δθ . Therefore, the tilted velocity for pole Ek is Vk. t 0k is the time
that pole hasn’t tilted for the pole Ek.
ck × hk sin Δθ
Vk = (1)
tnow − t0 k
The forewarned time of the pole Ek alerting to the residence Ri denotes Ri,AlertTime.
Di
Ri , AlertTime = (2)
ck × hk sin Δθ tnow − t0 k
5 Conclusion
The mudslides and landslides have consumed many lives. AMI communication net-
work is able to reach every residence that makes a substantial candidate for construct-
ing a mudslide detection network. The location and status of each electricity pole are
recorded in AMI MDMS. With an extra movement detector attached to each pole, an
AMI-associated sensor network is able to identify the possible mudslide areas from
the numerous gathered data. To obtain a “good” prediction, the density of the sensors
must be high enough. In other words, if there is a broad area requiring monitoring, the
investment must be sky-high. Ironically, the mudslide usually happens in an area that
draws little economical interests. Therefore, to protect people living in these areas, the
detection devices must be inexpensive. Furthermore, the mounting of these devices
must be associated with the public utilities system to reach where people are living.
Therefore, this paper proposes an inexpensive detection device associated with AMI
network, and a mudslide analysis method that indicates the severity and urgency of a
mudslide.
Identifying Mudslide Area and Obtaining Forewarned Time 379
This proposed device and method is still a concept rather than a realization. The fu-
ture work is to install such system in an AMI pilot installation to prove this concept.
References
1. Whitworth, M.C.Z., Giles, D.P., Murphy, W.: Airbone remote sensing for landslide hazard
assessment: a case study on the Jurassic escarpment slopes of Worcestershire, UK. The
Quarterly Journal of Engineering Geology and Hydrogeology 38(2), 197–213 (2005)
2. Ostir, K., Veljanovski, T., Podobnikar, T., Stancic, Z.: Application of satellite remote sens-
ing in natural hazard management: the Mount Mangart landslide case study. International
Journal of Remote Sensing 24(20), 3983–4002 (2003)
3. Nichol, J., Wong, M.S.: Satellite remote sensing for detailed landslide inventories using
change detection and image fusion. International Journal of Remote Sensing 9, 1913–1926
(2005)
4. Cheng, K.S., Wei, C., Chang, S.C.: Locating landslides using multi-temporal satellite im-
ages. Advances in Space Research 33, 296–301 (2004)
5. Hervas, J., Barredo, J.I., Rosin, P.L., Pasuto, A., Mantovani, F., Silvano, S.: Monitoring
landslides from optical remotely sensed imagery: the case history of Tessina landslide, It-
aly. Geomorphology 1346, 1–13 (2003)
6. Rosin, P.L., Hervas, J.: Remote sensing image thresholding methods for determining land-
slide activity. International Journal of Remote Sensing 26, 1075–1092 (2005)
7. Lin, W.T., Chou, W.C., Lin, C.Y., Huang, P.H., Tsai, J.S.: Vegetation recovery monitoring
and assessment at landslides caused by earthquake in Central Taiwan. Forest Ecology and
Management 210, 55–66 (2005)
8. Nichol, J., Wong, M.S.: Detection and interpretation of landslides using satellite images.
Land Degradation and Development 16, 243–255 (2005)
9. Yamaguchi, Y., Tanaka, S., Odajima, T., Kamai, T., Tsuchida, S.: Detection of a landslide
movement as geometric misregistration in image matching of SPOT HRV data of two dif-
ferent dates. Int. J. Remote Sensing, preview article:1 12 (2002)
10. Terzis, A., Anandarajah, A., Moore, K., Wang, I.-J.: Slip Surface Localization in Wireless
Sensor Networks for Landslide Prediction. In: IPSN 2006, Nashville, Tennessee, USA,
April 19-21 (2006)
11. Khairunniza-Bejo, S., Petrou, M., Ganas, A.: Landslide Detection Using a Local Similarity
Measure. In: Proceedings of the 7th Nordic Signal Processing Symposium, NORSIG 2006
(2006)
12. Mouse (computing) Wikipedia. Wikipedia (September 24, 2009),
http://en.wikipedia.org/wiki/Mechanical_mouse#Mechanical_mice
Utilization of Ontology in Health for Archetypes
Constraint Enforcement
Abstract. Most existing works in ontology deployment within the health indus-
try are mainly focusing on the standardization and interoperability goals. In this
paper, we propose the utilization of an ontology to apply a new constraint in
health archetypes, i.e. the slot filling constraint. An archetype is a model that
represents functional health concept such as admission record. It can reuse
other existing archetypes through a slot. The name of a slot represents a more
specific health concept such as head. The slot filling constraint restricts the se-
lection of archetypes to fill in that specific slot so that only relevant archetypes
are chosen from the available ones. Ontology is used to enforce this constraint.
An approach on how to apply the constraint is presented based on the semantic
similarity/relevance concept. The evaluation shows that the approach is a better
alternative to the current slot filling process which depends on manual decision
by the archetype author.
1 Introduction
Ontologies have been used widely in many areas today. It is primarily employed in
the applications where formal specifications of concepts are needed. One of the fields
that utilize ontology is health. There are even some ontologies related to health which
have been defined and become standards such as SNOMED CT (Systematized No-
menclature of Medicine-Clinical Terms)1 and LOINC (Logical Observation Identifi-
ers Names and Codes)2. However, the main use of ontologies in health so far is still
limited to the effort of achieving the uniform semanticity of health terms, in which the
clinical ontologies become standards. Another common aim of ontology usage in
health is to achieve interoperability between the different frameworks of electronic
health records (EHR), which are presented in e.g. [1], [2], and [3].
1
See: http://www.ihtsdo.org/
2
See: http://loinc.org/
D. Taniar et al. (Eds.): ICCSA 2010, Part III, LNCS 6018, pp. 380–394, 2010.
© Springer-Verlag Berlin Heidelberg 2010
Utilization of Ontology in Health for Archetypes Constraint Enforcement 381
This paper will present another kind of ontology employment in health, specifically
in the area related to EHR. Our focus is the EHR architecture proposed by openEHR3,
in which archetype lays as the basis of the model. The archetype concept was first
mentioned as part of the Australian project GEHR (Good Electronic Health Record)
[4]. The development of the archetype concept is mainly motivated by the inability of
the existing medical information system to keep up with the dynamic changing of
health knowledge. In the existing framework, which is commonly referred as “single-
level” methodologies, both informational and knowledge concepts are built into one
level of object and data models [5]. This methodology will become obsolete soon
after there is knowledge updating, and then should be replaced. This produces high
maintenance cost. The archetype concept is proposed to cope with the shortcoming of
the existing medical information model. The key feature of an archetype-based sys-
tem is that it is made up of a two-level architecture: a RIM (Reference Information
Model), consisting of a set of reference types, and a domain or knowledge model,
consisting of a set of archetypes [4]. The reference model is the repository of the
patient clinical data. As clinical knowledge is stored completely separately in
archetypes, resulting EHR systems are more flexible as changes in the clinical knowl-
edge can be embraced by modifications in archetypes, without compromising the
integrity of information in the repository [5]. An archetype itself formally expresses a
distinct health concept such as blood pressure. The concept is expressed in the form
of constraints on data whose instances conform to some RIM [6].
In this work, we are proposing a new archetype constraint, named as the archetype
slot filling constraint, which is defined as the complement of the existing constraints
specified in the archetype specifications. The proposal of the new constraint is moti-
vated by an example of the data inconsistency on the absence of this constraint. With
the enforcement of the constraint, we believe that new archetypes can be defined
more accurately and semantic inconsistency can be avoided. At the end, this will
improve the quality of the whole clinical systems in which archetypes are utilized.
As the constraint is not specified in the archetype specifications, a method will also
be proposed to make the constraint feasible to be applied without influencing the
internal structure of archetypes. Ontology, in this case clinical ontology, will be em-
ployed for the application of the approach. The method also uses the concept of se-
mantic similarity/relevance of an ontology. Of course, the concept needs to be ad-
justed to fit in the clinical ontology to be used.
Actually, the concept of ontology has been used in an archetype as one of its sec-
tions, but the usage is limited to terminologies (clinical ontology) linkings and bind-
ings. The binding is aimed at the uniformity in the meaning of terms contained in the
archetype, even though there is no special mechanism to check the uniformity. The
ontology is simply viewed as collection of terms, not as a ‘real’ ontology. This is
different to the approach in this work in which ontology will be utilized as the integra-
tion of all its concepts, relationships, and attributes.
The rest of the paper will be organized as follows. Section 2 will present the moti-
vating example on the proposal of the new constraint in archetype concept. Related
works is presented in Section 3. The approach to enforce the application of the
3
See: http://www.openehr.org/
382 A.K. Sari, W. Rahayu, and D. Wollersheim
constraint, as well as the related algorithms, is then explained in Section 4. The ap-
proach is evaluated in Section 5. Section 6 closes this paper with some conclusion and
future woks.
The semantic violation problem occurs because the user, i.e. the archetype author
should fill in the archetype slots without the guidance on which archetypes can be
chosen and which ones should be excluded due to their irrelevancy to the sub section
name. In this example, user should inspect the list of the fifty Observation archetype
names and decide by himself which names can fill in the slots. Since archetype con-
cept is expected to be used widely in the future, and thus, more and more new arche-
types will be created, it will be more complicated for the user to choose the relevant
archetypes when there will be many more than fifty Observation archetypes.
This problem could be prevented if there is a constraint in filling in the slot for ar-
chetypes. This constraint should determine the dependency of the archetype slot fill-
ing to the sub section it is related to or to the archetype itself. We refer this constraint
as archetype slot filling constraint. A method that utilizes ontology can be applied to
enable the slot filling constraint. This method will provide the user with the relevant
archetypes to be filled in a particular slot. The relevancy is measured based on a par-
ticular clinical ontology. However, since the archetype should neutral with respect to
terminologies [6], the utilization of terminology/ontology should be conducted out-
side the archetype definition. The detail of the method is discussed in Section 4.
3 Related Works
There are not many works which utilize ontologies for the works related to arche-
types. However, to the best of our knowledge, there is no existing paper which
discusses the employment of ontology for the validation of archetype constraints.
384 A.K. Sari, W. Rahayu, and D. Wollersheim
Rather, most of the works use ontology as a mediator to achieve semantic interop-
erability between the different frameworks of EHR. For example, [1] shows that the
interoperability between archetypes based web service messages could be accom-
plished by using OWL (Web Ontology Language) as the representation of both ar-
chetype definitions and instances to semantically annotate the web service messages.
Similarly, in [2] one of clinical ontologies i.e. SNOMED CT is used as the standard
into which archetypes and messages are bound. The aim is to allow different systems
to interoperate. And finally, Fernandez-Breis, et al., in [3], presents an ontological
approach to promote interoperability among CEN(European Committee for Stan-
dardization)/TC 2514 and openEHR compliant information systems by facilitating
the construction of interoperable clinical archetypes. This is achieved by construct-
ing a one to one mapping between the models, utilizing OWL as the ontological
representation.
The concept of semantic relevance is also employed in this paper. It is used in the
approach to apply the new constraint in the archetypes. Some works which are related
to the relevance semantic issue have been done. In most of them, e.g. [9], [10], [11],
[12], [13], and [14], the semantic relevance term is related to, or based on, or even
used interchangeably with the terms semantic similarity and/or semantic distance.
Hence, in this work we will also base the semantic relevance measurement on seman-
tic similarity measurement.
There have been many papers related to semantic similarity. According to [15],
based on the structures and input considered, there are three kinds of similarity meas-
ures: edge-based, information content-based, and feature-based approaches. Edge-
based approach is used in a graph representation of an ontology. This approach is
used in e.g. [16] and [17]. Information content-based method, which is applied in
[18], [19], uses text corpus to evaluate the semantic similarity. The feature-based
technique utilizes the properties of concepts for assessing the similarity and is used in
[20]. Some works, such as [21], [22], and [23], use the combination of those methods
or compare them.
To the best of our knowledge, none of the papers related to semantic similarity dis-
cuss its application to clinical ontology, which is the focus of this paper. Clinical
ontology, in this case SNOMED CT, has some special characteristics. First, the mean-
ing of attribute is not the same as attribute in common ontologies. In SNOMED,
attribute is a kind of relationship which relates a concept to another. This different
meaning should be taken into account when feature-based similarity techniques are
used, as they commonly base the measurement on attribute. The next characteristic is
that in SNOMED, the hierarchy is based only on the IS-A relationships. Other types
of relationships are not considered. This will influence the calculation when the edge-
based similarity approach, which is based on all types of relationships, is used. Thus,
there should be adjustment when the approach is applied for SNOMED CT clinical
ontology.
4
CEN/TC 251 is standard for health informatics. The complete published standards, see:
http://www.cen.eu/CENORM/Sectors/TechnicalCommitteesWorkshops/CENTechnicalCom
mittees/CENTechnicalCommittees.asp?param=6232&title=CEN/TC%20251
Utilization of Ontology in Health for Archetypes Constraint Enforcement 385
As previously mentioned, we propose the ontology based approach to apply the slot
filling constraint to archetype. We believe that ontology is suitable to be utilized to
force this constraint for the following reason. The constraint is much related to the
semantic data inconsistency. As ontology is a formal specification, it can maintain
semantic consistency of the concepts it specifies. As the result, the problem of data
inconsistency can be avoided.
In this work, SNOMED CT is chosen as the ontology for the approach. The choice
is based on some reasons. Firstly, SNOMED CT is considered to be the most compre-
hensive, multilingual clinical healthcare ontology in the world [24]. It covers areas
that are closely related to EHR, the main concern of archetypes. Secondly, some
works such as [2] and [25] use SNOMED CT as the ontology for archetype related
works. This shows that SNOMED CT fits well to the concepts defined in archetypes.
Since the slot filling constraint involves only particular classes of archetypes, i.e.
Composition, Section, and Entry, this paper only considers those types of archetypes
as well. For the purpose of clearness of the method used, we categorize those classes
of archetypes into three class layers: the bottom layer, the middle layer, and the top
layer. The bottom layer is the classes of archetypes which can only be the filler of
slots. They are not able to have any slot to be filled in by another archetype. All the
entry classes, i.e. the Instruction class, the Observation class, the Action class, the
Evaluation class, and the Admin Entry class, are included in this layer. The top layer
is the class of archetypes which can only have slots inside them. They themselves
cannot fill in another archetype’s slot. This layer includes the Composition class. The
middle layer comprises the Section class. This layer of archetypes can have slots to be
filled in by other archetypes as well as be the filler of other archetypes. In the rest of
this paper, we will use the layers categorization to refer to the classes of archetypes.
Table 1. summarizes this categorization.
The ontology based approach to force the constraint application is explained intui-
tively in this following scenario. An archetype author needs to fill in the slots with the
include and exclude assertion. The archetypes to be included or excluded in the
slot are chosen from some groups of a specific type of archetypes which can be one of
the bottom or middle layer class. The archetype type is chosen by the author. Each
386 A.K. Sari, W. Rahayu, and D. Wollersheim
archetype of the determined type is a candidate for the include assertion. Since
there can be many candidates, it is essential to assure that the chosen archetypes are
indeed relevant to the sub section or the Section/Composition archetype that contains
the slot. For this purpose, ontology is employed. It is used to annotate each archetype
with the concepts it specifies which are related to the archetype. The annotation proc-
ess is performed directly after an archetype of bottom or middle layer is built. The
archetypes of upper layer, i.e. the Composition archetypes, do not need to be anno-
tated because they are never used as slot fillers.
When an archetype is included as an archetype candidate to fill in a slot, its annota-
tion will be used in the filtering process. This process filters the archetypes in the
candidate list which are relevant to the slot. The relevancy is measured by inspecting
the relevance of each annotation of each archetype candidate with the concept in the
ontology which is related to the keyword. The user chooses the keyword from the
name of the sub section where the slot is contained, the names of all the ancestral sub
sections, the name of the archetype, or the combination of them which then become a
set of keywords. Each archetype candidate is then evaluated whether it is considered
relevant or not to fill in the slot. If it is, then the candidate is included in the list of the
relevant archetypes for the slot, otherwise it is excluded. The list is then used to force
the slot filling constraint to the slot filling activity by the user.
From the above description of the method, we determine two technical steps in the
approach: the preliminary process and the filtering process. In this paper, we focus on
the filtering process, thus we will describe the preliminary process briefly, while the
filtering process is described in detail in Section 5.2.
The preliminary process refers to the archetype annotation with respect to the
SNOMED CT ontology. Each archetype is annotated with some terms which repre-
sent the concepts in SNOMED CT which the archetype relates to. The annotation is
applied to any archetype of bottom and middle layer upon its development. Thus, it is
not limited to the archetypes in the candidate list. The process involves two main
steps: the ontology extraction and the archetypes annotation. The ontology extraction
is performed to obtain only the relevant part of the whole ontology, which is called
sub-ontology or ontology view. We reuse the ontology extraction proposed in [26] for
this work. The annotation is used in this work to annotate a specific archetype docu-
ment with the concepts specified in SNOMED CT which are related to its content.
Some annotation tools, such as SMORE [27] and Ontomat [28] can be utilized for the
purpose.
This is the main process of the method. In this process, the relevant archetypes for the
slot to be filled in are determined using a specific evaluation method. First, the arche-
type candidate list is prepared. This is based on the specific type of archetype chosen
by the user, i.e. the archetype author, for the archetype slot to be filled in. then, the
evaluation of relevant archetypes is conducted. Two steps need to be conducted here,
i.e. the determination of the set of keywords and the relevance calculation, which
result in the list of relevant archetypes. For the user, choosing archetypes not included
in the list will violate the archetype slot filling constraint.
Utilization of Ontology in Health for Archetypes Constraint Enforcement 387
The method for the relevance evaluation will be elaborated. The relevance of a
concept included in the annotation of a specific archetype to the keyword is evaluated
in a simple way. Algorithm 1 below describes the evaluation method.
The relevance evaluation is calculated for each keyword. The keywords are chosen
by the user based on the name of archetypes and the name of sub sections and all its
descendants where the slot is placed. For each keyword, each archetype is evaluated its
relevance to the keyword using its annotations. Each annotation is inspected whether it
is considered relevant to the keyword or not. Once it finds that an annotation of the
388 A.K. Sari, W. Rahayu, and D. Wollersheim
archetype candidate is considered relevant to the keyword based on the ontology, it will
add that archetype to list R, which includes the relevant archetypes for the slot. Thus,
the evaluation of the rest of the annotations is not needed. Moreover, the archetype will
be deleted from list P (list of candidate archetypes). It means that for the relevance
evaluation of the next keyword, it will not be considered anymore.
For evaluating the relevance between the keyword k with concept Pij in the arche-
type annotation, first one or more concepts which have the names similar to k are
searched in SNOMED. Each concept found is then evaluated its relevance to Pij based
on SNOMED CT ontology. Since the basis is an ontology, certainly the word rele-
vance here refers to the semantic relevance, which means the closeness of the rela-
tionship between two terms which in this work refers to the concepts contained in
SNOMED CT ontology. As previously mentioned in Section 3, semantic relevance is
much related to semantic similarity. Thus, some previous works on semantic similar-
ity are reused here with some adjustment.
To measure the semantic relevance between two concepts in SNOMED CT, we
first determine some features which influence the assessment. The factors are based
on the edge-based and information content based approaches of similarity calculation.
They are described below.
1. The distance between the concepts
In edge-based approach, the shortest path between two concepts in the semantic
net is commonly used to evaluate their similarity. The shortest path of concept a
and b is the path where the number of edges in the path is minimum among all
the paths connecting a and b in the nets. Rada et.al. [16] propose a basic semantic
distance measure by only using taxonomic links and computing the shortest path
between the concepts. A variant of the work is [29] in which distance is substi-
tuted by similarity. Thus, we consider the equation is more suitable to our work
and it is reused in this paper. By adopting the equation in [29], we define the
function for semantic relevance as follows:
(1)
where rele(a,b) denotes the edge-based relevance between concept a and concept
b, dist(a, b) represents the number of edges of the shortest path between concept
a and b, and Max is the maximum depth of the taxonomy.
2. The link type connecting the concepts
There are 4 types of relationships in SNOMED CT: defining characteristics (IS-A
relationships and defining attributes), qualifying characteristics (non-defining,
qualifying attributes), historical relationships (relate inactive concepts to active
concepts), and additional relationships (other non-defining characteristics such as
PART OF) [30]. The type of relationship that form the hierarchy of SNOMED
CT is the IS-A relationship. The other types of relationships are not represented
as links in the hierarchy. Two special cases occur here and we consider that the
relevance between two concepts is absolute if at least one of them happens. The
first case is when the two concepts are actually the same concept; the other is
when one of the concepts or one of their descendants becomes an attribute of the
Utilization of Ontology in Health for Archetypes Constraint Enforcement 389
other concept. The second case is based on the consideration that if the annota-
tion (which is a SNOMED CT concept) of an archetype becomes an attribute of
the keyword (which is also a SNOMED CT concept) of a slot, then the archetype
should be relevant for the slot. It is unlikely that the attribute-based relationships
form a path. Thus, we only consider their existence, not their path length. On the
other hand, the PART-OF relationships often form a path. Thus, the PART-OF
path distance between concept a and b needs to be considered. To differentiate
the IS-A and PART-OF relationships, we define the equations below which are
based on (1):
(2)
(3)
where simIS (a, b) denotes similarity between concepts a and b by using the IS-A
hierarchy, distIS (a,b) is the shortest path between concepts a and b based on the
IS-A hierarchy, MaxIS is the maximum depth of the IS-A taxonomy, simPART (a,
b) represents the similarity between concepts a and b based on the PART-OF re-
lationship, distPART (a,b) is the shortest distance between concepts a and b on the
PART-OF path, and LenIS (a, b) is the maximum length of PART-OF path which
contains a and b.
Then, we need to combine the equations and the consideration of attribute re-
lationship to get the relevance value between concepts a and b, as follows:
(4)
(5)
To get the final relevance equation, we combine (4) and (5). The following is the final
equation.
(6)
390 A.K. Sari, W. Rahayu, and D. Wollersheim
Rel(a, b) is the relevance between concepts a and b, while α1 and α2 are adjustable
factors, α1 + α2 = 1. α1 and α2 can be adjusted to the requirement. For example, when
the information about the text corpus is missing, α2 can be set to 0 to eliminate the
information content-based relevance from the equation.
Finally, equation (6) is connected with Algorithm 1 to determine the value of vari-
able relv. Equation (9) is used for the purpose.
(7)
Table 2. The comparison on some statistical metrics between the result using the method and
On the other hand, the recall value of the result using the method is generally
higher than the current slot filling. Again, the exception is for two slots, i.e. the En-
counter and the Antenatal check slots, as those two slots included all Evaluation ar-
chetypes. This makes the recall value to be maximum, because the inclusion of all
archetypes results in the inclusion of all relevant archetypes as well.
The accuracy and F-measure metric calculations normalize the opposite results of
the recall and precision metrics. The method generally has higher accuracy than the
current slot filling condition, which means that the method produces the closer results
to the actual relevant archetypes for each specific slot. The result of the F-measure
calculation is also higher. This shows that the method is more effective in retrieving
relevant archetypes for a specific slot than the current slot filling. From the result of
the two metrics, the slot filling process using the method seems to be a better alterna-
tive for the current slot filling condition.
The number of keywords does not seem influencing much on the result. For the
slots with multiple keywords, the accuracy value ranges from 0.69 to 0.85, while
that of the slots with single values ranges from 0.65 to 0.81. It shows that the num-
ber of keywords does not influence the accuracy. A slightly difference measure
happens to the F-measure metric, in which the slots with multiple keywords have a
little bit higher F-measure value. This means that the method is more effective to be
applied to the slots which are contained in a sub section rather than directly in an
archetype.
392 A.K. Sari, W. Rahayu, and D. Wollersheim
6 Conclusion
In this paper, we have presented the new constraint, i.e. the slot filling constraint as
the complement of the existing archetypes’ built-in constraints. The constraint is im-
portant to maintain semantic consistency between the archetype or sub section names
containing the slot and the filling archetypes. An ontology-based approach has been
proposed to validate the constraint on the development of new archetypes. This ap-
proach also shows a different way of utilization of ontology in health from most of the
existing ontology employment, which goal is for interoperability purpose.
The evaluation result shows that the method is a better alternative to the current
slot fillings condition. Based on the manual decision of the slots filling, the precision,
the recall, the accuracy, and the F-measure of the slot filling results using the method
are calculated. Compared to the current slot filling condition, the accuracy and F-
measure of the slot filling using the method is generally better. The evaluation also
shows that the F-measure for the slots contained in a sub section is better than those
contained directly in an archetype. However, the accuracy of the method is not much
different.
While there are two classes of archetypes, i.e. Section and Composition, can have
slots, in the evaluation the method is only applied to some Section archetypes. Further
application to all other archetypes, including the Composition archetypes, can be done
as part of our future works. Another interesting issue is the application of the ap-
proach to another existing clinical ontology, as some archetypes have the binding to
them as well.
Acknowledgement
This work is partially supported by the Ministry of National Education of the Repub-
lic of Indonesia through the scholarship granted to the first author.
References
1. Bicer, V., Kilic, O., Dogac, A., Laleci, G.B.: Archetype Based Semantic Interoperability of
Web Service Message in the Health Care Domain. International Journal on Semantic Web
and Information Systems 1, 1–23 (2005)
2. Qamar, R., Rector, A.: Semantic Issues in Integrating Data from Different Models to
Achieve Data Interoperability. In: MEDINFO 2007, Brisbane (2007)
3. Fernández-Breis, J.T., Menárguez-Tortosa, M., Moner, D., Valencia-García, R.,
Maldonado, J.A., Vivancos-Vicente, P.J., Miranda-Mena, T.G., Martínez-Béjar, R.: An
Ontological Infrastructure for the Semantic Integration of Clinical Archetypes. In: Hoff-
mann, A., Kang, B.-h., Richards, D., Tsumoto, S. (eds.) PKAW 2006. LNCS (LNAI),
vol. 4303, pp. 156–167. Springer, Heidelberg (2006)
4. Wollersheim, D., Sari, A., Rahayu, W.: Archetype-based Electronic Health Records: a Lit-
erature Review and Evaluation of their Applicability to Health Data Interoperability and
Access. Health Information Management Journal 38, 7–17 (2009)
5. Leslie, H., Heard, S.: Archetypes 101. In: Westbrook, J., Callen, J. (eds.) HIC 2006, Syd-
ney (2006)
Utilization of Ontology in Health for Archetypes Constraint Enforcement 393
6. Beale, T., Heard, S.: Archetypes Definitions and Principles, The openEHR Foundation
(2007)
7. Beale, T.: Archetypes: Constraint-based Domain Models for Future-proof Information
Systems. In: Baclawski, K., Kilov, H. (eds.) The International Conference on Object Ori-
ented Programming, Systems, Languages and Applications 2002, Seattle, Washington, pp.
1–18 (2002)
8. Beale, T., Heard, S.: Archetype Definition Language ADL 1.4, The openEHR Foundation
(2007)
9. Ruotsalo, T., Hyvönen, E.: A Method for Determining Ontology-Based Semantic Rele-
vance. In: Wagner, R., Revell, N., Pernul, G. (eds.) DEXA 2007. LNCS, vol. 4653, pp.
680–688. Springer, Heidelberg (2007)
10. Lombardi, L., Sartori, G.: Concept Similarity: An Abstract Relevance Classes Approach.
In: Fum, D., Missier, F.D., Stocco, A. (eds.) The 7th International Conference on Cogni-
tive Modeling, Trieste, pp. 190–195 (2006)
11. Sartori, G., Lombardi, L.: Semantic Relevance and Semantic Disorders. Journal of Cogni-
tive Neuroscience 16, 439–452 (2004)
12. Zhang, G., Yu, C., Cai, D., Song, Y., Sun, J.: Research on Concept-Sememe Tree and Se-
mantic Relevance Computation. In: The 20th Pacific Asia Conference on Language, In-
formation and Computation, vol. 20, pp. 398–402 (2006)
13. Rhee, S.K., Lee, J., Park, M.-W.: Ontology-based Semantic Relevance Measure. In: The
First International Workshop on Semantic Web and Web 2.0 in Architectural, Product and
Engineering Design, Korea (2007)
14. Ricklefs, M., Blomqvist, E.: Ontology-Based Relevance Assessment: An Evaluation of
Different Semantic Similarity Measures. In: Meersman, R., Tari, Z. (eds.) OTM 2008, Part
II. LNCS, vol. 5332, pp. 1235–1252. Springer, Heidelberg (2008)
15. Raftopoulou, P., Petrakis, E.: Semantic Similarity Measures: a Comparison Study. Techni-
cal University of Crete, Department of Electronic and Computer Engineering (2005)
16. Rada, R., Mili, H., Bicknell, E., Blettner, M.: Development and Application of a Metric on
Semantric Nets. IEEE Transactions on Systems, Man, and Cybernetics 19, 17–30 (1989)
17. Lee, J.H., Kim, M.H., Lee, Y.J.: Information Retrieval Based on Conceptual Distance in
IS-A Hierarchies. Journal of Documentation 49, 188–207 (1993)
18. Resnik, P.: Semantic Similarity in a Taxonomy: An Information-Based Measure and Its
Application to Problems of Ambiguity in Natural Language. Journal of Articial Intelli-
gence Research 11, 95–130 (1999)
19. Ljubešić, N., Boras, D., Bakarić, N., Njavro, J.: Comparing Measures of Semantic Similar-
ity. In: 30th International Conference on Information Technology Interfaces, Cavtat (2008)
20. Rodriguez, M.A., Egenhofer, M.J.: Comparing Geospatial Entity Classes: An Asymmetric
and Context-Dependent Similarity Measure. International Journal of Geographical Infor-
mation Science 18, 229–256 (2004)
21. Liu, M., Shen, W., Hao, Q., Yan, J.: A Weighted Ontology-based Semantic Similarity Al-
gorithm for Web Service. Expert Systems with Applications 36, 12480–12490 (2009)
22. Guisheng, Y., Qiuyan, S.: Research on Ontology-Based Measuring Semantic Similarity.
In: International Conference on Internet Computing in Science and Engineering, Harbin,
pp. 250–253 (2008)
23. Xu, X.-h., Huang, J.-l., Wan, J., Jiang, C.-f.: A Method for Measuring Semantic Similarity
of Concepts in the Same Ontology. In: 2008 International Multi-symposiums on Computer
and Computational Sciences, Washington DC, pp. 207–213 (2008)
24. IHTSDO: SNOMED-CT, http://www.ihtsdo.org/snomed-ct/
394 A.K. Sari, W. Rahayu, and D. Wollersheim
25. Sundvall, E., Qamar, R., Nyström, M., Forss, M., Petersson, H., Åhlfeldt, H., Rector, A.:
Integration of Tools for Binding Archetypes to SNOMED CT. In: SMCS 2006, Copenha-
gen, pp. 64–68 (2006)
26. Wouters, C.: A Formalization and Application of Ontology Extraction. PhD Thesis. De-
partment of Computer Science and Computer Engineering, La Trobe University, Mel-
bourne (2005)
27. Mindswap: SMORE - Create OWL Markup for HTML Web Pages,
http://www.mindswap.org/2005/SMORE/
28. Ontomat Homepage - Annotation Portal,
http://annotation.semanticweb.org/ontomat/index.html
29. Leacock, C., Chodorow, M.: Combining Local Context with WordNet Similarity for Word
Sense Identification. In: WordNet: A Lexical Reference System and its Application, pp.
265–283. MIT Press, Cambridge (1998)
30. IHTSDO: SNOMED Clinical Terms User Guide,
http://www.ihtsdo.org/fileadmin/user_upload/Docs_01/
Technical_Docs/SNOMED_CT_User_Guide_20080731.pdf
31. Jiang, J.J., Conrath, D.W.: Semantic Similarity Based on Corpus Statistics and Lexical
Taxonomy. In: International Conference Research on Computational Linguistics X, Tai-
wan (1997)
32. Mihalcea, R., Corley, C., Strapparava, C.: Corpus-based and Knowledge-based Measures
of Text Semantic Similarity. In: AAAI 2006 Conference, pp. 775–780 (2006)
Time-Decaying Bloom Filters for Efficient
Middle-Tier Data Management
Kai Cheng
1 Introduction
D. Taniar et al. (Eds.): ICCSA 2010, Part III, LNCS 6018, pp. 395–404, 2010.
c Springer-Verlag Berlin Heidelberg 2010
396 K. Cheng
2 Related Work
we say data is living, otherwise we say it is dead. Later we will discuss in details
how to set the value of vi ’s adaptively.
The recency statistic can be obtained by timestamp when the data item was
referenced. As users may not actually download a data item from the system,
instead they may go directly to the origin site. Thus we have to clarify the
meaning of data item reference.
in the set is hashed into several locations of the bit–vector using different hash
functions. Bits at these locations are then turned on. In order to find whether
an item is a member of the set, we first compute the locations to which the
item is mapped according to the same hash functions. We then check if all
bits at these locations are on. If so, the answer is yes; otherwise answer no. As
a bit can be set by other items due to the hash collisions, false positives are
possible.
Standard Bloom Filters have several limitations when extended to deploy
in other applications. First, Bloom Filter does not support deletes as simply
turning off the corresponding bits may introduce false negative errors (some bits
of other items, although still in the set, may be turned off). Second, it is note
suitable for dealing with multiset (set with duplicates), where multiplicities of
items should be reported. To extend standard Bloom Filters several variants
have been developed [6,7,8,9].
......
0000
0010
0000
0111
0000
0012
0000
0101
0000
0000
0110
0000
0000
0000
0000
0001
0100
Fig. 1. Basic Time-Decaying Bloom Filters
T and λ are parameters that control the behavior of how fast the counter
will decay. T is the period between two consecutive applications of φ. In other
words, it controls the granularity of time–sensitivity. A time period of T time
units is referred to as an epoch. The parameter λ ∈ [0, 1], called exponential
decaying factor or simply decaying factor, controls the speed of exponential
decaying. In a time–decaying Bloom Filter, basically we use the exponential
decay since it is most commonly used decay function in practice. An expo-
nential decay is a function of the the form g(x) = λx , λ ∈ [0, 1]. A decaying
counter c is maintained in such a way that its value ct at time t is obtained as
follows:
ct = λ · Δt + (1 − λ) · ct−1 (1)
Using this equation, we need only one counter for each value. We can increment
the counter directly after the decay function was applied, that when a new
(1 − λ)ct−1 was obtained. Whenever, say at time t we need the value of c, the
result of ct can be obtained by the following equation.
ct = λ · ct (3)
Note that when the value of a counter can become very large and the exact
counts are not necessary, some form of probabilistic counting can be adopted
[12,13].
Time-Decaying Bloom Filters for Efficient Middle-Tier Data Management 401
BF SBF Pool
0 00
h01 01
1 h11 12 Lookup
02
q h02 0 h12 02
q 05
0 00
1 00
h03 0 h13 00
1 03
0 15
values of 01 and 02 respectively. Assuming all are 4-bit counters, the frequency
count for q by the SBF counter Ch12 (q) is 256 × 1 + 2 × 16 + 0 = 288.
The usage of web is well known to be quite biased, with a small fraction of
popular sites getting very high hits, while the rest and also the majority are
rarely used. As the values of counters vary significantly from very small values
for most ”cold” pages to thousands of hits for a few ”hot” pages it is not suitable
to allocate the bits to count each of these items.
To avoid allocating large counters for many ”cold” pages, we optimize the
basic SBF as well as TMF using a dynamic representation of large counters.
The optimization is do in the following way:
1. Allocate a small counter to each cell, for example 4 bits or 8 bits.
2. Introduce a free counter pool (Pool) and dynamically allocate to large coun-
ters, each may be 16 or 32 bits. The Pool can also be organized as a hierarchy
with counters of different length.
3. Allocate a flag bit for each small counter to indicate whether it contains
the real frequency value or just a index to a large counter. Its initial value
is on or 1 indicating real value contained. When a small counter becomes
overflow, turn off the flag bit and allocate a free large counter from the Pool
and record its index in the small counter.
The query process will begin with a flag bit check. If it is on, return the value
in small counter as the frequency count. Otherwise use the value as an index to
find the large counter and return value of that counter.
5 Preliminary Experiments
We evaluate the optimization of the proposed Space–Optimized Counting Bloom
Filters. The data set we will use is proxy web access logs of 12 days from June 13
to 24,2003 provided by IRCache 1 . The sanitized cache access logs contain 541.4
MB data in gzipped form. We preprocess the data set by excluding accesses
without a success status code 200 and keep requests of htmltext mime type. The
cleaned data set consists of as many as total 2, 626, 434 requests in total and
1, 027, 679 distinct requests. Fig. 3 shows that it follows Zipf-like distribution.
First we compare the basic TBF, or B–TBF with its optimization called O–
TBF, with different values of λ. The result is shown in Figure 4. With the
increase of λ, the number of large counters in use increase. Thus more memory
space is consumed.
We then evaluate the the basic TBF, or B–TBF with its optimization called
O–TBF, with different window sizes. The result is shown in Figure 5. With the
increase of window size, the number of large counters in use increase and more
memory space is required.
From Figure 4 and Figure 5, we see the optimization by allocating large coun-
ters dynamically on demand is effective.
1
http://www.ircache.net/
Time-Decaying Bloom Filters for Efficient Middle-Tier Data Management 403
㪈㪇㪇㪇㪇㪇
㪈㪇㪇㪇㪇
㩺㩷㫆㪽㩷㫉㪼㪽㪼㫉㪼㫅㪺㪼㫊
㪈㪇㪇㪇
㪈㪇㪇
㪈㪇
㪈
㪈 㪉㪐 㪌㪎 㪏㪌 㪈㪈㪊 㪈㪋㪈 㪈㪍㪐 㪈㪐㪎 㪉㪉㪌 㪉㪌㪊 㪉㪏㪈 㪊㪇㪐 㪊㪊㪎 㪊㪍㪌 㪊㪐㪊 㪋㪉㪈 㪋㪋㪐 㪋㪎㪎 㪌㪇㪌
Frequenct Rankings
㪈㪍㪇
㪈㪋㪇
㪤㪼㫄㫆㫉㫐㩷㪪㫀㫑㪼㩷㫀㫅㩷㪢㪙
㪈㪉㪇
㪈㪇㪇
㪙㪄㪫㪙㪝
㪏㪇
㪦㪄㪫㪙㪝
㪍㪇
㪋㪇
㪉㪇
㪇
㪈 㪉 㪊 㪋 㪌 㪍 㪎 㪏 㪐 㪈㪇 㪈㪈 㪈㪉 㪈㪊
㫍㪸㫃㫌㪼㫊㩷㫆㪽㩷㫃㪸㫄㪹㪻㪸㩷㫇㪸㫉㪸㫄㪼㫋㪼㫉㩿㩼㪀
㪈㪍㪇
㪈㪋㪇
㪤㪼㫄㫆㫉㫐㩷㪪㫀㫑㪼㩷㫀㫅㩷㪢㪙
㪈㪉㪇
㪈㪇㪇
㪙㪄㪫㪙㪝
㪏㪇
㪦㪄㪫㪙㪝
㪍㪇
㪋㪇
㪉㪇
㪇
㪈 㪉 㪊 㪋 㪌 㪍 㪎 㪏 㪐 㪈㪇 㪈㪈 㪈㪉 㪈㪊
6 Concluding Remarks
To capture the evolving nature of data in middle–tier data management, we
proposed a semi–persistence model, which enhances the cache approach by fea-
tures like trend–report queries as well autonomous life cycle management. The
proposed model relies on efficient management of statistics of the underlying
data over time. We proposed Time-Decaying Bloom Filters as an efficient data
structure to manage time-sensitive access patterns of the underlying data. As
a future direction, a suitable scheme for adaptive management of the time to
live value of vi for each element is importance. Results from cache coherency
maintenance can be used but more semantic factors should be considered.
References
1. Cooper, B.F., Sample, N., Franklin, M.J., Olshansky, J., Shadmon, M.: Middle-tier
extensible data management. World Wide Web 4, 209–230 (2001)
2. Luo, Q., Krishnamurthy, S., Mohan, C., Pirahesh, H., Woo, H., Lindsay, B.G.,
Naughton, J.F.: Middle-tier database caching for e-busines. In: Proceedings ACM
SIGMOD Conference on Management of Data (SIGMOD 2002), Madison, Wiscon-
sin, pp. 600–611 (2002)
3. Team, T.: Mid-tier caching: The timesten approac. In: Proceedings ACM SIGMOD
Conference on Management of Data (SIGMOD 2002), Madison, Wisconsin, pp.
588–593 (2002)
4. Stonebraker, M.: Managing persistent objects in a multi-level store. In: Proceedings
ACM SIGMOD Conference on Management of Data (SIGMOD 1991), Denver, CO,
pp. 2–11 (1991)
5. Bloom, B.H.: Space/time trade-offs in hash coding with allowable errors. Commun.
ACM 13, 422–426 (1970)
6. Bonomi, F., Mitzenmacher, M., Panigrahy, R., Singh, S., Varghese, G.: An im-
proved construction for counting bloom filters. In: Azar, Y., Erlebach, T. (eds.)
ESA 2006. LNCS, vol. 4168, pp. 684–695. Springer, Heidelberg (2006)
7. Cohen, S., Matias, Y.: Spectral bloom filters. In: SIGMOD 2003, pp. 241–252.
ACM Press, New York (2003)
8. Deng, F., Rafiei, D.: Approximately detecting duplicates for streaming data using
stable bloom filters. In: SIGMOD 2006, pp. 25–36. ACM Press, New York (2006)
9. Fan, L., Cao, P., Almeida, J., Broder, A.Z.: Summary cache: a scalable wide-area
web cache sharing protocol. IEEE/ACM Trans. Netw. 8, 281–293 (2000)
10. Cheng, K., Xiang, L.: Efficient web profiling by time-decaying bloom filters. DBSJ
Letters 4, 137–140 (2005)
11. Cohen, E., Strauss, M.: Maintaining time-decaying stream aggregates. In: PODS
2003: Proceedings of the twenty-second ACM SIGMOD-SIGACT-SIGART sym-
posium on Principles of database systems, pp. 223–233. ACM Press, New York
(2003)
12. Morris, R.: Counting large numbers of events in small registers. Commun. ACM 21,
840–842 (1978)
13. Whang, K.Y., Vander-Zanden, B.T., Taylor, H.M.: A linear-time probabilistic
counting algorithm for database applications. ACM Trans. Database Syst. 15, 208–
229 (1990)
Soft Decision Making for Patients Suspected Influenza
1 Introduction
The reliance on information obtained from databases is very critical and important.
Almost in every part of our life, there are lots of instances where we have either direct
or indirect dealing with databases. One aspect that database plays an important role is
in the field of decision making. Inputs obtained from data are stored in terms of re-
cords and attributes in databases do contribute a lot in the process of decision making.
To this, one practical problem is faced: for a particular property, whether all the at-
tributes in the attributes set are always necessary to preserve this property [1]. Soft set
theory [2], proposed by Molodtsov in 1999, is a new general method for dealing with
uncertain data. We note that the soft set is designed to replace a Boolean-valued in-
formation system. The research described in this paper is a part of our short term ef-
fort in applying soft set theory in order to make a decision and further grouping of
patients under certain symptoms of influenza. Since soft set theory and soft decision
making techniques are not yet widely known, we start from a tutorial introduction.
The notions are first introduced by means of simple examples and later formalized.
The second part of this research elucidates a soft decision making technique in sig-
nificant detail. We present a problem of decision making through an artificial data of
patient suspected influenza. The data contains 50 patients with 7 symptoms. Using
soft set and maximal symptoms co-occurences in patients, we explore how soft
decision making technique can be used to reduce the number of dispensable
symptoms and further make a correct decision. This technique may potentially
D. Taniar et al. (Eds.): ICCSA 2010, Part III, LNCS 6018, pp. 405–418, 2010.
© Springer-Verlag Berlin Heidelberg 2010
406 T. Herawan and M.M. Deris
2 Information System
The syntax of information systems is very similar to relations in relational data bases.
Entities in relational databases are also represented by tuples of attribute values. An
{ }
information system is a quadruple S = (U , A, V , f ) , where U = u1 , u 2 ,L , u U is a
{ }
non-empty finite set of objects, A = a1 , a 2 ,L , a A is a non-empty finite set of attrib-
utes, V = Ua∈A Va , Va is the domain (value set) of attribute a, f : U × A → V is an
information function such that f (u, a ) ∈ Va , for every (u, a ) ∈ U × A , called informa-
tion (knowledge) function. An information system is also called a knowledge repre-
sentation systems or an attribute-valued system. In an information system
S = (U , A, V , f ) , if Va = {0,1}, for every a ∈ A , then S is called a Boolean-valued
information system. In many medical information systems, there is an outcome of
classification that is known. This a posteriori knowledge is expressed by one (or
more) distinguished attribute called decision attribute; the process is known as super-
vised learning. An information system of this kind is called a decision system. A de-
cision system is an information system of the form D = (U , A U {d }, V , f ) , where
d ∉ A is the decision attribute. The elements of A are called condition attributes. A
simple example of a decision table can be found in Table 1.
Definition 2. (See [2].) A pair (F , E ) is called a soft set over U, where F is a map-
ping given by
F : E → P(U ) .
In other words, a soft set over universe U is a parameterized family of subsets of uni-
verse U. For e ∈ E , F (e ) may be considered as the set of e-elements of the soft set
(F , E ) or as the set of e-approximate elements of the soft set. Clearly, a soft set is not
a (crisp) set.
Definition 3. (See [3].) The class of all value sets of a soft set (F , E ) is called value-
class of the soft set and is denoted by C ( F , E ) .
Clearly C ( F , E ) ⊆ P(U ) .
Example 4. Let we consider a soft set (F , E ) which describes the “conditions of pa-
tients suspected influenza” that a hospital is considering to make a decision. The six
influenza symptoms, i.e., fever, respiratory, nasal discharges, cough, headache and
sore throat are adopted from [4] and one symptom added is lethargic. Suppose that
there are six patients in the hospital under consideration,
U = {h1 , h2 , h3 , h4 , h5 , h6 } ,
and E is a set of decision parameters
E = {e1 , e2 , e3 , e4 , e5 , e6 , e7 } ,
where
e1 stands for the parameter “fever”,
e2 stands for the parameter “respiratory”,
e3 stands for the parameter “nasal discharges”,
e4 stands for the parameter “cough”,
e5 stands for the parameter “headache”
e6 stands for the parameter “sore throat”
e7 stands for the parameter “lethargic”.
Consider the mapping F : E → P(U ) given by “patients (⋅) ”, where (⋅) is to be filled
in by one of parameters e ∈ E . Suppose that
F (e1 ) = {p1 , p3 , p4 , p5 , p6 } ,
F (e2 ) = {p1 , p2 } ,
408 T. Herawan and M.M. Deris
F (e3 ) = {p1 , p2 , p4 } ,
F (e4 ) = {p1 } ,
F (e5 ) = {p3 , p4 } ,
F (e6 ) = {p2 , p4 } ,
F (e7 ) = {p1 , p3 , p5 , p6 } .
Therefore, F (e1 ) means “patients suffer fever”, whose functional value is the set
{p1 , p3 , p 4 , p5 , p6 } . Thus, we can view the soft set (F , E ) as a collection of approxi-
mations as below
⎧ fever = {p1 , p3 , p 4 , p5 , p6 }, ⎫
⎪ respiratory = {p1 , p 2 }, ⎪
⎪ ⎪
⎪nasal discharges = {p1 , p2 , p 4 },⎪
(F , E ) = ⎪⎨ cough = {p1 },
⎪
⎬
⎪ head ache = {p3 , p4 }, ⎪
⎪ ⎪
⎪ sore throat = {p 2 , p4 }, ⎪
⎪ ⎪
⎩ lethargic = {p1 , p3 , p5 , p6 } ⎭
U e1 e2 e3 e4 e5 e6 e7
p1 1 1 1 1 0 0 1
p2 0 1 1 0 0 1 0
p3 1 0 0 0 1 0 1
p4 1 0 1 0 1 1 0
p5 1 0 0 0 0 0 1
p6 1 0 0 0 0 0 1
Each approximation has two parts, a predicate e and an approximate value set p.
For example, for the approximation “ fever = {p1 , p3 , p4 , p5 , p6 } ”, we have the predi-
cate name of patients with fever and its value set is {p1 , p3 , p 4 , p5 , p6 } .
The relation between a soft set and a Boolean-valued information system is given in
the following proposition.
Proposition 3. If (F, E ) is a soft set over the universe U, then (F, E ) is a binary-
valued information system S = (U , A, V{0,1} , f ) .
where
⎧1, x ∈ F (ei )
f i : U → Vi and f i (x ) = ⎨ , for 1 ≤ i ≤ A .
⎩0, x ∉ F (ei )
U e1 e2 e3 e4 e5 e6 e7 f E ( pi )
p1 1 1 1 1 0 0 1 5
p2 0 1 1 0 0 1 0 3
p3 1 0 0 0 1 0 1 3
p4 1 0 1 0 1 1 0 4
p5 1 0 0 0 0 0 1 2
p6 1 0 0 0 0 0 1 2
The problem of parameterization reduction and decision making through this view,
then is to reduce the number of parameter (symptoms) which preserved the consis-
tence of optimal and sub-optimal decisions.
410 T. Herawan and M.M. Deris
Maji et al. presented a reduction of soft sets and its applications in a decision making
problem, which can be briefly explained as follows. The most optimal decision
derived by Maji will be only be deduced by identifying the rough set-based reduction
set first. This can by maintained using a partition on U based on the indiscernibility
relation on a set of attributes in rough set theory [8]. As for our example based on
Table 3, the partition induced by the set of all attributes E = {e1 , e 2 , e3 , e4 , e5 , e6 , e7 } ,
denoted by U / E is given by
{{p }, {p }, {p }, {p }, {p , p }} .
1 2 3 4 5 6
Therefore, if any attribute to be deleted, then the partition induced by any subset
Q ⊂ E can only be considered as a reduct if partition induced by Q / E is still
{{p }, {p }, {p }, {p }, {p , p }} . For example,
1 2 3 4 5 6
Q ⊂ E , where Q = {e1 , e5 , e6 , e7 } and Q / E
is still
{{p }, {p }, {p }, {p }, {p , p }} .
1 2 3 4 5 6
Therefore, for Maji’s techniques, the set of attributes, Q can also be considered as a
reduct of E.
U e1 e5 e6 e7 f E ( pi )
p1 1 0 0 1 2
p2 0 0 1 0 1
p3 1 1 0 1 3
p4 1 1 1 0 3
p5 1 0 0 1 2
p6 1 0 0 1 2
However, for Maji’s techniques, it can be seen from Table 4, that the optimal deci-
sion of patients is no longer p1 . Patient p1 has instead become sub-optimal patient
together with that of p5 and p6 . The maximum value of parameters after reduction,
i.e., f E ( p3 ) = f E ( p4 ) = 3 , is now the maximum choice. Thus, the new optimal decision
of patient in this case will be both p3 and p4 . It is very obvious, selection for opti-
mal patient will yield inconsistent result as shown from one example of reduct
Q ⊂ E , as in the above example. Even if we are to consider the selection of the next
optimal patient, in the case of Q, it will also provide inconsistent result. If for another
example, G ⊂ E , G = {e1 , e3 , e5 , e6 , e7 } and G / E is still producing the partition like
{{p1 }, {p2 }, {p3 }, {p 4 }, {p5 , p6 }} . Therefore G / E can also be considered as a reduct of E.
Soft Decision Making for Patients Suspected Influenza 411
U e1 e3 e4 e5 e7 f E ( pi )
p1 0 0 1 1 1 3
p2 1 0 0 0 1 2
p3 0 1 1 1 0 3
p4 1 1 0 1 1 4
p5 0 0 1 1 0 2
p6 0 0 1 1 0 2
As can be seen clearly again, now the optimal decision of patients has now
changed to p4 , for this new set of reduct G which is G ⊂ E . Again the issue of
changing the value of the optimal decision of patient due to the change of maximum
value of different set of reduct will is very critical and of our main concern here. This
will not be an ideal technique of Maji’s proposal, since it does not provide consistent
result in terms of different optimal decision of patients that is suggested. Furthermore,
Maji did not discuss the sub-optimal patient that can be derived from the reduced set.
And as we have explored, the sub-optimal patient if addressed by Maji, then again
different set of reduct will yield different sub-optimal patient, thus again not able to
provide consistency in selecting sub-optimal patient. As pointed out earlier in this
paper, that the main intention of reduction is to reduce complexity in decision making
but at the same time, still able to provide consistencies in decision making.
For the sake of comparison results from different proposed techniques, let us again
use the Table 3 as example for analysis purposes. The main purpose of Chen tech-
niques, was to maintain consistency in optimal value of decision thus improving the
∑
process of decision making from Maji. Chen has defined f E ( pi ) = j pij where hij
are the entries in the Boolean-table of (F , E ) and M E denoted for collection of
U e1 e2 e3 f E ( pi )
p1 1 1 1 3
p2 0 1 1 2
p3 1 0 0 1
p4 1 0 1 2
p5 1 0 0 1
p6 1 0 0 1
412 T. Herawan and M.M. Deris
objects in U which has the maximum value of f E . Chen has in fact defined a dispen-
sable set A ⊂ E if only if M E \ A = M E . Clearly, parameter reduction of Chen has been
able to provide consistency in optimal object’s decision.
Let assume J, where J ⊆ E and J = {e4 , e5 , e6 , e7 } . Hence, M E \ J = M E is true, and
therefore J is dispensable and we are left with S = {e1 , e2 , e3 } and S ⊆ E . In the Table
6 above, p1 is still the optimal decision of patient, thus parameter reduction thus still
up-hold consistency in selecting optimal patient as opposed to Maji’s attribute reduc-
tion. However, Chen’s parameter reduction is only able to provide in consistency in
selecting optimal decision but not in the sub-optimal decision. As can be seen from the
Table 6 above, that the sub-optimal patients are p2 and p4 . Meanwhile, from Table 3,
p4 is the only patient to be referred clearly as the sub-optimal patient. Therefore pa-
rameter reduction of Chen does provide consistency in selecting optimal patient but
falls short in providing consistency for the selection of sub-optimal patients.
{
CE = {p1 ,L , pi }f , {pi+1 ,L , p j }f ,L , {pk ,L , pn }f .
1 2 s
}
In the case that for A ⊂ E , if f A ( p1 ) = f A ( p2 ) = ... = f A ( p n ) implies C E = C E − A , then
A is called dispensable set. For this definition, Kong has termed E − A as normal
parameter reduction.
Based on on the Boolean table in Table 3, the decision partition induced will be
C E = {{p1 }5 , {p4 }4 , {p2 , p3 }3 , {p5 , p6 }2 } , and p1 is the optimal decision and p4 will be
U e1 e2 e3 e4 e5 f E ( pi )
p1 1 1 1 1 0 4
p2 0 1 1 0 0 2
p3 1 0 0 0 1 2
p4 1 0 1 0 1 3
p5 1 0 0 0 0 1
p6 1 0 0 0 0 1
Soft Decision Making for Patients Suspected Influenza 413
the first sub-optimal decision following by p2 , p3 and p5 , p6 as the second and third
sub-optimal patients, respectively. To do a normal parameter reduction from Table 3,
since f {e ,e } (h1 ) = f {e ,e } (h2 ) = ... = f {e ,e } (h6 ) = 1 , thus Z = {e6 , e7 } ⊂ E , is dispensable
6 7 6 7 6 7
since the decision partition has not changed, that is C E = C E \ Z . By deleting parameter
Z = {e1 , e4 } ⊂ E , we will have as what is been displayed in Table 7. And Kong has
successfully shown that the optimal and sub-optimal decisions, p1 and p4 , {p2 , p3 } ,
{p5 , p6 } respectively, thus maintaining consistency in patients decision after the
reduction.
Throughout this sub-section the pair (F, E ) refers to the soft set over the universe U
representing a Boolean-valued information system S = (U , A, V{0,1} , f ) .
Definition 4. Let (F, E ) be a soft set over the universe U and u ∈ U . A parameter
co-occurrence set of an object u can be defined as
coo(u ) = {e ∈ E : f (u , e ) = 1} .
Obviously, Coo(u ) = {e ∈ E : F (e ) = 1} .
Definition 5. Let (F, E ) be a soft set over the universe U and u ∈ U . Support of an
object u is defined by
supp(u ) = card({e ∈ E : f (u, e ) = 1}) .
Definition 6. Let (F, E ) be a soft set over the universe U and u ∈ U . An objects u is
said to be maximally supported by a set of all parameters E, denoted by Msupp(u ) if
Definition 7. Let (F, E ) be a soft set over the universe U and A ⊂ E . A is said to be
indispensable if U / A = U / E . Otherwise, A is said to be dispensable.
414 T. Herawan and M.M. Deris
Based on Definition 7, we can reduce the number of parameters without changing the
optimal and sub-optimal decisions.
Definition 8. For soft set (F, E ) over the universe U and A ⊆ E . A is reduction of E
if only if A is indispensable and supp A (u ) = A supp(v ) , for every u, v ∈U .
Definition 9. For soft set (F, E ) over the universe U and u ∈ U . An object u will be
the optimal decision if u is maximally supported by E.
Example 10. As for example, the following will be the co-occurrence objects derived
from Table 3.
coo( p1 ) = {e1 , e2 , e3 , e4 , e7 } coo( p2 ) = {e2 , e3 , e5 } , coo( p3 ) = {e1 , e5 , e7 } ,
,
coo( p4 ) = {e1 , e3 , e5 , e7 } , coo( p5 ) = {e1 , e7 } and coo( p6 ) = {e1 , e7 } .
Thus, support of each object is given as follow
supp( p1 ) = 5 > supp( p4 ) = 4 > supp( p2 ) = supp( p3 ) = 3 > supp( p5 ) = supp( p6 ) = 2 .
Based from the Definition 7, the partition contain 4 clusters, i.e.,
{{p }, {p }, {p , p }, {p , p }} ,
1 4 2 3 5 6
Furthermore, our proposed technique confirming that the reduction also provide the
right objects for decision making.
Patients e1 e2 e3 e4 e5 e6 e7
1 1 0 1 0 0 0 1
2 1 1 0 1 0 1 1
3 1 1 0 1 0 1 1
4 1 0 1 1 1 0 0
5 1 0 0 0 0 0 1
6 1 0 0 0 0 0 1
7 1 0 1 1 1 0 0
8 1 0 0 0 0 0 1
9 1 0 0 0 0 0 1
10 1 0 0 0 0 0 1
11 0 0 0 1 1 1 0
12 1 0 1 1 1 0 0
13 1 0 1 0 0 0 1
14 1 0 1 0 0 0 1
15 1 1 0 1 0 1 1
16 1 1 0 1 0 1 1
17 1 0 0 0 0 0 1
18 1 1 0 1 0 1 1
19 1 0 1 1 1 0 0
20 1 0 0 0 0 0 1
21 1 0 0 0 0 0 1
22 1 0 0 0 0 0 1
23 0 0 0 1 1 1 0
24 0 0 0 1 1 1 0
25 1 0 0 0 0 0 1
416 T. Herawan and M.M. Deris
Table 8. (continued)
26 0 0 0 1 1 1 0
27 1 0 0 0 0 0 1
28 0 0 0 1 1 1 0
29 1 0 0 0 0 0 1
30 0 0 0 1 1 1 0
31 0 0 1 1 0 0 0
32 1 0 1 0 1 1 0
33 0 0 1 1 0 1 0
34 1 0 0 0 1 0 1
35 0 0 1 1 0 0 0
36 0 0 1 1 0 1 0
37 0 0 1 1 0 0 0
38 1 0 1 0 1 1 0
39 0 0 1 1 0 0 0
40 0 0 1 1 0 0 0
41 0 0 1 1 0 0 0
42 1 0 1 0 1 1 0
43 1 0 1 0 1 1 0
44 1 1 1 1 0 0 1
45 0 0 1 1 0 1 0
46 1 1 1 1 0 0 1
47 1 1 1 1 0 0 1
48 0 0 1 1 0 1 0
49 1 0 0 0 1 0 1
50 0 0 1 1 0 1 0
6.2 Result
After calculating co-occurrences of symptoms in each patient, figure 1 lists the de-
tailed results of dispensable symptoms and reduct obtained. It is clear that symptoms
severe malaise (feeling unwell) (e5 ) and runny nose (e7 ) are indicated as dispensable
symptoms. Therefore, reduct of the symptoms is fever (e1 ) , cough (usually dry) (e2 ) ,
headache (e3 ) , muscle and joint pain (e4 ) , and sore throat (e6 ) .
Dispensable symptoms e5 e7
Reduct e1 e2 e3 e4 e6
After reduction, we can make easily make decisions. The decisions are based on
the support symptoms of each patient. Patients having the same support are clustered
in the same class. There are 9 clusters decision we can made and they are given in
Soft Decision Making for Patients Suspected Influenza 417
Cluster 1 (fever) 5 6 8 9 10 17 20 21 22 25 27 29 34 49
Figure 2. Patient with higher supported symptoms are indicated having flu rather than
lower supported symptoms.
7 Conclusion
In this paper, we have presented an application of soft decision making technique for
decision making of patients suspected influenza. We have proposed an alternative ap-
proach for parameterization reduction and decision making under soft set theory. We
have used the co-occurrence of parameters concept in a Boolean-valued information
system for defining support of each object by parameters. Based on maximal supported
objects, we have defined the notion of indispensable subset and reduct. We have pre-
sented an artificial dataset of 50 patients suspected influenza to show that the proposed
technique can be used to make a correct decision. We have shown that the results ob-
tained are equivalent with that normal parameterization and decision making problem
under soft set theory. For future research, we will investigate the possibility of parame-
terization reduction and decision making problem under a certain criteria of parameter,
e.g., under spatial ordered parameter. With this approach, it is possible that the reduct
and decision obtained are different with the proposed technique in this paper.
Acknowledgement
This work was supported by the FRGS under the Grant No. Vote 0402, Ministry of
Higher Education, Malaysia.
References
1. Zhao, Y., Luo, F., Wong, S.K.M., Yao, Y.Y.: A General Definition of an Attribute Reduct.
In: Yao, J., Lingras, P., Wu, W.-Z., Szczuka, M.S., Cercone, N.J., Ślȩzak, D. (eds.) RSKT
2007. LNCS (LNAI), vol. 4481, pp. 101–108. Springer, Heidelberg (2007)
418 T. Herawan and M.M. Deris
2. Molodtsov, D.: Soft Set Theory-First Results. Computers and Mathematics with Applica-
tions 37, 19–31 (1999)
3. Maji, P.K., Biswas, R., Roy, A.R.: Soft Set Theory. Computers and Mathematics with Ap-
plications 45, 555–562 (2003)
4. Abbas, K., Mikler, A.R., Gatti, R.: Temporal Analysis of Infectious Diseases: Influenza.
In: The Proceeding of ACM Symposium on Applied Computing (SAC), pp. 267–271
(2005)
5. Maji, P.K., Roy, A.R., Biswas, R.: An Application of Soft Sets in a Decision Making Prob-
lem. Compututer and Mathematics with Application 44, 1077–1083 (2002)
6. Chen, D., Tsang, E.C.C., Yeung, D.S., Wang, X.: The Parameterization Reduction of Soft
Sets and its Applications. Computers and Mathematics with Applications 49, 757–763
(2005)
7. Kong, Z., Gao, L., Wang, L., Li, S.: The Normal Parameter Reduction of Soft Sets and Its
Algorithm. Computers and Mathematics with Applications 56, 3029–3037 (2008)
8. Pawlak, Z., Skowron, A.: Rudiments of Rough Sets. Information Sciences 177(1), 3–27
(2007)
9. http://www.who.int/mediacentre/factsheets/fs211/en/index.html
Personal Identification by EEG Using ICA
and Neural Network
1 Introduction
Biometrics such as fingerprints, retinal or iris scanning, face recognition are ac-
tively used for identifications [1]. Cognitive biometrics using brain signals have
become interesting as identification tools. To understand how the brain functions,
many techniques such as electroencephalography (EEG), magnetoencephalogra-
phy (MEG), function magnetic resonance imaging (fMRI), and positron emission
tomography (PET) have been utilized. Each technique has its own strengths and
weaknesses. In this paper EEG has been used to analyze patterns because EEG has
a desirable property for excellent time resolution and low cost of instrumentation.
The objectives of this work strive to acquire locations on the scalp (channels)
that are the most promising locations for personal identification. Furthermore,
the minimum numbers of channel necessary for the identifications would also
be explored. Identifications by EEG using Independence Component Analysis
technique has been used to identify individual signals from different areas of the
brain. Neural network technique was then performed on pattern recognition for
identification. Our study is organized as follows. Section 1 is the introduction,
using EEG for identifications. Section 2 presents the background work; tech-
niques employed in the experiment such as ICA and neural networks and the
D. Taniar et al. (Eds.): ICCSA 2010, Part III, LNCS 6018, pp. 419–430, 2010.
c Springer-Verlag Berlin Heidelberg 2010
420 P. Tangkraingkij et al.
experimental process. Then, the results are displayed and discussed. Finally, the
conclusion is included in Section 3.
Find the best ICA algorithms applied with 5,10,20 subjects; Reduce the number
of channels from 16 channels; Test the neural network with 20 subjects.
Collect EEG signal. EEG signals were collected from 20 normal patients (eight
men and twelve women) from the Chulalongkorn Comprehensive Epilepsy Pro-
gram (CCEP). The age range of subjects is between 12 and 40 years. The EEG
signal was recorded while subjects were resting with their eyes open. The elec-
trodes using 16 gold cups were placed on the scalp. Electrode placement are used
to record EEG signals from the locations FP 1 , F7 , T3 , T5 , FP 2 , F8 , T4 , T6 , F3 , C3
P3 , O1 , F4 , C4 , P4 , O2 according to 10-20 system shown in Figure 2. In this pa-
per, channels are labeled as ch1, ch2, ..., ch16 representing EEG from location
FP 1 , F7 , ..., O2 as mentioned earlier.
Recording sessions used monopolar montage with reference at the mastoid
area. The EEG amplifier was Grass model 8 plus. The sampling rate was 200 Hz.
The data was digitized by BMSI board using Stellate Harmony EEG software.
The data was exported as EDF (European Data Format). Electromyography
(EMG) and electrocardiogram (ECG) signals were initially removed from the
EEG signals since they were irrelevant. Finally, there were 3,000 sample data
sets of 16 channels collected from 20 subjects. A sample of the raw EEG signals
is shown in Figure 3(a).
Feature extractions using ICA. The purpose of this step is to identify the
individual signal coming from each channel using 22 different ICA algorithms.
The signal data obtained from the previous step were processed using ICALAB.
The ICALAB for signal processing is the package for MATLAB that implements
422 P. Tangkraingkij et al.
Fig. 2. The locations of electrode placements on the scalp using 10-20 system [3]
Fig. 3. The samples of raw and feature extraction using ICA. (a) Raw EEG signals
from four channels (ch2, ch10, ch11, ch12) of a subject are displayed. The x-axis of
each channel denotes number of samples, whereas y-axis represents the amplitude of
the EEG signal. (b) The EEG signals after being processed by ICA algorithm (ERICA).
Personal Identification by EEG Using ICA and Neural Network 423
a number of efficient algorithms for ICA [4]. The list of these 22 different ICA
algorithms used in the experiment is shown in Table 1. The sample of raw EEG
signals in Fig. 3(a) after being processed by ICA algorithm for feature extraction
is shown in Fig. 3(b).
Find the best ICA algorithms using a neural network. The objective of
this part is to explore which ICA algorithm performs the best among 22 different
ICA algorithms. The experiments were conducted by using pattern recognition
feature of the neural network on MATLAB 2008a. All EEG data processed from
feature extractions were applied to 5, 10, and 20 subjects, respectively. Then,
the results were compared as shown in Table 1. The first column shows the
name of 22 ICA algorithms. The second column shows accuracy, i.e. the per-
centage of correctly classified cases from training, validation and testing data
sets for five subjects. Note that there are several algorithms achieving 100 per-
cent. To see whether this accuracy is still retained for the other sizes of ex-
perimented subjects, the numbers of study subjects were increased to 10 and
20 as shown in the third and fourth columns, respectively. The experimental
results of 5, 10 and 20 subjects were studied. ERICA, EWASOBI, JADEop,
SIMBEC, SOBI, SOBI-BPF, SOBIRO, and WASOBI algorithm gave high per-
centage of accuracy. Therefore, one of these algorithms was chosen to employ
in the next step to reduce the number of channels. Here, ERICA is the selected
algorithm.
Table 1. The accuracy percentage of each ICA algorithm when applied with 5, 10, 20
subjects. The details of all algorithms can be seen in ICALAB [2], [4].
by using the first six signal values from each channel. The next two signal values
are used for forming the validation patterns and another next two signal val-
ues next to validation patterns are formed as testing patterns. The network was
trained with five initial weight sets to test whether the training, validation, and
testing sets are properly formed. If they are properly formed, then the network
will always be converged. Otherwise, it may be diverged by some initial weight
sets.
Table 2 displays the results of groups of two channels when applied with 5,
10 and 20 subjects. When 5 subjects are used in experiments, nine combinations
have 100% accuracy percentage. When 10 and 20 subjects are used in the ex-
periment, the best accuracy percentage in the group of Ch2 and Ch7 is dropped
to 91.96% and 61.67%, respectively.
The results in Table 2, suggested that only two channels may not be enough
to use for identification since the accuracy percentage with 20 subjects is about
60%. Therefore, the number of channel in a group of three is investigated to
increase the accuracy percentage. Again, the possible combinations of 3 out of
16 channels are considered (16 C3 ), which is equal to 560 combinations. From the
three-channel experiment, the training set consists of three channels of 5, 10,
Personal Identification by EEG Using ICA and Neural Network 425
Table 2. The accuracy percentage of 2 channels when applied with 5 and 10 subjects
using ERICA algorithm
and 20 subjects with each subject using 500 samples. The ratio of the training
set, validation set and testing set is 60:20:20 as in the case of two channels. The
results of three channels are shown in Table 3.
Table 3. The accuracy percentage of 3 channels when applied with 5,10,20 subjects
using ERICA algorithm
Table 3 shows only partial of the three channels combinations. The accuracy
percentages with 5 subjects reach 100%. With 10 subjects, the accuracy per-
centages reach 100% in the groups of Ch4, Ch5, Ch10 and Ch4, Ch13, Ch15.
When numbers of subjects were increased to 20 subjects, the result shows that
the highest percentage is 90.52% in the group of Ch4, Ch5, Ch10.
The accuracy percentage of the experiment using three channels for 20 sub-
jects is still not high enough (the ideal accuracy should get close to 100%). So
the number of channels was increased to a group of four channels to increase the
accuracy percentage of the 20 subjects.
In case of four channels, total combinations of 4 out of 16 channels are con-
sidered (16 C4 ), which is equal to 1,820 combinations. The training set consists of
four channels of 5, 10, and 20 subjects, with each subject employing 500 samples.
The ratio of the training set, validation set and testing set of four channels is the
same as in the case of two and three channels. The results four channels groups
are shown in Table 4.
426 P. Tangkraingkij et al.
Table 4. The accuracy percentage of 4 channels when applied with 5,10,20 subjects
using ERICA algorithm
Table 4 shows only a part of the four channels combinations. The accuracy
percentages with 5 and 10 subjects reach 100%. When the numbers of subjects
were increased to 20 subjects, the result shows that the highest percentage is
98.71% in the group Ch2, Ch10, Ch11, Ch12. The results of the experiments for
5, 10, and 20 subjects are acceptable. Therefore, four channels are the lowest
possible channels for this experiment.
Algorithms The best accuracy The best accuracy The best accuracy
for 5 subjects for 10 subjects for 20 subjects
(combination) (combination) (combination)
ERICA 1,502 combinations 185 combinations ch2 ch10 ch11 ch12
reach 100% reach 100% 98.71%
EWASOBI 862 combinations 34 combinations ch4 ch 5 ch 8 ch13
reach 100% reach 100% 91.17%
JADEop 40 combinations 4 combinations ch9 ch11 ch14 ch15
reach 100% reach 100% 93.48%
SIMBEC 110 combinations 14 combinations ch8 ch12 ch13 ch16
reach 100% reach 100% 93.48%
SOBI 110 combinations 4 combinations ch6 ch 7 ch12 ch13
reach 100% reach 100% 96.22%
SOBI-BPF 1,511 combinations 218 combinations ch3 ch 8 ch 9 ch12
reach 100% reach 100% 98.57%
SOBIRO 170 combinations 22 combinations ch5 ch10 ch13 ch15
reach 100% reach 100% 97.87%
WASOBI 1,814 combinations 65 combinations ch3 ch 6 ch 9 ch10
reach 100% reach 100% 94.26%
Test the neural network with 20 subjects. The purpose of this step is
to test the accuracy of recognition. The neural network that has been trained
earlier using ERICA algorithm on ch2, ch10, ch11, ch12 will be used to test
the accuracy percentage. From Table 5, ERICA algorithm performs the best. In
order to further investigate the performance of this recognition, 3,000 samples
of the EEG signals are classified into 6 groups G1 - G6 (shown in Fig. 4). Each
group contains 500 samples. The first group: G1 consists of sample data from 1
- 500, the second group: G2 consists of sample data from 501-1000, and so on
until the sixth group: G6 is the sample data from 2501-3000. Note that G1 has
been used for training and testing the recognition in the previous step. However,
G2 to G6 are samples that have not been used to train and test the recognition
performance before. Thus, these samples: G2 to G6 are then used to test the
recognition performance on 20 subjects. The results are shown in Table 6.
Fig. 4. The EEG signals are divided into six groups for testing
428 P. Tangkraingkij et al.
Table 6. The accuracy percentage of 20 subjects when test with the best combination
of ERICA algorithm 4 channels 500 samples
Subject G1 G2 G3 G4 G5 G6
(%) (%) (%) (%) (%) (%)
1 100.00 100.00 99.20 98.40 93.40 98.80
2 100.00 100.00 100.00 98.20 100.00 100.00
3 100.00 100.00 100.00 100.00 100.00 100.00
4 100.00 100.00 100.00 100.00 100.00 100.00
5 100.00 99.60 100.00 99.80 100.00 100.00
6 97.00 99.60 99.80 99.60 97.60 99.40
7 99.60 97.80 95.00 99.00 94.60 94.00
8 100.00 99.40 92.40 100.00 100.00 99.60
9 100.00 100.00 100.00 99.80 100.00 100.00
10 97.40 96.40 93.40 97.80 96.20 90.00
11 100.00 99.00 95.60 99.00 98.80 99.60
12 100.00 100.00 100.00 99.20 99.80 95.00
13 99.80 98.60 99.20 100.00 100.00 100.00
14 95.40 81.00 81.80 91.60 82.40 81.20
15 99.40 92.80 95.40 97.00 92.60 94.20
16 96.80 96.20 97.60 94.80 89.00 93.00
17 94.60 92.60 94.60 94.40 98.00 96.40
18 95.80 89.40 93.20 94.80 86.60 90.60
19 98.40 98.40 99.40 98.40 99.60 90.60
20 100.00 99.80 100.00 100.00 99.80 100.00
Average
G(I) 98.71 97.03 96.83 98.09 96.42 96.59
Numbers in Table 6 are the accuracy percentage for each subject in a par-
ticular group. In the first row, for example, the accuracy percentage of subject
1 was evaluated with all groups of data (G1 to G6) . The accuracy percent-
age of subject 1 using G1 data is 100%. The accuracy percentages using G2 -
G6 data are 100%, 99.20%, 98.40%, 93.40%, 98.80%, respectively. The last row
of the Table 6 shows the average accuracy percentage in each test data group.
It can be seen that the average accuracy of these six groups is in the range
of 96% - 99%. The results show that identifications on 20 subjects are very
encouraging.
In this paper, a practical technique for identifying from standard 16 EEG lo-
cations of EEG signals is proposed. The signals are cleaned first by apply-
ing Independent Component Analysis. Then, a supervised neural network is
used to test the accuracy of the identification process. We found that minimum
Personal Identification by EEG Using ICA and Neural Network 429
References
1. Jain, A.K., Ross, A., Prabhakar, S.: An Introduction to Biometric Recognition.
IEEE Trans. Circuits Syst. Video Technol. 14(1), 4–20 (2004)
2. Cichocki, A., Amari, S.: Adaptive blind signal and image processing: learning al-
gorithms and applications. Wiley, New York (2003)
3. Niedermeyer, E., da Silva, F.L.: Electroencephalography: Basic Principles. In: Clin-
ical Applications, and Related Fields, pp. 139–141. Lippincott Williams, Wilkins
(2004)
4. Cichocki, A., Amari, S., Siwek, K., Tanaka, T., et al.: ICALAB toolboxes,
http://www.bsp.brain.riken.jp/ICALAB
5. Poulos, M., Rangoussi, M., Alexandries, N.: Neural network based person identifi-
cation using EEG features. In: Proceeding IEEE of the international conference on
Acoustic, Speech and Signal Processing, ICASSP 1999, Arizona, USA, pp. 1117–
1120 (1999)
6. Poulos, M., Rangoussi, M., Kafetzopoulos, E.: Person identification via the EEG
using computational geometry algorithms. In: Proc. Intl. Conf. EUSIPCO 1998,
Rhodes, Greece (September 1998)
7. Marcel, S., Millan, J.: Person authentication using brain-waves (EEG) and maxi-
mum a posteriori model adaptation, Tech. Rep. 81, p. 11, IDIAP Research Report,
Valais, Switzerland (2005)
8. Paranjape, R.B., Mahovsky, J., Benedicenti, L., Koles, Z.: The electroencephalo-
gram as a biometric. In: Proceedings of Canadian Conference on Electrical and
Computer Engineering, Toronto, Canada, May 2001, vol. 2, pp. 1363–1366 (2001)
9. Mohammadi, G., Shoushtari, P., Ardekani, B., Shamsollahi, M.: Person identifica-
tion by using AR model for EEG signals. In: Proceedings of the 9th International
Conference on Bioengineering Technology (ICBT 2006), Czech Republic, vol. 5
(2006)
10. Daly, D.D., Markand, O.N.: Focal brain lesions. In: Daly, D.D., Pedley, T.A. (eds.)
Current Practice of Clinical Electroencephalography, 2nd edn. Raven Press, New
York (1990)
11. Poulos, M., Rangoussi, M., Alexandris, N., Evangelou, A.: On the use of EEG
features towards person identification via neural networks. Medical Informatics
and the Internet in Medicine 26(1), 35–48 (2001)
430 P. Tangkraingkij et al.
12. Poulos, M., Rangoussi, M., Alexandris, N., Evangelou, A.: Person identification
from the EEG using nonlinear signal classification. Methods of Information in
Medicine 41(1), 64–75 (2002)
13. Palaniappan, R., Ravi, K.V.R.: A New Method to Identify Individuals Using Sig-
nals from the Brain. In: Proc. Fourth Int’l Conf. Information Comm. and Signal
Processing, pp. 15–18 (2003)
14. van Drongelen, W.: Signal Processing for Neuroscientists An introduction to the
analysis of physiological signals. Academic Press, London (2007)
15. Saeid Sanei, J.A.: Chambers, EEG Signal Processing. Wiley, Chichester (2007)
A Formal Concept Analysis-Based Domain-Specific
Thesaurus and Its Application in Document
Representation
1 Introduction
With the growth of internet technology, ever more information is becoming available
online. However, this causes the problem of information overload, in that users have
to make greater efforts to find the information they need from web sites or document
bases. Many techniques, such as information retrieval, personalized portal or docu-
ment management are developed to help users solve this problem, by organizing the
search results or document bases in hierarchical structures, or providing users addi-
tional vocabularies to reformulate their query text.
D. Taniar et al. (Eds.): ICCSA 2010, Part III, LNCS 6018, pp. 431–442, 2010.
© Springer-Verlag Berlin Heidelberg 2010
432 J.-C. Jehng, S. Chou, and C.-Y. Cheng
Most of the techniques in the processing of document retrieval and clustering rep-
resent documents by vectors. Usually, the dimensions of the vectors are the terms
extracted from documents. This approach has been widely applied to information
retrieval and related techniques, such as document retrieval, document clustering, and
document classification.
The main problem with this bag of terms approach is that it ignores the conceptual
similarity of terms [1][2][3]. Suppose there are two documents s1 and s2; term "query
expansion" appears in the topic of s1 and "document clustering" appears in s2. In the
traditional approach, the two documents will be identified irrelevant, because there is
no hypernym like "information retrieval" to make them relevant.
The application of conceptual relationships of terms to document retrieval and
clustering in some cases can improve the precision and recall rate [1][2]. Usually, the
authors of documents use different words to describe the same thing. However, tradi-
tional approaches do not take these relevant words into synonym sets. A thesaurus
like WordNet [4] can provide the information of conceptual relationships of terms
such as synonyms, hypernyms or hyponyms. Previous studies, such as Hotho et al. [2]
and Recupero [1], used WordNet as background knowledge to improve the results of
document clustering.
A thesaurus like WordNet is a general thesaurus, and does not aim at the use of
specific domains. Therefore, some domain-specific words or phrases cannot be found
in WordNet. These words or phrases are always thought of as important features in
domain-specific documents. For example, phrases such as "information retrieval",
"query expansion" and "document clustering" are important features in information
retrieval-related documents; however, they cannot be found in WordNet. To deal with
the problem of general thesauri, some researches construct domain-specific thesaurus
by ontology [3][5]. However, the building of domain-specific ontology always re-
quires the participation of humans and is manual or semi-automatic.
To tackle the problems mentioned above, an automatic domain-specific thesaurus
building methodology is needed. Formal concept analysis (FCA) is mainly used for
data analysis and in particular for extracting inherent relationships between objects in
a given context [6]. This feature helps us to extract the relationships of terms and
form the terms into a thesaurus. Previous researches applied FCA to identify the re-
lated terms in a given document set and used these terms to generate expanding terms
for query expansion [7].
In this paper, we propose an automatic domain-specific thesaurus building ap-
proach by FCA. Each concept of the domain-specific thesaurus consists of related
terms, and the relations between concepts are inherent. This thesaurus provides the
relations of terms for a given specific domain. We also use it as background knowl-
edge to represent documents by concept dimension vectors.
The rest of this paper is organized as follows: In section 2, we introduce the FCA
and its applications in information retrieval. Major definitions of FCA are also de-
scribed. In section 3, the proposed method of constructing the domain-specific
thesaurus by FCA is explained. In this section, we first introduce the procedures of
document preprocessing, then build our domain-specific thesaurus by FCA. Finally,
we explain the method of representing documents by concept dimension vectors. In
section 4, evaluation of our method is conducted. We make a comparison with tradi-
tional approaches to show the improved results of the use of the domain-specific
thesaurus. Finally, we come to the conclusions of this paper.
A Formal Concept Analysis-Based Domain-Specific Thesaurus 433
Definition 1
A formal context is defined as a triple (G, M, I), where G and M are two sets and
I ⊆ G × M is a binary relation between G and M. The elements of G are called objects
and those of M are called attributes. I is the relation defined between G and M. To
represent a relation I between object g and attribute m, we write gIm or ( g , m) ∈ I .
In Table 1, we provide an example of a formal context which contains four objects
from Document1 to Document4, and attributes are the words appearing in the given
documents. The elements of Table 1 represent the binary relation between G and M
that indicates whether terms appeared in the documents.
Definition 2
For A ⊆ G , we define
A' = {m ∈ M | ( g , m ) ∈ I for all g ∈ A}
denotes the set attributes common to all the objects in A.
Dually, for B ⊆ G ,
B' = {g ∈ G | ( g , m) ∈ I for all m ∈ B}
Definition 3
A formal concept of context (G, M, I) is defined as a pair (A, B)
where A ⊆ G , B ⊆ G , A'=B and B'=A.
For example, {Document1, Document3}={"Query expansion"} is called a formal
concept in our example of formal context. A of the formal concept (A, B) is called
extent and B is called intent. Extent contains all objects belonging to the formal con-
cept (A, B), and intent is the attributes shared by the objects.
Definition 4
Two formal concepts of a given formal context can establish the super-concept and
sub-concept relation. Let (A1, B1) and (A2, B2) be formal concepts of a formal context;
(A1, B1) is a sub-concept of (A2, B2) if and only if A1 ⊆ A2 (or if and only if B2 ⊆ B1);
then (A2, B2) is called a super-concept of(A1, B1). We write (A1, B1) p (A2, B2), where
p denotes a super-concept and sub-concept relation of the formal concepts. Consider
the set of all formal concepts and their super-concept and sub-concept relations of the
formal context; then CL = (G , M , I ) is a complete lattice called Concept Lattice.
A Concept Lattice can be drawn as a diagram with nodes representing formal con-
cepts and edges representing super-concept and sub-concept relations between formal
concepts. Fig. 1 is the Concept Lattice constructed from the formal context of Table 1.
Each node in Fig. 1 represents a formal concept and the edges are their inherent rela-
tions. From Fig. 1, we find a sub-concept inherits all attributes from its super-
concepts and is more specific than its super-concepts. Many methodologies for
generating Concept Lattices can be found in [15]. In this paper, we use the tool of
Concept Explorer1 for producing Concept Lattices.
where tf(di,tn) is the frequency of term tn in document di; df(tn) is the document fre-
quency that indicates the number of documents containing term tn. This weighting
schema is used to measure the importance of terms in a given document. Some further
applications, such as document retrieval, document clustering, or document classifica-
tion, apply Cosine or Jaccard measure to compute the distance between two vectors
which are represented documents or queries.
This very typical method of document representation, called the vector space
model, was first introduced by Salton [16] and has been successfully and widely ap-
plied to information retrieval. However, one of the main drawbacks of this approach
is its ignoring of conceptual relationships of terms. To tackle this problem, we pro-
posed an automatic domain-specific thesaurus building approach based on FCA. This
domain-specific thesaurus is mainly used to represent the relations between terms via
the relations between formal concepts. We also applied this thesaurus as background
knowledge to represent documents by concept dimension vectors.
In our method, FCA is applied to construct the domain-specific thesaurus. The Con-
cept Lattice generated by FCA for a given document set consists of a set of formal
concepts in which related terms are formed. By the definitions introduced in Section 2,
all relations between formal concepts in a Concept Lattice are inherent, and these rela-
tions are called super-concept and sub-concept relations. Therefore, we assume that the
terms in the same formal concept can be regarded as synonyms; and the terms in super-
concept are hypernyms for those terms in their sub-concepts; on the other hand, terms
in sub-concept can be thought as hyponyms for those terms in their super-concepts. In
the following sections, we introduce the procedure for our method in detail.
1
Concept Explorer can be download at http://sourceforge.net/projects/conexp/
436 J.-C. Jehng, S. Chou, and C.-Y. Cheng
Each token after the slash is the POS tag of words. For example, IN is preposition; DT
is determiner; and NN is noun. A complete list of POS tags can be found in [17].
Second, we keep all nouns, and in order to extract the domain-specific phrases
from documents, we referr to a list of POS patterns constructed by Ou, et al. [18] for
recognizing phrases (See Table 2). Then, stop-words are removed from remained
nouns.
Part-of-speech tag
1 2 3 Example term
NN QE, IR, FCA
JJ NN unexpanded query
NN NN query expansion, document clustering
JJ NN NN interactive query expansion
Third, we stemmed all remained nouns and phrases by Porter stemming algorithm
[19].
2
http://nlp.stanford.edu/software/tagger.shtml
A Formal Concept Analysis-Based Domain-Specific Thesaurus 437
Definition 6
A domain-specific thesaurus is defined as a tuple DST=(C, p )consisting of a set of
concept C and inherent relation p with C. These inherent relations are also called
sub-concept and super-concept relations. For each concept ci ∈ C, it contains a set of
term Tci.
Definition 7
For c1, c2 ∈ C, if c1 p c2, we called c1 is a sub-concept of c2; on the other hand, c2 is
super-concept of c1.If c1 is sub-concept of c2, then Tc2 is a subset of Tc1 (Tc2 ⊆ Tc1).
Definition 8
A concept ci is the root concept in DST, if and only if there is no super-concept for ci.
We write a root concept as cRoot. A root concept ci is the only concept that its Tci may
be empty in DST.
In order to transform the Concept Lattice CL generated in Section 3.2 into a do-
main-specific thesaurus DST defined above, the following steps are executed. First,
all information about objects is removed from a formal concept, while the object
information is no longer needed in the following process. Second, we delete the most
bottom formal concept whose object set is empty and term set contains all terms of T.
Third, we eliminate the formal concept whose set of terms Tci equals the union of sets
of terms of its all direct super-concepts, and the edges connected to this concept.
This pruned Concept Lattice with no circle is used to represent the domain-specific
thesaurus DST defined above. We keep all relations between formal concepts and use
the pruned formal concepts to represent concept ci defined in DST.
In order to deal with the problem with traditional bag of terms approaches which do
not take account of the conceptual similarity of terms in their applications, we apply
the domain-specific thesaurus defined in Section 3.3 as background knowledge to
represent a given document by concept dimension vector. A new concept dimension
vector can be represented as
→
d i = (cw(d i , c1 ), cw(d i , c2 ),..., cw(d i , cn )) ,
where cw(di,cn) is the weight of concept ci ∈ C in the document di. For each docu-
ment, all of its concept weights are determined by the following steps.
438 J.-C. Jehng, S. Chou, and C.-Y. Cheng
First, we compute all concept weights for all concepts ci ∈ C. We define a concept
weighting schema for each concept as follows.
Definition 9
For all ci ∈ C,
cw(d i , cn ) = ∑ tf (d i , t )
t∈Tci '
where Tci' is term set in ci excluded the terms appeared in the direct super-concepts of
concept ci.
Second we update concept weights for all concepts, except the concepts at most
bottom level, from its sub-concepts. In our definitions of DTS, a concept ci may in-
herit from multiple super-concepts. This is significantly different to hierarchical struc-
ture. In hierarchical structure, one concept only has one super-concept. Therefore, we
developed a bottom-up approach which updates the weight of concepts by moving the
weight of a given concept up to all its super-concepts.
A recursive algorithm called UpdateConceptWeights (Fig. 2) is used to update the
concept weights for all super-concepts for a given concept. A given concept c and its
weight SubWeight will be the inputs of UpdateConceptWeights. If concepts c is not
root concept, this algorithm will get the summation value NumberOfSuperTerms of
the number of terms of c’s direct super-concepts (lines 5-7). For each direct super-
concept csuper of c, their weight is determined by the maximum function (line 9). Two
values will be the inputs of this maximum function, one is the original concept weight
of csuper and another is distributed value of SubWeight. The distributed value of Sub-
Weight for csuper is defined as
NumberOfTerms (c Super )
×ω ,
NumberOfSuperConceptTerms
Algorithm: UploadConceptWeights
Input: Concept c, SubConceptWeights SubWeight
Output: Concept weights for all concepts in DST
1. if c=cRoot
2. return
3. else
4. NumberOfSuperConceptTerms=0
5. foreach csuper ∈ SuperConcept(c)
6. NumberOfSuperConceptTerms+=NumberOfTerms(csuper)
7. end foreach
8. foreach csuper ∈ SuperConcept(c)
9. cw(di,csuper)=max(cw(di,csuper),
NumberOfTerms (cSuper )
× ω × SubWeights )
NumberOfSuperConceptTerms
10. UploadConceptWeights (csuper,cw(di,csuper))
11. end foreach
12. end if
4 Evaluation
In this section, we demonstrate our method introduced in Section 3 through a
simple example given as follows: We first generate a domain-specific thesaurus
from an example training document set. Furthermore, this thesaurus is used as
background knowledge for representing the given document set by concept dimen-
sion vectors. Finally, we analyze and compare the results of the concept dimension
vectors with traditional term dimension vectors in the measuring of document
similarities.
The test data set we use in this evaluation are the sentences selected from [20],
which is a research paper mainly focused on the topic of query expansion. We select
sentences, which describe the techniques about local and global query expansion, as
training and test document sets.
A training document set which consists of four documents from TrainDoc1 to Tra-
inDoc4 is given as follows:
Highlighted terms, including single nouns and phrases, are selected by the proce-
dure of document preprocessing introduced in Section 3.1. All selected terms are
stemmed by Porter’s algorithm. We use these terms as the attributes and training
documents as objects to form the formal context. Then, according to the method de-
scribed in Section 3.2 and Section 3.3, a domain-specific thesaurus for this context
can be generated as Fig. 3.
440 J.-C. Jehng, S. Chou, and C.-Y. Cheng
Three test documents are given as follows: The important terms in documents are
selected and highlighted.
TestDoc1: One of the earliest global techniques is term clustering.
TestDoc2: Information from the top-ranked documents is used to
reestimate the probabilities of query terms in the relevant set
for a query.
TestDoc3: Local techniques expand a query based on the information in
the set of top-ranked documents retrieved for the query.
According to the procedures described in Section 3.4, each concept in domain-
specific thesaurus, as shown in Fig. 3, is the dimension of vectors which will be used to
represent test documents. Then, we compute concept weights for all concepts in do-
main-specific thesaurus for each test document by concept weighting schema defined in
definition 9, and execute the algorithm of UpdateConceptWeights to update the concept
weights for concepts which have sub-concepts. In this example, the value of ω is set to
0.5. The values of concept weights for each test document are used as the dimension
values for vectors. The results of document representation are shown in Table 3.
In evaluating the method, we are especially interested that the sub-concept and su-
per-concept relationships between terms extracted by our method may create a chance
to make two documents relevant. These two documents may be identified irrelevant
by traditional approaches. Table 4 shows the document representations by traditional
bag of terms approach with tf-idf weights value for the test documents. After Cosine
measuring, the similarity between test documents TestDoc1 and TestDoc2 is zero,
because, there are no common terms between these two documents. However, accord-
ing to the results of our method shown in Table 3, TestDoc1 and TestDoc2 are
accounted with small similarity by Cosine measuring, because they have a shared
concept in c1. The terms {"top-ranked documents", "query"} in c4 are sub-concept of
term {"local"} in c2, it seems reasonable, because selecting useful expansion terms
from top-ranked documents is the basic idea of local based query expansion. The term
{"local"} in c2 and term {"global"} in c3 inherit root concept c1 which contains term
{"technique"}; in this context, it may mean the technique of query expansion. There-
fore, although there is no term in TestDoc2 matching the terms in concept c1, Test-
Doc2, however, contains terms in c4 which indirectly inherits c1, and therefore c1
becomes a shared concept between TestDoc1 and TestDoc2 and makes these two
documents relevant.
Term dimensions
global technique local information top-ranked probability query query terms term
documents cluster
TestDoc1 0.48 0.18 0 0 0 0 0 0 0.48
TestDoc2 0 0 0 0.18 0.18 0.48 0.18 0.48 0
TestDoc3 0 0.18 0.48 0.18 0.18 0 0.35 0 0
5 Conclusion
References
1. Recupero, D.R.: A New Unsupervised Method for Document Clustering by Using Word-
Net Lexical and Conceptual Relations. Information Retrieval 10(6), 563–579 (2007)
2. Hotho, A., Staab, S., Stumme, G.: Wordnet Improves Text Document Clustering. In: Proc.
of the SIGIR 2003 Semantic Web Workshop (2003)
3. Hotho, A., Staab, S., Stumme, G.: Ontologies Improve Text Document Clustering. In:
Proc. of the IEEE International Conference on Data Mining (ICDM), pp. 541–544. IEEE
Press, New York (2003)
4. Miller, G.: WordNet: A Lexical Database for English. CACM 38(11), 39–41 (1995)
5. Hotho, A., Staab, S., Maedche, A.: Ontology-based Text Clustering. In: IJCAI 2001,
Workshop on Ontology Learning. Seattle, Washington, USA (2001)
6. Cimiano, P., Hotho, A., Staab, S.: Learning Concept Hierarchies from Text Corpora Using
Formal Concept Analysis. J. of Artificial Intelligence Research 24, 305–339 (2005)
7. Groojten, F.A., Weide, T.P.: Conceptual Query Expansion. Data & Knowledge Engineer-
ing 56(2), 174–193 (2006)
8. Wille, R.: Restructuring Lattice Theory: an Approach Based on Hierarchies of Concepts.
In: Rival, I. (ed.) Ordered sets, pp. 445–470. D. Reidel Publishing Company, Dordrecht
(1982)
9. A Formal Concept Analysis Homepage,
http://www.upriss.org.uk/fca/fca.html
10. Ganter, B., Wille, R.: Formal Concept Analysis - Mathematical Foundations. Springer,
Heidelberg (1999)
11. Brikoff, G.: Lattice Theory. American Mathematical Society, Providence (1967)
12. Rajapakse, R.K., Denham, M.: Text Retrieval with more Realistic Concept Matching and
Reinforcement Learning. Information Processing and Management 42(5), 1260–1275
(2006)
13. Carpineto, C., Romano, G.: A Lattice Conceptual Clustering System and its Application to
Browsing Retrieval. Machine Learning 24(2), 95–122 (1996)
14. Carpineto, C., Romano, G.: Exploiting the Potential of Concept Lattices for Information
Retrieval with CREDO. J. of Universal Computer Science 10(8), 985–1013 (2004)
15. Carpineto, C., Romano, G.: Concept Data Analysis: Theory and Applications. John Wiley
& Sons, Ltd., Hoboken (2004)
16. Salton, G.: A Vector Space Model for Automatic Indexing. CACM 18, 613–620 (1975)
17. Marcus, M.P., Santorini, B., Marcinkiewicz, M.A.: Building a Large Annotated Corpus of
English: The Penn Treebank. Computational Linguistics 19(2), 313–330 (1993)
18. Ou, S., Khoo, C.S., Goh, D.H.: Design and Development of a Concept-Based Multi-
Document Summarization System for Research Abstracts. J. of Information Science 34(3),
308–326 (2008)
19. Porter, M.F.: An algorithm for suffix stripping. Program. 14(3), 130–137 (1980)
20. Xu, J., Croft, W.B.: Improving the Effectiveness of Information Retrieval with Local Con-
text Analysis. ACM Tran. on Information Systems 18(1), 79–112 (2000)
On the Configuration of the Similarity Search
Data Structure D-Index for High Dimensional
Objects
1 Introduction
New similarity search applications are constantly being developed, ranging from
music similarity search to open source license violation detectors [1]. To support
such development, metric access methods (MAMs) are valuable tools as they can
be used by researchers to find suitable similarity measures for objects. Among
centralized metric access methods, the D-index [2] is regarded as one of the
fastest MAMs available . The index implements an addressing scheme based on
the relationship between each object and a set of objects called pivots taken
from the database. Pivoting techniques have been traditionally used to estimate
a lower bound of the distance function employed [3,4,5]. The combination of
these techniques is what makes this index unique.
The D-index requires several parameters whose configuration severely affects
performance. The optimal configuration of these parameters still remains an
open problem to date. In this paper, we study the structure of the D-index
in the context of costly distance functions. Our main objective is therefore to
minimize distance computations.
The D-index employs a clustering technique where data is divided into levels
composed of buckets. Additionally, it employs a pivot filtering technique [5] used
D. Taniar et al. (Eds.): ICCSA 2010, Part III, LNCS 6018, pp. 443–457, 2010.
c Springer-Verlag Berlin Heidelberg 2010
444 A.J. Müller-Molina and T. Shinohara
– The D-index tries to distribute evenly data among a set of buckets however,
in practice, about 80% to 99% of the data stays in one bucket.
– The search algorithm can access an exponential number of buckets with
respect to the cardinality of the pivot set employed.
2 Background
∀x, y ∈ D, d(x, y) ≥ 0
∀x, y ∈ D, d(x, y) = d(y, x)
∀x, y ∈ D, d(x, y) = 0 ⇐⇒ x = y
∀x, y, p ∈ D, d(x, y) ≤ d(x, p) + d(p, y)
The technique selects n pivot objects, and obtains the distances of all the objects
to these pivots. These pre-computed distances can be used to obtain a lower
bound of the real distance. Formally, for a metric space M = (D, d) and a set
of pivots P = {p1 , . . . pn }, the mapping Ψ : (D, d) → (Rn , L∞ ) can be defined:
The resulting vector is called the pivot vector. A lower bound of d can be obtained
with:
n
L∞ (Ψ (x), Ψ (y)) = max |d(x, pi ) − d(y, pi )| ≤ d(x, y).
i=1
This lower bound can be used in queries to eliminate objects if L∞ (Ψ (x), Ψ (y)) >
r, thus avoiding costly distance computations. Naturally, as the number of piv-
ots increases the pruning effect improves. The technique has been studied by
different authors in the past [5,6,3,7].
2.3 D-Index
The D-index was introduced by [8] and refined by [2]. The index partitions data
into h levels and 2ni − 1 buckets per level i. A function decides in which level
and bucket any given object must be placed. The main characteristic of the
index is that when a range query has a search radius up to some predefined ρ,
only one bucket must be accessed per level. Figure 1 shows how the structure
would look like for a particular example. Each square drawn with solid lines
represents a bucket that is used to store objects. For levels 0 to 2, these buckets
are called separable buckets. Dashed squares represent exclusion buckets that hold
objects that are recursively partitioned to the next level. At the very bottom,
the exclusion bucket E is employed to capture all the objects that cannot be
stored in all the previous levels.
To construct the D-index, for each level i, ni pivots pik (0 ≤ k < ni ) are
selected, and the corresponding median distances mik of all the objects that
entered the level and each pivot pik are calculated. The excluded middle parti-
tioning scheme [9] is employed to split the data. For x ∈ D, the split function
bps implements this partitioning scheme:
⎧
⎨0 if d(x, pik ) ≤ mik − ρ
bpsk,ρ
i (x) = 1 if d(x, pik ) > mik + ρ (2)
⎩
− otherwise
Objects within the interval ]mik − ρ, mik + ρ], belong to the exclusion set. In this
case, the function returns the value “−”. Separable sets are created from objects
446 A.J. Müller-Molina and T. Shinohara
Level 0, n0 = 2
(4 buckets)
Level 1, n1 = 2
(4 buckets)
Level 2, n2 = 1
(2 buckets)
Exclusion bucket
for which the bps function returns 0 or 1. An example of the split function bps
is illustrated in Figure 2. The exclusion set of width 2ρ contains the triangle-
shaped objects. The innermost circle represents one separable set, containing
circle-shaped objects. Finally, the square-shaped objects belong to the second
separable set. Given a range query such that r ≤ ρ, the function bps guarantees
that only one separable set is accessed. This property is called the separable
property.
mik
pik 2ρ
The separation of data into three sets is not sufficient for most applications.
Therefore, a higher order split function can be defined by concatenating the
results bk of each function bpsk,ρ ρ
i . The joint n-split function is denoted by bpsi (x)
and the return value is a string b = (b0 , . . . , bni −1 ) where bk ∈ {0, 1, −}.
To store each bucket, an addressing technique that receives a string and re-
turns a bucket identification number was proposed:
ni −1 k
k=0 2 bk if ∀k bk = −
b = (3)
2ni otherwise
When no “−” elements are returned by any of the bpsk,ρi functions, a number
smaller than 2ni is returned. By using the function bpsρi and the . operator,
On the Configuration of the Similarity Search Data Structure D-Index 447
Each bucket Bi,j contains a header that holds the pivot vectors for each object
stored in the bucket. The header is sorted by the first pivot, and the objects
can be stored in an alternate location. This bucket layout was called elastic
bucket [2].
Insert. Figure 3 shows the insertion procedure. For every level, the algorithm
attempts to find a bucket. If a bucket is found, the object is inserted (line 4)
and the process terminates. If all the levels are tried without success, the object
is inserted into the exclusion bucket E (line 8).
Range search. Figure 4 shows the range search algorithm. The algorithm eval-
uates each level and reads one or more buckets depending on r. If r is not greater
than ρ (line 4), two possibilities arise. In the case where both the separable set
and the exclusion set are intersected by the query (line 5), one separable bucket is
accessed. Otherwise, no separable bucket is accessed for the current level because
the query region is contained exclusively in the exclusion set of the level.
When the radius is greater than ρ, more than one separable set must be ac-
cessed per level. Recall that when invoking bpsr−ρ
i , a string b = (b0 , b1 , . . . , bni −1 )
is returned. To explore all the possible buckets, we must replace each bk with
value “−” alternately with 0 and 1, and generate all the possible bucket identifi-
cation number combinations. This is accomplished by calling function si in line
10.
In Figure 5, the function si receives a string b initialized with each bk = 0.
This string is used as a template to generate all the possible bucket identification
numbers. The function will process each split function bpsk,r−ρ i until a bucket
identification number can be created. When k < ni , the kth split function’s
value is computed(line 7). If the resulting value is “−” (line 9), the bucket
identification numbers where bk = 0 and bk = 1 are set in b, and the next split
function is recursively computed. Otherwise, the value obtained by bpsk,r−ρ i is
replaced and the next split function k + 1 is processed in line 12. When k = ni ,
a bucket identification number is already in b and bucket Bi,
b can be searched.
Insertion Algorithm
1: function insert(o ∈ D)
2: for i = 0 to h − 1 do
3: if bpsρi (o) < 2ni then o is not in E?
4: Store o in bucket Bi,bpsρi (o)
5: exit
6: end if
7: end for
8: Store o in bucket E
9: end function
store a small subset of the data. This represents a waste, as pivot vectors of the
upper levels are the most helpful in reducing distance computations because the
pivot set size is larger.
To illustrate this problem, in Table 1 the distribution of objects per level for
different ρ and h = 9 in two real data-sets is displayed. The description of the
data-sets is given in Section 4. The percentage of objects that pass to the next
level is shown for each level. In this example, the D-index will search data in
a limited number of buckets. As ρ grows bigger, the percentage of data stored
in the exclusion bucket increases. This behavior can be observed with relatively
large ρ and pivot sets. The large number of objects stored in the exclusion
bucket suggests that D-index behaves like sequential search on pivot vectors [6].
A solution is to reduce the number of pivots. For fast distance functions this is
possible. Nevertheless, as the complexity of a distance function increases, it is
desirable to have more pivots so that the lower bound becomes more precise.
The second problem occurs when ρ is small. In this case, the probability
that an object will fall into the exclusion zone will be also reduced. Recall from
Section 2.3 that when r > ρ, the function si must be called. This implies that
h−1
at most i=0 2ni buckets would have to be accessed. The generation of bucket
identification number combinations by si is an expensive operation.
Matching with a large set of pivots can negatively affect the distribution of buck-
ets. On the other hand, if distribution is improved by changing ρ, exponential I/O
450 A.J. Müller-Molina and T. Shinohara
Table 1. Level distribution for each level h. The column “%” contains the percentage
of objects that fall into the next level i + 1. The last row represents the percentage of
objects stored in the exclusion bucket. The column ni contains the number of pivots
for level i.
MTD ρ = 8 TED ρ = 7
i ni % ni %
0 16 95.06 30 99.99
1 15 93.26 28 99.99
2 13 92.16 26 99.99
3 12 91.65 24 99.99
4 11 90.61 22 99.99
5 10 90.11 20 99.99
6 9 89.68 19 99.99
7 8 88.88 18 99.99
8 7 88.14 17 99.99
3 Proposed Technique
The first heuristic is designed to guide si to buckets that are more likely to contain
objects close to the query. Our current implementation works only for discrete
numbers but it can be extended for real valued numbers. We pre-compute n0 arrays
fk (0 ≤ k < n0 ) that contain the distance distribution for each pivot pk0 . The
value fk [i] represents the number of objects that are at distance i from pivot pk0 .
For o ∈ D and a range r, the functions lk (o, r) and rk (o, r) are defined as:
On the Configuration of the Similarity Search Data Structure D-Index 451
m 0k
lk (o, r) = fk [j],
j=d(p0k ,o)−r
d(p0k ,o)+r
rk (o, r) = fk [j].
j=m0k
4 Experiments
– M T D: 350000 trees compared by the mtd [12] tree distance metric (O(n2 )).
The trees are binary program fragments [10] used by an open source license
violation detector. Each tree has up to 500 nodes.
– T ED: 245000 trees compared by the tree edit distance [13] metric (O(n3 )).
This data-set is a subset of M T D and includes trees that have up to 20
nodes.
We have chosen distance functions with relatively high complexity that require
many pivots to become feasible to match. As a rough estimation, one distance
computation takes as long as the computation of the L2 distance on a pair of
integer vectors of 6500 dimensions for mtd and 890000 dimensions for ted.
All our experiments present the averaged results of querying the data-set with
an exclusive set of 400 different objects. The maximum range query radii was
selected so that roughly 20% of the data is returned. The distance distributions
between data and query sets are illustrated in Figure 7.
We measured disk access operations, distance computations and total execu-
tion time. We implemented the algorithms in Java and executed them on Intel
Xeon computers running at 2.66GHz and with 4GB of memory on Ubuntu Linux
On the Configuration of the Similarity Search Data Structure D-Index 453
Frequency
0.06
0.05
0.04
0.03
0.02
0.01
0
1 10 100 1000
Distance
7.10. The GPL source code along with the data-sets can be downloaded1 . All
the indexes were built on top of the same framework2.
4.2 Indexing
Index M T D T ED
D 0.80 97
SD 0.15 13.75
P + 0.11 7
Table 3. Object Distribution per Bucket. Average, σ, and minimum and maximum
number of objects per bucket. The “Count” column contains the total number of buck-
ets created for SD-index and different ρ values of D-index.
6
00
0.
Seq = 350000 Seq = 3.575
00
60
10
0
5
00
00
0.
ρ=1
10
50
ρ=4
0
00
4
ρ=8
00
0.
10
40
ρ = 12
0
0
3
SD
00
10
0.
30
P+
0
10
2
00
0.
20
0
1
1
00
0.
10
1
0
,1
,3
,1
,3
,1
,3
0.
1
1
3
1
3
1
3
1
1
3
1
3
1
3
1
1
3
1
3
1
3
12
12
12
12
12
12
0,
3,
3,
6,
6,
9,
9,
0,
3,
3,
6,
6,
9,
9,
0,
3,
3,
6,
6,
9,
9,
Parameters r and k Parameters r and k Parameters r and k
30
00
00
25
Seq = 245000
10
Seq = 418
25
0
00
0
00
10
20
20
00
10
0
00
SD
15
15
0
10
ρ=7
0
P+
10
00
10
10
5
00
1
50
0
1
3
0,
2,
2,
5,
5,
7,
7,
1
0
0.
1
3
0,
2,
2,
5,
5,
7,
7,
0,
2,
2,
5,
5,
7,
7,
Fig. 9. Results for data-set T ED. Performance comparison among SD-index, P+tree
and a ρ = 7 D-index.
4.4 Discussion
When ρ is large, D-index resembles a sequential search on the pivot projection.
The fact that the D-index performs well for large ρ is not surprising as it has been
experimentally verified that sequential search based on pivot vectors outperforms
other hierarchical methods under certain circumstances [6]. The P+tree performs
the largest amount of read operations in all the experiments. On the other hand,
SD-index and P+tree are very close to each other in high dimensions because
456 A.J. Müller-Molina and T. Shinohara
their pivot set cardinality is the largest possible. The advantage obtained by
SD-index comes from heuristic one, as we descend first to buckets that are more
likely to have more objects.
5 Conclusion
We have studied the D-index in the context of high dimensional objects. Our
modification, the SD-index, consistently outperforms the D-index in different
configurations. Additionally, it guarantees that fewer distance computations will
be executed as the pivot set cardinality is maximized.
The D-index cannot scale well for small ρ and range r ≥ ρ and when the
number of pivots is large. In this case, the index must access an exponential
number of buckets that depends on the number of pivots. In our experiments
D-Index with 30 pivots was considerably slower than plain sequential search of
a very costly distance function. On the other hand, if we set a large ρ such that
r ≤ ρ, the index degenerates into a sequential search.
References
11. Bloom, B.H.: Space/time trade-offs in hash coding with allowable errors. ACM
Commun., 422–426 (1970)
12. Müller-Molina, A.J., Hirata, K., Shinohara, T.: A tree distance function based on
multi-sets. In: Chawla, S., Washio, T., Minato, S.-i., Tsumoto, S., Onoda, T., Ya-
mada, S., Inokuchi, A. (eds.) PAKDD 2008. LNCS, vol. 5433, pp. 87–98. Springer,
Heidelberg (2009)
13. Demaine, E., Mosez, S., Rossman, B., Weimann, O.: An optimal decomposition
algorithm for tree edit distance. In: ALP, pp. 146–157. Springer, Heidelberg (2007)
14. Zhang, R., Ooi, B.C., Tan, K.L.: Making the pyramid technique robust to query
types and workloads. In: ICDE 2004, p. 313. IEEE Computer Society, Washington
(2004)
15. Zezula, P., Amato, G., Dohnal, V., Batko, M.: Similarity Search: The Metric Space
Approach. Springer, Secaucus (2005)
Automatic Chinese Text Classification
Using N-Gram Model
1 Introduction
In recent years, there are more and more digital text services on the internet with the
rapid growth of the World Wide Web. Examples include Google, Yahoo, etc. Search
engine does not only provide user with the information, but also manage the informa-
tion effectively. In order to make it easy, many search engine systems classify the
texts in advance to organize their taxonomy. The methods of automatic text classifica-
tion help people to classify the documents. At present, many techniques have been
proposed to deal with the text classification problems. Yang and Liu [21] experiment
on the news text of Reuter. They reported that support vector machine (SVM) and k-
nearest neighbor (KNN) classifiers achieved the best accuracy in comparison to the
other four methods. Sebastian [14] pointed out that SVM had better performance in
general case.
However, support vector machine classifiers still have many problems. First, it
is designed to solve the binary-class classification problem. Therefore it should be
converted to handle multi-class classification problem. Second, the support vector
D. Taniar et al. (Eds.): ICCSA 2010, Part III, LNCS 6018, pp. 458–471, 2010.
© Springer-Verlag Berlin Heidelberg 2010
Automatic Chinese Text Classification Using N-Gram Model 459
machine classifiers generally could not generate probability. Most support vector
machine classifiers represent the similarity between the class and the text using the
cosine similarity.
Unlike probability, the cosine similarity is very abstractive to represent the similar-
ity between the class and the text. These problems are solved by Tipping [17]. He
combines the logistic regression in the support vector machine classifiers to generate
the probability. Meanwhile, the multi-class classification problem also can be trans-
formed into several binary-class classification problems. Nevertheless, a known
problem with the Tipping’s approaches is their relative inability to scale with large
problems like text classification. Fortunately, Silva and Ribeiro [15] solved the
problem by the method of dimension reduction.
Nonetheless, these standard approaches to Chinese text categorization has so far
been using a document representation in a word-based “input space” [6], i.e. as a
vector in some high (or trimmed) dimensional Euclidean space where each dimension
corresponds to a word. The advantages of this method are effective and high feasibil-
ity, but it assumes each word is independent with each other. Hence we need to solve
word segmentation first for Chinese text. Most text classifiers treat each feature are
mutually independent, word segmentation in Chinese is a difficult problem and we
could not ensure that every words which is segmented by some word segmentation
process are mutually independent. To solve it, character level N-gram models
[3][4][12][13][16] could be applied. These approaches generally outperform tradi-
tional word-based methods in term of accuracy. Moreover, these approaches can
avoid the word segmentation step and model the word-by-word relations [5][11].
Cavnar and Trenkle [3] proposed a classifier for English categorization, named
N-gram-based text categorization, which works with letters level feature. They used
N-grams as features for traditional feature selection process and then deployed classi-
fiers based on calculating feature-vector similarities. Cavnar and Trenkle’s approach
has a shortcoming. It was intractable when there is a large feature set, and standard
feature selection approaches do not always cope well in such circumstances. For ex-
ample, given sufficient features, the cumulative effect of uncommon features can still
have an important effect on classification accuracy, even though infrequent features
contribute less information than common features individually. Teahan and Harper
[16] used a PPM (prediction by partial matching) model for text classification where
they seek a model that obtains the best compression on a new document. Later, Peng,
et al. [12][13] proposed a classifier for Chinese classification without the word seg-
mentation step which works with letters character feature by means of using N-gram
model. In their experiment, their approach outperforms Cavnar and Trenkle’s ap-
proach [3], Support Vector Machine and naive Bayes in most case. Although Peng, et
al. proposed a novel classification to improve the traditional classifier for Chinese text
classification, the approach also involve in several problems. First, the research of
Peng, et al. entirely does not process the feature selection. If the testing text is too
large, the classifier maybe uses a lot of unnecessary conditional probability to esti-
mate the probability of a document to assign some class. Because the N-gram model
uses the expression (1) to estimate the probability of an event that a sequence of the
words w1…wT occur in a data set.
T
P( w1 ...wT ) = ∏ P( w | w
i =1
i i − n +1 ...wi −1 ) (1)
460 S.-J. Yen et al.
select important n-grams. In the third step, we build a logistic regression model to
estimate each conditional probability P(wi|wi-n+1…wi-1). According to the third step we
can make sure each conditional probability P(wi|wi-n+1…wi-1) converge within the
interval 0 to 1(i.e. 0<P(wi|wi-n+1…wi-1)<1).
According to the previous researches for feature selection [1], the chi-square statistic
is better than mutual information and information gain in selecting useful features.
However, the chi-square statistic cannot be applied to N-gram-based model without
amendment. Therefore, we amend it to suit N-gram-based model. First, we use an
example to expound what kind feature is significant in N-gram-based model as fol-
lows: Assume we wish to calculate a probability of a word sequence s=w1w2w3w4
occurred in two class, C1 and C2, using bi-gram model and the distribution of w1w2
and w1 between C1 and C2 is given in Table 1.
#( w1w2) #( w1)
C1 6 12
C2 1 2
PC ( w1w2 w3 w4 ) = PC ( w1 ) × PC ( w2 | w1 ) × PC ( w3 | w2 ) × PC ( w4 | w3 )
1 1 1 1 1
PC ( w1w2 w3 w4 ) = PC ( w1 ) × PC ( w2 | w1 ) × PC ( w3 | w2 ) × PC ( w4 | w3 )
2 2 2 2 2
In Tables 2 and 3, Let A1 be the number of times w1w2…wn and c co-occur, let B1
be the number of times w1w2…wn occurs without c, let P1 be the number of times c
occurs without w1w2…wn, let Q1 be the number of times neither w1w2…wn nor c, and
let N1 be the total number of n-gram occurred in corpus.
The chi-square statistic of n-gram w1w2…wn with class c is represented as follows:
N1 × ( A1Q1 − B1 P1 )
2
χ 2 ( w1w2 ...wn , c) =
( A1 + P1 )× (B1 + Q1 )× ( A1 + B1 )× (P1 + Q1 )
Identically, the chi-square statistic of (n-1)-gram w1w2…wn-1 with class c is repre-
sented as follows:
N 2 × ( A2Q2 − B2 P2 )
2
χ 2 ( w1w2 ...wn−1 , c) =
( A2 + P2 )× (B2 + Q2 )× ( A2 + B2 )× (P2 + Q2 )
It is clear that if an n-gram w1w2…wn whose chi-square value is more different than
chi-square value of (n-1)-gram w1w2…wn-1, then their distributions over all classes are
more different. Therefore, we propose a method as follows:
∑ F (w ...w , c)
1
F ( w1...wn , C ) = 1 n
C c∈C
⎧ χ 2 ( w1...wn , c)
⎪ 2 , if χ 2 ( w1...wn , c) ≥ χ 2 ( w1...wn−1 , c)
⎪ χ ( w ...wn −1 , c )
where F ( w1...wn , c) = ⎨ 2
1
⎪ χ ( w ...w , c)
1 n −1
, if χ 2 ( w1...wn , c) < χ 2 ( w1...wn−1 , c)
⎪⎩ χ 2 ( w1...wn , c)
The χ 2 ( w1 ...wn , c) is the chi-square statistic of n-gram w1w2…wn with class c, and the
χ 2 ( w1...wn−1 , c) is the chi-square statistic of (n-1)-gram w1w2…wn-1 with class c. We
can determine them by contingency table as follows: If the significance probability
(i.e. p-value) of the chi-square statistic of some n-gram w1w2…wn is less than the level
of significance, the n-gram is called significant n-gram. In experiment, we set the
level of significance equal to 5%, 10%, 15%, 20%, and 25%. Therefore, we can cal-
culate the significance probability of each n-gram. If the significance probability of
the n-gram is greater than the level of significance we set, we will not consider the n-
gram into the classifier.
Automatic Chinese Text Classification Using N-Gram Model 463
P (r = 1) =
1 + eα +α x +α x +...+α x
0 1 1 2 2 n n
We can treat #(wi-1) as the frequency we observe and #(wi-1wi) as the frequency of the
response which bi-gram is occurred. Assume there is a probability pi to represent the
probability of the response which bi-gram is occurred, the joint probability density
function is
j
⎛ ni ⎞ r
∏ ⎜⎜ ⎟⎟ pi (1 − p i )n − ri i i
i =1 ⎝ ri ⎠
464 S.-J. Yen et al.
1
According to the assumption of the logistic regression pi = , the joint
1 + eα +α x
0 1 1i +α 2 x2 i
⎜⎜ ⎟⎟⎜⎜ α +α x +α x ⎟
⎟ ⎜ α +α x +α 2 x2 i
⎟
i =1 ⎝ ri ⎠⎝ 1 + e ⎠ ⎝1+ e ⎠
0 1 1i 2 2i 0 1 1i
⎛ ni ⎞ α +α x +α x r
⎜⎜ ⎟⎟(e ) (1 + eα +α x +α x )
j
∏
− ni
= 0 1 1i 2 2i
i
0 1 1i 2 2i
i =1 ⎝ ri ⎠
Plo ( wi | wi −1 ) =
1 + eα +α P ( w |w )+α P ( w )
0 1 ML i i −1 2 ML i
Unlike the approach of Peng, et al. [12], our approach considers that a document is a
set of sentences, and a sentence is a word sequence. Because we believe that calculat-
ing the probability of a word considering the end of last sentence is unreasonable.
There is no poof the end of last sentence would affect the probability of a word in the
start of a sentence.
Assume we wish to classify a document d={s1,s2,…,s|d|} into a category
c ∈ C = {c1 ,..., c C } , where si = w1w2 ...w s . A natural choice is to pick the category c
i
that has the largest posterior probability given the text. That is,
c* = arg max Pr(c | d )
c∈C
3 Empirical Evaluation
The Chinese data set we used has been previously investigated in [19]. The corpus is
a subset of the TREC-5 People's Daily news corpus published by the Linguistic Data
Consortium (LDC) in 2001. The entire data set contains 164,789 documents on 59
categories, including IT, economy, etc. However, we could not use the entire data for
our experiments, because there many categories include too few documents. There-
fore, if we use the entire data for our experiments, there will be the imbalanced distri-
bution data problem [22] in it. To void the imbalanced distribution data problem, we
just choose particular of the entire data. The entire data set which we choose contains
105,733 documents on 22 categories.
Our goal is to classify text into a set of categories, and each document is assigned a
single category. Each n-gram-based classification, such as Peng et al’s approach [12],
466 S.-J. Yen et al.
is generally set the parameter n. For the experiments on Chinese data, we follow
Peng et al’s approach and experiment on uni-gram model, bi-gram model, and tri-
gram model. In each case, we randomly select 90% of documents evenly from
among the remaining categories to form the training data. The testing set contains the
other 10% documents of original data set. The training set and testing set do no over-
lap and do not contain replicated documents. Then we divide the training data into
9 parts uniformly. We pick 5 parts as training data 1, and pick all parts as training
data 2.
Predicted
Text belong to ci Text NOT belong to ci
Actual
Text belong to ci ai pi
Text NOT belong to ci bi qi
The precision, recall, and F-measure of the category c are shown as follows.
ai ai 2 × precision × recall
precision = , recall = , F − measure =
ai + bi ai + pi precision + recall
The aggregated confusion matrix and the micro-averaged precision, recall, and F-
measure of the category c are shown as follows.
Predicted
Actual Text belong to c ∈ C Text NOT belong to c ∈ C
C C
Text belong to c ∈ C A = ∑ ai P = ∑ pi
i =1 i =1
C C
Text NOT belong to c ∈ C B = ∑ bi Q = ∑ qi
i =1 i =1
Automatic Chinese Text Classification Using N-Gram Model 467
A A
micro-averaged precision = , micro-averaged recall =
A+ B A+ P
2 × micro-averaged precision × micro-averaged recall
micro-averaged F − measure =
micro-averaged precision + micro-averaged recall
Figures 3 and 4 give the results of comparisons of our approaches and Peng et al.’s
approach with feature selection on training data set 2 in tri-gram and bi-gram
models. In Figures 3 and 4, we can see that our approach outperform Peng et al.’s
approach in most case. Tri-gram model outperform bi-gram model in most case.
Remarkably, tri-gram model smoothing by logistic regression is more stable than
bi-gram model smoothing by logistic regression, because the difference in F-
measure of tri-gram model smoothing by logistic regression between best case and
worst case is less than 25%, but the difference in F-measure of bi-gram model
smoothing by logistic regression between best case and worst case is greater
than 43%.
Fig. 3. Comparisons of our approaches and Peng et al.’s approach with feature selection on
training data set 2 in tri-gram model
Figures 5 and 6 give the results of comparisons of our approaches and Peng et al.’s
approach with feature selection on training data set 1 in tri-gram and bi-gram models.
In Figures 5 and 6, we can see that our approach outperform Peng et al.’s approach in
most case. There are evidences that the F-measure of back-off smoothing method will
not be improved by feature selection. Therefore, the F-measure of back-off smoothing
method we be deteriorated following the decreasing of the number of features.
468 S.-J. Yen et al.
However, if we just select 80% number of features to train the classifiers, the F-
measure of back-off smoothing method looks few better then the F-measure of logis-
tic regression smoothing method, because there are too many “significant” features
are filtered out in this case. We also can find that the effect of sentence segmentation
is marginally in bi-gram model, because bi-gram model considers that each word is
observed is just related to the last word. In other words, there are rare bi-grams in
head and tail of sentences.
Fig. 4. Comparisons of our approaches and Peng et al.’s approach with feature selection on
training data set 2 in bi-gram model
Fig. 5. Comparisons of our approaches and Peng et al.’s approach with feature selection on
training data set 1 in tri-gram model
Automatic Chinese Text Classification Using N-Gram Model 469
Fig. 6. Comparisons of our approaches and Peng et al.’s approach with feature selection on
training data set 1 in bi-gram model
In the following, the results of the best cases of logistic regression smoothing
method and back-off smoothing method on each training data set in bi-gram model
and tri-gram model is represented in Table 7.
Table 7. The best cases of logistic regression smoothing method and back-off smoothing
method
Logistic Back-off
bi-gram tri-gram bi-gram tri-gram
training data set 2 0.800323 0.931512 0.596471 0.839820
training data set 1 0.739898 0.912540 0.513090 0.803358
In summary, first, generally our approach is better than Peng et al.’s approach in
most case. Second, feature selection has no help to classifier smoothing by back-off,
but very helpful to classifier smoothing by logistic regression. Third, sentence seg-
mentation impact on F-measure is marginally in classifier smoothing by back-off.
Fourth, logistic regression smoothing method generally outperform than back-off
smoothing method in most case. Fifth, tri-gram model is much better than bi-gram
model.
our experiment, we prove that it could improve the F-measure in most case. Third, we
consider a text as a set of sentences. According to our experimental result, this could
improve system performance especially in the case which use tri-gram model or
whose training data set is large enough.
In the future, we plane to find out a method to evaluate the relationship between
sentences. Since we believe that considering a text as a set of sentences is not an op-
timal assumption and there exist some fact could respect the relationship between
sentences. We also try to reduce the consumption of memory and calculative time of
CPU on feature selection.
References
1. Aizawa, A.: Linguistic Techniques to Improve the Performance of Automatic Text Catego-
rization. In: 6th Natural Language Processing Pacific Rim Symposium, pp. 307–314
(2001)
2. Chen, S., Goodman, J.: An Empirical Study of Smoothing Techniques for Language Mod-
eling. In: 34th Annual Meeting of the Association for Computational Linguistics, pp. 310–
318 (1998)
3. Cavnar, W., Trenkle, J.: N-Gram-Based Text Categorization. In: 3rd Annual Symposium
on Document Analysis and Information Retrieval, pp. 161–175 (1994)
4. Damashek, M.: Gauging Similarity with N-Grams: Language-Independent Categorization
of Text. Science 267, 843–848 (1995)
5. Dumais, S., Platt, J., Heckerman, D., Sahami, M.: Inductive Learning Algorithms and Rep-
resentations for Text Categorization. In: 7th International Conference on Information and
Knowledge Management, pp. 148–155 (1998)
6. He, J., Tan, A., Tan, C.: On Machine Learning Methods for Chinese Document Categori-
zation. Applied Intelligence 18, 311–322 (2003)
7. Jiang, E.: Learning to Semantically Classify Email Messages. In: 2nd International Con-
ference on Intelligent Computing, pp. 664–675 (2006)
8. Joachims, T.: Text Categorization with Support Vector Machines: Learning with Many
Relevant Features. In: European Conference on Machine Learning, pp. 137–142 (1998)
9. Lam, W., Ruiz, M., Srinivasan, P.: Automatic Text Categorization and Its Application to
Text Retrieval. IEEE Transactions on Knowledge and Data Engineering 11, 865–879
(1999)
10. Manning, C.D., Schuetze, H.: Fundations of Statistical Natural Language Processing, pp.
191–227. MIT Press, Cambridge (2004)
11. Peng, F., Huang, X., Schuurmans, D., Cercone, N.: Investigating the Relationship of Word
Segmentation Performance and Retrieval Performance in Chinese IR. In: 15th Interna-
tional Conference on Computational Linguistics, pp. 72–78 (2002)
12. Peng, F., Huang, X., Schuurmans, D., Wang, S.: Text Classification in Asian Languages
without Word Segmentation. In: 6th International Workshop on Information Retrieval with
Asian Languages, pp. 41–48 (2003)
Automatic Chinese Text Classification Using N-Gram Model 471
13. Peng, F., Schuurmans, D.: Combining Naive Bayes and N-Gram Language Models for
Text Classification. In: 25th European Conference on Information Retrieval Research, pp.
335–350 (2003)
14. Sebastian, F.: Machine Learning in Automated Text Categorization. ACM Computing
Surveys 34, 1–47 (2002)
15. Silva, C., Ribeiro, B.: Scaling Text Classification with Relevance Vector Machines. In:
IEEE International Conference on Systems, Man, and Cybernetics, pp. 4186–4191 (2006)
16. Teahan, W., Harper, D.: Using Compression-Based Language Models for Text Categoriza-
tion. In: Workshop on Language Models for Information Retrieval, pp. 83–88 (2001)
17. Tipping, M.: Sparse Bayesian Learning and the Relevance Vector Machine. Journal of
Machine Learning Research 1, 211–214 (2001)
18. Vapnik, V.: The Nature of Statistical Learning Theory. Springer, Heidelberg (1995)
19. Wu, Y.C.: Chinese Text Categorization with Term Clustering. M.S. Thesis, Mining Chuan
University (2003)
20. Yang, Y.: An Evaluation of Statistical Approaches to Text Categorization. Information Re-
trieval Journal 1, 69–90 (1999)
21. Yang, Y., Liu, X.: A Re-examination of Text Categorization Methods. In: 22nd Interna-
tional ACM SIGIR Conference on Research and Development in Information Retrieval,
pp. 42–49 (1999)
22. Yen, S., Lee, Y., Lin, C., Ying, J.: Investigating the Effect of Sampling Methods for Im-
balanced Data Distributions. In: IEEE International Conference on Systems, Man, and Cy-
bernetics, pp. 4163–4168 (2006)
Genetic Algorithms Evolving Quasigroups with
Good Pseudorandom Properties
1 Introduction
The need for random and pseudorandom sequences arises in many applications,
e.g. in modeling, simulations, and of course in cryptography. Pseudorandom
sequences are the core of stream ciphers. They are popular due to their high
encryption/decryption speed. Their simple and cheap hardware design is often
preferred in real-world applications. The design goal in stream ciphers is to
efficiently produce pseudorandom sequences - keystreams (i.e. sequences that
possess properties common to truly random sequences and in some sense are
”indistinguishable” from these sequences).
The use of quasigroups and quasigroup string transformations is a recent
but successful tendency in cryptography and coding [17]. With quasigroups in
the hearth of advanced cryptosystems and hash functions, a need to find good
quasigroups becomes hot topic.
D. Taniar et al. (Eds.): ICCSA 2010, Part III, LNCS 6018, pp. 472–482, 2010.
c Springer-Verlag Berlin Heidelberg 2010
Genetic Algorithms Evolving Quasigroups 473
2 Quasigroups
Definition 1. A quasigroup is a pair (Q, ◦), where ◦ is a binary operation on
(finite) set Q such that for all not necessarily distinct a, b ∈ Q, the equations
a ◦ x = b and y ◦ a = b.
have unique solutions.
The fact that the solutions are unique guarantees that no element occurs twice
in any row or column of the table for (◦). However, in general, the operation (◦)
is neither a commutative nor an associative operation.
Quasigroups are equivalent to more familiar Latin squares. The multiplication
table of a quasigroup of order q is a Latin square of order q, and conversely, as
it was indicated in [6,9,23], every Latin square of order q is the multiplication
table of a quasigroup of order q.
Definition 2. Let A = {a1 , a2 , . . . , an } be a finite alphabet, a n× n Latin square
L is a matrix with entries lij ∈ A, i, j = 1, 2, . . . , n, such that each row and each
column consists of different elements of A.
For i, j, k ∈ A the ordered triple (i, j; k) is used to represent the occurrence of
element k in cell (i, j) of the Latin square. So a Latin square may be represented
by the set {(i, j; k)| entry k occurs in cell (i, j) of the Latin square L.}
474 V. Snášel et al.
All reduced Latin squares of order n are enumerated for n ≤ 11 [19]. Let Ln
be the number of Latin squares of order n, and let Rn be the number of reduced
Latin squares of order n. It can be easily seen that
Ln = n!(n − 1)!Rn .
Number of distinct Latin squares of a given order grows exceedingly quickly
with the order and there is no known easily-computable formula for the number
of distinct Latin squares. The problem of classification and exact enumeration
of Latin squares of order greater than 11 probably still remains unsolved. Thus,
there are more than 1090 quasigroups of order 16 and if we take an alphabet L =
{0 . . . 255} (i.e. data are represented by 8 bits) there are at least 256!255! . . . 2! >
1058000 quasigroups.
Multiplication in quasigroups has an important property; it is proven that
each element occurs exactly q times among the products of two elements of Q,
q 2 times among the products of three elements of Q and, generally q t−1 among
the products of t elements of Q. Since there are q t possible ordered products of
t elements of Q, this shows that each element occurs equally often among these
q t products (see [10]).
◦0123
00321
12103
21032
33210
The table was created from table of modular subtraction. The second and the
third rows were exchanged. Permutations π, ρ were identities and ω = [0213]. A
multiplication in this quasigroup can be illustrated by e.g. 1 ◦ 0 = ω(1) ◦ 0 =
2 ◦ 0 = 2.
Starting with the quasigroup of modular subtraction, we can explore a large class
of quasigroups isotopic to the quasigroup of modular subtraction [11,22]. This
allows us to utilize quasigroups with very large number of elements without the
necessity of their storage in program memory. The multiplication in such isotopic
quasigroup is defined as follows:
a ◦ b = π −1 ((ω(a) + n − ρ(b)) mod n). (1)
We call the quasigroup defined by its multiplication formula and three selected
permutations an analytic quasigroup[15,27].
The notion of analytic quasigroup enables efficient work with large quasi-
groups. Previos studies in this field used mostly quasigroups of small order [14],
or just a small parts of certain quasigroup were utilized mainly as a key for
Message Authentication Code. Such small quasigroups are represented as look-
up tables in main memory. Larger quasigroup of order 2256 is used by NIST’s
SHA-3 competition1 candidate, hash function EdonR [13].
The properties of one analytic quasigroup isotopic to the quasigroup of mod-
ular subtraction were studied in [15]. The quasigroup was created using three
static functions that divided the sequence of n elements of the quasigroup into
several parts. The parts were rotated in various directions and exchanged among
themselves. It was shown that the investigated quasigroup has some faults in its
properties.
A quasigroup over a set of n elements requires log2 (n) bits to express each
element. Each permutation of bits in the element representation represents also
a permutation of all elements of the quasigroup (if n is a power of 2). Bit per-
mutation can be implemented easily as a function of q ∈ Q.
The bit permutation is an elegant way of implementing permutations over n
elements of Q. Although it enables us to explore only a fragment (log2 (n)!log2 (n)!
log2 (n)!) of all possible permutation triples over the quasigroup of n elements, it
is useful because it does not require all n elements in main memory and therefore
fits into the framework of analytic quasigroups.
Bit permutations are computationaly more expensive than the static functions
used to implement permutation in [15]. However, there are ongoing efforts to
implement bit permutation instructions in hardware, which would improve the
performance of the proposed algorithm significantly [12].
3 Genetic Algorithms
Genetic algorithms are generic and reusable population-based metaheuristic soft
optimization method [5,16,21]. GAs operate with a population of chromosomes
encoding potential problem solutions. Encoded individuals are evaluated using
a carefully selected domain specific objective function which assigns a fitness
value to each individual. The fitness value represents the quality of each can-
didate solution in context of the given problem. Competing individuals explore
the problem domain towards an optimal solution [16]. The solutions might be
encoded as binary strings, real vectors or more complex, often tree-like, hierar-
chical structures (subject of genetic programming [18]). The encoding selection
is based on the needs of particular application area.
The emulated evolution is driven by iterative application of genetic opera-
tors. Genetic operators algoritmize principles observed in natural evolution. The
crossover operator defines a strategy for the exchange of genetic information
between parents (sexual reproduction of haploid organisms) while the muta-
tion operator introduces the effect of environment and randomness (random
perturbation of genetic information). Other genetic operators define e.g. parent
selection strategy or the strategy to form new population from the current one.
Genetic operators and algorithm termination criteria are the most influential
parameters of every evolutionary algorithm. The operators are subject to do-
main specific modifications and tuning [21]. The basic workflow of the standard
generational GA is shown in Fig. 1.
Many variants of the standard generational GA have been proposed. The dif-
ferences are mostly in particular selection, crossover, mutation and replacement
strategy [16].
In the next section, we present genetic algorithm for the search for good
analytic quasigroups. It is an extended version of the initial GA for quasigroup
evolution introduced in[27]. In this study, we present modified GA for quasigroup
optimization with reengineered fitness function to find quasigroups that produce
good pseudoramdom sequences.
Genetic Algorithms Evolving Quasigroups 477
4.1 Encoding
As noted in section 2.3, any analytic quasigroup isotopic to quasigroup of mod-
ular subtraction is defined by three permutations. Such permutation triple rep-
resents a problem solution and should be mapped to one GA chromosome. Per-
mutations can be for the purpose of genetic algorithms encoded using several
strategies. In this study, we use random key encoding.
Random key (RK) encoding is an encoding strategy available for problems
involving permutation optimization [24]. In random key encoding, the permu-
tation is represented as a string of real numbers (random keys), whose relative
position changes after sorting corresponds to the permutation index. An example
or random key encoding is shown in (2).
0.2 0.3 0.1 0.5 0.4
Π5 = (2)
2 3 1 5 4
n
(Oi − Ei )2
χ2 = (3)
i=1
Ei
5 Experimental Optimization
Parameter value
Population size 20
Probability of mutation (PM ) 0.02
Probability of recombination (PC ) 0.8
Selection operator elitist
Max number of generations 1000
Genetic Algorithms Evolving Quasigroups 479
6
max
avg
4
Fitness
0
0 100 200 300 400 500
Generation
0.16
max
avg
0.14
0.12
0.1
Fitness
0.08
0.06
0.04
0.02
0
0 100 200 300 400 500
Generation
0.025
max
avg
0.02
0.015
Fitness
0.01
0.005
0
0 100 200 300 400 500
Generation
proved to 0.153 and for quasigroup of the length 2048 improved the optimization
procedure fitness from 0.00053 to 0.023. In all cases was the final improvement
achieved before generation 500.
The results demonstrate that genetic algorithms can improve randomly se-
lected quasigroups (isotopic to the quasigroup of modular subtraction) towards
properties defined by used fitness function. The resulting quasigroups are better
PRSG generators in terms of normality of symbol frequency in the generated
pseudo random sequence. Therefore, the optimized quasigroups are more suit-
able for applications in computer security.
6 Conclusions
In this paper was described a genetic algorithm for optimization of an ana-
lytic quasigroup. The genetic algorithm looks for good bit permutations that
are used to construct analytic quasigroups with desired properties. Both, the
analytic quasigroup and bit permutation, do not rely on the lookup table of
the quasigroup stored in memory. Therefore, large quasigroups can be used and
optimized efficiently.
A χ2 based goodness of fit statistics evaluating the randomness of pseudo-
random sequence produced by the quasigroup operation was used as a basis
of the fitness function. When used as a pseudorandom number generator, the
optimized quasigroups generate better pseudorandom sequences than randomly
chosen quasigroups of the same length.
Conducted numerical experiments suggest that the genetic algorithm is a
good tool to find optimized quasigroups. In our future work, we aim to find
more appropriate fitness function that will allow us to find quasigroups suitable
for cryptography and also other application areas.
Genetic Algorithms Evolving Quasigroups 481
Acknowledgment
This work was supported by the Czech Science Foundation under the grant no.
102/09/1494.
References
17. Knapskog, S.J.: New cryptographic primitives. In: CISIM 2008: Proceedings of the
2008 7th Computer Information Systems and Industrial Management Applications,
pp. 3–7. IEEE Computer Society, Washington (2008)
18. Koza, J.: Genetic programming: A paradigm for genetically breeding populations of
computer programs to solve problems. Technical Report STAN-CS-90-1314, Dept.
of Computer Science, Stanford University (1990)
19. McKay, B.D., Wanless, I.M.: On the Number of Latin Squares. Journal Annals of
Combinatorics 9(3), 335–344 (2005)
20. Merkle, R.C.: Secrecy, authentication, and public key systems. Stanford Ph.D.
thesis, pp. 13-15 (1979), http://www.merkle.com/papers/Thesis1979.pdf
21. Mitchell, M.: An Introduction to Genetic Algorithms. MIT Press, Cambridge
(1996)
22. Ochodková, E., Snášel, V.: Using Quasigroups for Secure Encoding of File Sys-
tem. In: Proceedings of the International Scientific NATO PfP/PWP Conference
Security and Information Protection 2001, Brno, Czech Republic, May 9-11, pp.
175–181 (2001)
23. Smith, J.D.H.: An introduction to quasigroups and their representations. Chapman
& Hall/CRC (2007)
24. Snyder, L.V., Daskin, M.S.: A random-key genetic algorithm for the generalized
traveling salesman problem. European Journal of Operational Research 174(1),
38–53 (2006)
25. Vojvoda, M.: Cryptanalysis of One Hash Function Based on quasigroup. In: Con-
ference Mikulášská kryptobesı́dka, Praha, pp. 23–28 (2003)
26. TREC Web Corpus: GOV (2009),
http://ir.dcs.gla.ac.uk/test_collections/govinfo.html
27. Snášel, V., Abraham, A., Dvorský, J., Ochodková, E., Platoš, J., Krömer, P.:
Searching for Quasigroups for Hash Functions with Genetic Algorithms. In: Pro-
ceedings of the 2009 World Congress on Nature & Biologically Inspired Computing,
pp. 367–372. IEEE Computer Society, Los Alamitos (2009)
Software Openness: Evaluating Parameters of
Parametric Modeling Tools to Support Creativity and
Multidisciplinary Design Integration
Abstract. The ubiquitous computing era has pushed the Architecture, Engineer-
ing, and Construction (AEC) industry towards new frontiers of digitally enabled
practice. Are these the frontiers originally identified by the pioneers in the
field? Architectural design has progressively shifted from two-dimensional
paper based pencil sketched models to digital models drawn in various Com-
puter-Aided Design (CAD) tools. The recent adoption of parametric modeling
tools from the aerospace industry has been driven by the need for tools that can
assist in rapid flexible modeling. The adaptation of parametric modeling has re-
formed both pedagogy and practice of architectural design. The question
remains if parametric design has answered all the requirements specified by
Steven Anson Coons in his 1963 proposal for a Computer-Aided Design (CAD)
system. Given the growth of computational power and ubiquitous computing,
how has CAD met the visions of its pioneers with respect to the flexibility and
ease of communication with the computer and support of simultaneous design
conversations with many designers working on the same project? This paper
will revisit ideas conceived by the early inventors of CAD, explore the opportu-
nities for advancing parametric modeling with the existing ubiquitous comput-
ing infrastructure, and introduces the notion of software openness to support
creativity and multidisciplinary design integration.
1 Introduction
The rise of Computer-Aided Design (CAD) and Computer-Aided Manufacturing
(CAM) has led to widespread adoption of digital modeling in the field of manufactur-
ing and construction. There is an increasing demand for flexible modeling tools that
allow rapid adaptations to design variations to be produced in real time. CAD is
a term originally introduced by Steven Coons [1] and Ivan Sutherland [2] as an
integrated design approach which leverages both the creativity and imagination of
man and the analytical and computational power of the computer. CAD eventually
evolved to tools that truly assist designers. Parametric design software, which was
hitherto used mainly in the aerospace and automotive industry, had been adapted to
D. Taniar et al. (Eds.): ICCSA 2010, Part III, LNCS 6018, pp. 483–497, 2010.
© Springer-Verlag Berlin Heidelberg 2010
484 F.D. Salim and J. Burry
the Architecture, Engineering and Construction (AEC) industry [3], [4]. We, a com-
puter scientist (the first author) and a practicing architect in academia (the second
author), are involved in various integrated multidisciplinary architecture-engineering
projects which require explorations of existing design tools and techniques.
This paper aims to investigate the current state-of-the-art of digital modeling and
reevaluate parametric design software retrospectively in the light of the aims of early
CAD inventors. The review also seeks to discover the requirements for a truly flexible
and open parametric design tool that can leverage creative design thinking and mul-
tidisciplinary design integration.
The paper is organized as follows. Section 2 presents the summary of the require-
ments for CAD as proposed by the originators. Section 3 reviews the existing para-
metric CAD software in the light of the requirements stated in the previous section
and outlines the research gap that needs to be investigated. Section 4 presents our
proposed notion of software openness given the need to leverage creative design
thinking and multidisciplinary design integration. Section 5 concludes the paper.
Fig. 1. Left: The clerestory window in Sagrada Familia, Image courtesy of: Mark Burry. Right:
Dubai Towers. The Lagoons Image courtesy of: Thompson, Ventulett, Stainback & Associates
(www.tvsa.com).
Fig. 2. Left: An early digital model of the Bird’s Nest Beijing National Stadium as modeled
inside Digital Project. Right: the final model. Image courtesy of: Gehry Technologies LLP
(www.gehrytechnologies.com).
We perceive two different groupings of parametric design tools given the varying
approach to parametric design. The first group of parametric design tools is based on
associative-geometry, where parameterized mathematical descriptions and associa-
tions between points, curves, surfaces, and solids are possible. The tools belonging to
this group include Bentley GenerativeComponents and Rhino Grasshopper. The sec-
ond group of parametric design tools are focused on BIM, where parametric relation-
ships encapsulate parametric descriptions of components of a building design across
multiple disciplines [10]. BIM stands for Building Information Modeling, which is
regarded as “a data-rich, object-oriented, intelligent and parametric digital represen-
tation of the facility” by the American General Contractors (as quoted in [11]). The
Autodesk Revit 2010 package (Revit Architecture 2010, Revit MEP 2010, Revit
Structure 2010) belongs to this group as well as Gehry Technologies’ Digital Project.
Revit 2010, however, is unable to handle NURBS geometry modeling [12] and
Software Openness: Evaluating Parameters of Parametric Modeling Tools 487
therefore Revit is not generally used for modeling complex curved geometries. Digital
Project is known to be a very powerful parametric CAD package that handles both
complex parametric geometric associations as well as parametric design representa-
tions referring to diverse libraries across multiple disciplines. Nevertheless, it is also
known to be the most expensive CAD package that is currently around and customiz-
ing model via the script interface is not trivial. The first group of tools can also be
used for parametric representations that link different disciplines, however, BIM high-
level object definitions are not included in the tools by default.
In a reflective mode, we review the existing parametric modeling tools and revisit
the CAD requirements by the CAD visionary and retrospectively raise the question –
are we there yet? The following evaluation is raised in the light of the use of the cur-
rent parametric modeling tools and processes.
3.1 Are They Capable of Accepting, Inferring, and Storing Geometries Using
Graphical and Symbolical Means?
This section evaluates the existing parametric modeling tools’ capabilities to process
and manage the geometries in the design using graphical and symbolical means. The
abstractions of the design in parametric modeling tools are often represented in a
symbolic diagram. These features are now intrinsic to all the existing parametric de-
sign tools. GenerativeComponents (GC) encompasses a 3D interactive view, a sym-
bolic view, and an object view [13]. One distinguishing feature of GC is the way it
displays both the textual and symbolic representation of the parametric associations
between components in the model. This feature is also available in Grasshopper, the
new extension of Rhino. The symbolical representation of a parametric model is
analogous to an object diagram of a software application. In fact, the notion of com-
ponents, parameters, constraints, and associations in a parametric model resemble
those of the object-oriented programming paradigm. The graphical and symbolical
representations undoubtedly assist designers to rethink their design in the light of the
components and their associations in the model as well as the parametric variations
they would like to generate from the model. Both the graphical and symbolical repre-
sentations visualize design intent that is captured within the tool, particularly during
the early design stage when design options need to be generated rapidly.
Revit 2010 uses the more rigid tree structure to represent the higher level of ab-
straction of the model. Geometrical associations between components in the model
are not visualized in the tree structure. Since each component in a Revit model is an
instance of a Revit family type (e.g. an instance of a 36’ flush panel type of the door
family), the tree structure provides the abstraction and the detailed view of the
instances of each type and family that are created in the model. Symbolic representa-
tions are used to view the 3D model as engineering plans. However, there is no exist-
ing symbolical abstraction to represent the parametric associations of all the parts in a
Revit model. Considering that Revit is aimed to be a BIM software, the tree structure
is useful for viewing abstract representations of parts of the model to give a non disci-
pline specific overview.
CATIA/Digital Project (DP) primarily uses a hierarchy of linked files. No single
file need be very large. It also uses simple graphical files that just show the geometry
for representing large extents of the model where it might otherwise take hours or
488 F.D. Salim and J. Burry
days to open the Part and Product files. There is a tree structure in CATIA/DP that is
an organizational and navigational tool and another that directly maps the model’s
parametric relationships.
To date, CAD tools (i.e. parametric design tools) and analysis tools have been distinct
and autonomous. Cross-disciplinary analysis is a challenge since current CAD sys-
tems are not integrated well with CAE (Computer Aided Engineering) analysis tools
[Kasik05]. Some tools, such as Revit Architecture 2010 has released a plug-
in/extension for solar incidence and energy analysis to be performed within the soft-
ware. Another example is Google Sketchup, which has an extended capability for a
simple energy analysis when the IES Virtual Environment plug-in is installed with it.
However, there is no existing tool on the market that currently carries the capability to
perform all-in-one design and analysis, where the analysis can respond to the design
variations made in the same model. Some recent work has come close to realizing this
silver bullet [14].
In order for a parametrically-modeled design proposal to be assessed for structural
loads, energy efficiency, solar incidence analysis, lighting analysis, and other types of
analyses, it must undergo “expert translation” since geometrical representation in vari-
ous domains are interpreted differently [15]. Translation errors often arise from two
key sources, which are floating-point arithmetic and tolerances [6]. Approximations
used in numerical calculations reducing floating-point that substitute higher perform-
ance for high precision cause errors in models passed to disciplines that require high
precision analysis [6]. Tolerances are used to control the level of approximation and
need to be managed to produce geometry that can meet different discipline analysis
requirements. If the tolerances are set high, analysis can be performed more quickly
but with greater likelihood of error in the resulting geometry [6]. If the tolerance is too
small, analysis may fail to converge at all [6]. Therefore, selecting a tolerance to be
used requires good comprehension of the engineering analyses to be used, the envi-
ronment of the analyses, and the specific software to be employed [6].
3.3 Are the Models Generic and Adaptable to Various Domains Pertaining to
the Design Process?
This section deals with Coons’ requirement for a CAD system that can support cross-
disciplinary model interoperability. According to a survey by McGraw-Hill construc-
tion study [16], 3% of a project cost corresponds to software non-interoperability. In
the U.S. construction market worth $1.3 trillion in 2007, this corresponds to $36 bil-
lion of productivity waste. In the global market worth $4.6 billion, the figure escalates
to $138 billion [16]. In order to maintain building information efficiently and the vary
representations in a broadly reusable way, the building industry has recently at-
tempted to define and encourage the adoption of Building Information Modeling
(BIM). With too many industry players and major CAD vendors introducing a num-
ber of commercialized BIM tools, however, many have found themselves reinventing
the wheel in order to bridge tools created by one vendor to another tool by another
vendor.
Software Openness: Evaluating Parameters of Parametric Modeling Tools 489
Industry Foundation Classes (IFC), the accepted industry standard for interopera-
bility and model exchange, are too heavy and complex (IFC consist of 327 data types,
653 entity definitions, and 317 property sets [17]). This poses challenges to perform-
ance, scalability, and ease of use and adaptation into specific sub-domains. Green
Building XML (gbXML), which was introduced as a lighter option to IFC, only en-
compassing a subset of IFCs has become a de facto standard for interoperability of
green building design. However, semantic translations from those standards to BIM
tools and vice versa (import to/export from BIM tools) still generate translation errors.
Lee et al argue that parametric modeling needs to be central to BIM and provide
mechanisms for translation to various domains [7].
Over the top of any provision for collaboration and model sharing and merging within
existing parametric design software packages, there is a need for additional communi-
cation channels that combine graphical, symbol, and language based information.
Currently, separate tools are required Product Lifecycle management tools for manu-
facturing aim to fulfill this role by integrating information and communication with
customers/clients, suppliers, enterprise/planning and systems development informa-
tion and communication.
However, in practice it is a lightweight tool that allows rapid human interaction as
well as capturing the history of informal decision making and rapid change in the
design model that is required in combination with parametric modeling. Autodesk
Design Review and Octopz are examples of visual annotation environments that sup-
port this level of interaction. Versioning software has been integrated with Digital
Project parametric software as the means of team model sharing online, allowing
visual indicators of ownership of different parts of the model, locked in the versioning
software, at any given time within the modeling environment. The concept is familiar
to Archicad Teamwork and Bentley Projectwise users.
There are two fundamental truths that Jerry Laiserin has stated of the current state
of technology in the industry: that while collaboration across design teams is the most
critical factor for successful building information modeling, as model data integration
goes up, flexibility of workflow and performance in collaboration go down [18]. This
is true even for the most straight forward, explicit geometrical representation of build-
ings and currently leaves models shared between large teams of a fundamentally more
formative and flexible nature largely in the aspirational realm even 40 years on.
One of Coons’ manifestos is for CAD to promote design conversations between the
computer and the designer. In parametric design tools such as Digital Project and
GenerativeComponents (GC), the history of geometrical associations and parametric
value variations is recorded. In GC, parametric changes are recorded as transactions
and users can go back and forth in the transaction list to undo or redo changes in the
model. Whenever the value of a parameter or the parametric relationship between one
component to another is changed, the value and relationship changes are recorded in
490 F.D. Salim and J. Burry
the transaction file. This is one way for CAD tools to carry “design conversations”,
since designers constantly perform iterative loops in evaluating their design and men-
tal design conversations that are channeled through their modeling activities can be
recorded.
However, difficulties in communicating design intent for multidisciplinary model
sharing still persist. When standards of model interoperability exchange such as IFC
or gbXML are used for model sharing, parametric representations and associations are
not captured either by IFC or gbXML during translation, since the dynamic associa-
tions between the components in the model are replaced with static values represent-
ing the components themselves. The design intent that needs to be maintained in a
parametric model is lost in translation and unrecoverable once the model is imported
or exported to other tools or analysis programs.
parametric variations grows exponentially, the flexible model poses the threat of in-
flexibility given that the proliferation of backward/forward chaining of changes can
be unmanageable [21]. This is where customization, scripting, or coding comes in to
pull the weight off from visual design. Designers are now becoming more interested
in increasing the efficiency of model building through scripting or coding. To design-
ers who code, coding activity is considered as a channel for creativity and means of
representing design ideas that goes beyond visual or graphical drawing. Current tools
have indeed assisted creative thinking. However, to innovate, designers need to ex-
tend the exploration of their design ideas using complementary creative methods to
the use of digital/visual synthetic geometric modeling aids.
Lawson [22] stated his desire to see CAD becoming an agent for creativity. As an
agent, CAD would act as a subordinate who learns and understands the design inten-
tion of the master and is able to assist creative thinking by crawling the internet and
finding design ideas that can match the master’s interest [22]. Lawson’s desire can
potentially be realized if parametric modeling is integrated with such computational
techniques as machine learning or data mining. The existing work involving machine
learning techniques for the design process includes evolutionary design [23], [24] and
agent based modeling [25].
The initial need for parametric design tools is the requirement to support flexible
modeling. Flexibility should be an inherent attribute of any parametric design tool. In
the light of the converging physical and digital spaces, architectural form inside the
CAD should keep on transforming and producing emergent results which are reflec-
tive on the physical environmental changes as well as the virtual energy simulation
results. The current state-of-the-art technology in the cyber/digital world, such as
sensing technology, ubiquitous computing, and cloud computing, can potentially be
combined with the power of parametric modeling to enhance the flexibility, fluidity,
and creativity of designing. Therefore, there is an opportunity to leverage the existing
parametric modeling tools by keeping it open to the user, the environment, and other
existing tools in the space. In addition, given the complexity of the process, analysis,
and tools required in various disciplines, there is a need to escalate the degree of soft-
ware openness in order to carry trans-disciplinary design conversations. This is dis-
cussed further in the next section.
and open output, the more usable the tool is in the context of a multidisciplinary de-
sign project. Since a parametric design tool do not generally include the full range of
engineering analysis tools, such as building thermal performance analysis and struc-
tural analysis tools, the higher the degree of the software openness, the higher is the
possibility of the software being used in design model integration across disciplines
and design activities. Since performance analysis and software interoperability are
still challenges in multidisciplinary design model integration, the ability to customize
design representation, translation, and input/output from one software to another en-
ables engineers to customize the software for the purpose of the project. The need to
assess the software openness for interoperability is demonstrated in the PM4D project
[27] by Centre for Integrated Facility Engineering (CIFE), Stanford University, where
the input and output from one design, engineering analysis, or visualization software
to another are compared for the best selection of tools required for the integrated
project delivery.
Open Input. Input to a parametric tool can either be models (imported from another
CAD) or textual files containing analysis requirements (which could be a result im-
ported from external analysis tools). In a multidisciplinary design project, it is neces-
sary to analyze the input capabilities of the tools being used and the existing model
translation mechanisms between one software to another. The interoperability gap
between software, where it exists, could then be tackled in two ways. Firstly, the
‘tool-by-tool’ approach which employs a specific tool or plug-in for export and im-
port of a design model between one tool and another, such as demonstrated by Nicho-
las and Burry [15] in translating a model created in Rhino CAD software to a model
understandable by Radiance, a lighting analysis software. Secondly, the ‘bridge tool’
approach which employs an intermediate tool for model translation among various
CAD, analysis, and visualization software. This approach was exemplified in the
development of DesignLink Software Development Kit (SDK) [28], [29], by Arup
and Spatial Information Architecture Laboratory (SIAL), which relies on the transla-
tion of various models into DesignLink XML, an XML schema designed for model
information exchange in the early design stage.
In order to promote design conversations between human designers and the com-
puter, various sensor and haptic devices need to be explored as an input device to pa-
rametric modeling. Handheld devices help designers to converse with the parametric
494 F.D. Salim and J. Burry
models better. For example, a tangible view cube was developed to interface with
Autodesk tools in order to represent the 3D digital model in a physical environment
and promote tangible interaction between the user and the CAD [30]. The ClayTools
system from Senseable [31] is a free-form robot hand that allows model creation, craft-
ing, and detailing straight to the virtual model with a force feedback applied to the
user. The ClayTools can become input to 3DMax, Maya, and Rhino, which are not
parametric modeling software. However, with Grasshopper, Rhino’s parametric plug-
in, there is a potential in connecting Rhino-Grasshopper virtual parametric models with
ClayTools. Another example of a haptic device that can be used for exploring, design-
ing, and generating inputs for parametric models is the Nintendo Wii remote (or Wii-
mote). A Wiimote plug-in for Autodesk Design Review by Autodesk Labs was devel-
oped for designers to navigate through the 3D model and environment using the Wii-
mote [32]. Another existing work on development of haptic input devices to parametric
modeling software is the Wiimote features development for Generative Components
(GC) [33], which allows the Wiimote to interact with the GC to generate forms and
variations based on real-time data input. The Wiimote can be used to draw in 3D space,
modify existing 3D models, or used as a controller of the model (Fig. 5).
Open Process. Software openness can also be used to assess the suitability of the
software to be openly customized in order to generate a creative output. Creativity can
be achieved not only through physical or digital modeling, but also through creative
coding [34]. Parametric design and analysis tools need to accommodate for exten-
sions, customization, and open interoperability, in order that additional plug-ins, class
libraries, software components, and scripts can be developed to support creativity and
productivity. Interoperability checking and automation through code can be devel-
oped when the communication channel between software is open. In Digital Project,
customization through codes is available from their internal scripting window, which
uses VBA script or by expert use of the CATIA SDK. Given the nature of VBA as
procedural language, component and object reuse in the code is a challenge since
classes cannot be defined inside the VBA script. The open component based software
architecture adopted by GC and Grasshopper (Rhino) provides a better flexibility
for customization in comparison to non-parametric modeling tools which largely rely
on internal scripting techniques. GC and Grasshopper (Rhino) allow installation of
Software Openness: Evaluating Parameters of Parametric Modeling Tools 495
custom plug-ins or Dynamic Link Libraries (DLLs) to define new features, objects, or
parametric relationships or to extend/inherit from the existing type libraries. Autodesk
offers an Application Programming Interface (API) to access the classes, methods and
operations provided by Revit programmatically on .NET platform. Commercial and
non-commercial parametric modeling software could open the process of definition,
generation, modification, and deletion of design objects and parametric relationships
among the objects in the model to promote collaboration, interoperability, design
integration, creativity and interactivity with other software, sensors and haptic
devices.
Open Output. Coons’ paper opens with reference to the first automatically controlled
milling machine at MIT in the 1950s [1]. The CAD/CAM axis so far seems to have
proved if anything more vital in the design process than immersive virtual reality.
Designing and building in the physical world entails, as it always has, working from
and capturing a physical context, representing this virtually, whether through meas-
urement, drawing and physical modeling or through digital capture and representa-
tion. Similarly, collective design decision making is still frequently better supported
through interaction around physical models. While the cost and speed of traditional
physical modeling is in some cases uncompetitive with virtual equivalents during the
design process, the opportunity for rapid prototyping allows for the design conversa-
tion to flow back and forth between the virtual and physical embodiment.
often require real-time interaction and feedback that can be manifest in the physical
world. When a parametric design software unlocks the means for software openness,
sensors and haptic devices can be linked with parametric design software to achieve
tools that are capable of simulating adaptive and responsive forms of architecture.
Finally, the degree of software openness can be used to measure the flexibility of
the software to be openly customized for interoperability, model sharing, and support-
ing creativity.
Acknowledgment
The authors thank Australian Research Council (ARC) for the funding of the research
in the ARC Discovery project: Challenging the inflexibility of the flexible model. We
also thank the ARC and Queensland State Government’s Project Services for the
funding of the research in the ARC Linkage project: Assimilation of architectural and
services design in early design modeling. We acknowledge John Frazer, Robin Dro-
gemuller, and Bianca Toth from Queensland University of Technology (QUT) and
Mark Burry from RMIT University for their contribution to the research team. The
authors also acknowledge Daniel Davis from RMIT and Ruwan Fernando from QUT
for their assistance.
References
1. Coons, S.A.: An outline of the requirements for a computer-aided design system. In:
AFIPS Spring Joint Computer Conference, pp. 299–304. ACM, New York (1963)
2. Sutherland, I.E.: Sketchpad: a man-machine graphical communication system. In: AFIPS
Spring Joint Computer Conference, pp. 329–346. ACM, New York (1963)
3. Frazer, J.: Computing without computers. Architectural Design 75(2), 34–43 (2005)
4. Hernandez, C.R.B.: Thinking parametric design: introducing parametric Gaudi. Design
Studies 27(3), 309–324 (2006)
5. Dubendorf, V.A.: Wireless Data Technologies. John Wiley & Sons, West Sussex (2003)
6. Kasik, D.J., Buxton, W., Ferguson, D.R.: Ten CAD Challenges. IEEE Computer Graphics
and Applications 25(2), 81–92 (2005)
7. Lee, G., Sacks, R., Eastman, C.M.: Specifying parametric building object behavior (BOB) for
a building information modeling system. Automation in Construction 15(6), 758–776 (2006)
8. Monedero, J.: Parametric design: a review and some experiences. Automation in Construc-
tion 9(4), 369–377 (2000)
9. Burry, M.: Parametric Design and Sagrada Familia. Architectural Research Quarterly 1(4),
70–80 (summer 1996)
10. Drogemuller, R.M., Crawford, J., Egan, S.: Linking early design decisions across multiple
disciplines. In: Proceedings of the Conference on Product and Process Modelling in the
Building and Construction Industry (ECPPM 2004), Istanbul, Turkey (2004)
11. Aranda-Mena, G., Crawford, J., Chevez, A., Froese, T.: Building information modelling
demystified: does it make business sense to adopt BIM? International Journal of Managing
Projects in Business 2(3), 419–434 (2009)
12. Autodesk Inc.: Autodesk Conceptual Design Curriculum 2010 (2009),
http://usa.autodesk.com/adsk/servlet/
item?siteID=123112&id=14038871
Software Openness: Evaluating Parameters of Parametric Modeling Tools 497
13. Aish, R., Woodbury, R.: Multi-level Interaction in Parametric Design. In: Butz, A., Fisher,
B., Krüger, A., Olivier, P. (eds.) SG 2005. LNCS, vol. 3638, pp. 151–162. Springer, Hei-
delberg (2005)
14. Maher, A., Burry, M.: The Parametric Bridge: Connecting Digital Design Techniques in
Architecture and Engineering. In: The 2003 Annual Conference of the Association for
Computer Aided Design in Architecture, Indianapolis, Indiana, pp. 39–47 (2003)
15. Nicholas, P., Burry, C.M.: Import as: Interpretation and Precision Tools. In: Proc. of the
12th International Conference on Computer Aided Architectural Design Research in Asia,
CAADRIA 2007, Nanjing, China (2007)
16. McGraw-Hill Construction: Interoperability in the Construction Industry (2007),
http://construction.ecnext.com/mcgraw_hill/includes/SMRI.pdf
17. Steel, J., Drogemuller, R.: Model Interoperability in Building Information Modelling. In:
Industrialized Software: KISS workshop @ Code Generation 2009, Cambridge, UK (2009)
18. Laiserin, J.: Next Gen BIM (2009),
http://www.laiserin.com/features/issue25/feature01.pdf
19. Hoffmann, C.M., Joan-Arinyo, R.: Parametric Modeling. In: Farin, G., Hoschek, J., Kim,
M.-S. (eds.) Handbook of Computer Aided Geometric Design, ch. 21. Elsevier B. V, Am-
sterdam (2002)
20. Aish, R.: Migration from an individual to an enterprise computing model and its implica-
tions for AEC Research. In: Berkeley-Stanford CE&M Workshop: Defining A Research
Agenda, Stanford, California (2000)
21. Burry, J., Burry, M.: The Bonds of Spatial Freedom. In: Proc. of the 26th eCAADe Con-
ference - Architecture in Computro, Antwerpen, Belgium, pp. 301–308 (2008)
22. Lawson, B.: Oracles, draughtsmen, and agents: the nature of knowledge and creativity in
design and the role of IT. Automation in Construction 14(3), 383–391 (2005)
23. Frazer, J.: An Evolutionary Architecture. Architectural Association Publications, London
(1995)
24. Janssen, P.: A generative evolutionary design method. Digital Creativity 17(1), 49–63 (2006)
25. Coates, P.S.: Some experiments using agent modelling at CECA. In: Generative Arts 2004,
Milan, Italy (2004)
26. Hansen, M.: Wearable Space. Configurations 10(2), 321–370 (2002)
27. Fischer, M., Kam, C.: Product Model & 4D CAD - Final Report. Stanford University (2002)
28. Khemlani, L.: SmartGeometry 2009 Conference Day (2009),
http://www.aecbytes.com/feature/2009/SmartGeometry2009.html
(accessed April 27, 2009)
29. Holzer, D., Tengono, Y., Downing, S.: Developing a Framework for Linking Design Intel-
ligence from Multiple Professions in the AEC Industry. In: Dong, A., Moere, A.V., Gero,
J.S. (eds.) Computer-Aided Architectural Design Futures (CAADFutures) 2007, Sydney,
Australia. Springer, Netherlands (2007)
30. Babu, S.: Tangible View Cube. CAD Professor (2009),
http://cadprofessor.in/2009/04/tangible-view-cube
31. Senseable, http://www.sensable.com
32. Sheppard, S.: A Wii Bit of Fun. In: Autodesk Labs: It’s Alive in the Lab (2009),
http://labs.blogs.com/its_alive_in_the_lab/2007/12/
a-wii-bit-of-fu.html
33. Salim, F.D., Mulder, H., Burry, J.: A System for Form Fostering: Parametric Modeling of
Responsive Forms in Mixed Reality. In: Proc. of the 15th International Conference on
Computer Aided Architectural Design Research in Asia, New Frontiers. Hong Kong,
China (2010) (accepted for publication)
34. Maeda, J.: Creative Code. Thames & Hudson, London (2004)
Dynamic and Cyclic Response Simulation
of Shape Memory Alloy Devices
1 Introduction
Since the first large-scale application of shape memory alloys (SMAs) as couplings in
the Grumman F-14 aircraft, the implementations of SMA devices have been spread
from aerospace to medical field, smart structures, robots etc.. The noninvasive charac-
teristic and excellent bio-compatibility make SMA devices widely applied in biomed-
ical fields. The controllable deformation feature of SMA devices has attracted many
attentions in robot researching field. SMAs have been used as actuators and sensors in
several experimental robots, such as robots made by Sugiyama and Hirai [1]. Besides,
the adjustable stiffness feature of SMAs have been used in passive control devices of
vibration controlling, showing a remarkable capability in controlling vibrations with
unknown frequency. Even more, the damping capacity of SMAs has become another
active research field recently.
However, the practical applications are still limited in small scale such as applica-
tions in medical instruments and aerospace devices. One reason is the high price of
SMAs, which has made the applications of SMA devices limited in the fields which
are not sensitive to cost. Another reason is a lack of understanding of SMAs' proper-
ties and simulation tools, which has limited the flexibility when designing SMA de-
vices. The cost problem can be solved by improving the manufacturing process and
mass producing, which are jobs for material scientists and manufacturing engineers.
D. Taniar et al. (Eds.): ICCSA 2010, Part III, LNCS 6018, pp. 498–510, 2010.
© Springer-Verlag Berlin Heidelberg 2010
Dynamic and Cyclic Response Simulation of Shape Memory Alloy Devices 499
As for the second problem, developing reliable and accurate models for SMA devices
is the most important task for researchers in the field of mechanics.
It is clear that, for most of the potential application fields of SMAs, such as SMA
actuators, SMA sensors and smart structures with SMAs, dynamic behavior simula-
tion is even more important than static behavior simulation. Since there have been
enough static models developed for a variety of SMAs, this paper is focused on mod-
eling the dynamic behavior of NiTi SMA devices. A SMA dynamic model is devel-
oped in this paper, whose feasibility is proved in both one-dimensional scale and
three-dimensional scales. The model which is used in this paper is based on the static
model developed by Brinson [2] and extended by Toi et al. [3]. The dynamic simula-
tion model is inspired by both Seelecke [4], who performed several one-dimensional
SMA dynamic simulations using Achenbach's model [5], and by Toi and Choi [6],
who have done several impact response simulations of SMA devices by considering a
strain rate effect in the previously formulated multi-axial constitutive equation [7].
The present modeling takes into consideration asymmetric tensile and compressive
deformation as well as multi-axial stress state which are not considered in the dynam-
ic analysis by Seelecke [4]. However, the strain rate effect, which is included in the
constitutive modeling by Toi and Choi [6], is neglected in the present modeling, as
the present study discusses relatively slow, vibration problems.
Afterwards, an improvement of SMA dynamic model considering cyclic effect [8-
11] is developed and discussed. The cyclic effect on constitutive behaviors of SMAs
in cyclic loading as well as cyclic deformation conditions, which is important in vi-
bration behaviors, was studied experimentally or analytically by Tobushi et al. [8],
Dolce and Cardone [9], Gall and Maier [10] and Sun and Rajapakse [11]. As for the
constitutive modeling for the cyclic effect, Tobushi et al. [8] and Sun and Rajapakse
[11] proposed the use of critical transformation stresses depending on the number of
cycles and the loading frequency, respectively. In the present study, the critical trans-
formation stresses and the maximum residual strains are assumed to depend on the
accumulated equivalent strain, which can be more conveniently used than the number
of cycles in the general, dynamic analysis. The frequency dependence is not consi-
dered in the present modeling, as the target of the present study is relatively slow
vibration analysis.
because of increasing Gibbs free energy when the SMA element is heated in the ab-
sence of applied stress.
Figure 2 shows the mechanical characteristics of SMAs. In Fig. 2(left), solid line
and dotted line show the stress-strain curve of superelastic effect and shape memory
effect respectively. Figure 2(right) is the critical stress-temperature diagram which
can be used to determine the phase concentration in different temperature and stress
settings. M and M indicate the starting and finishing temperature of phase trans-
formation from austenite to martensite when stress is 0; while A and A indicate the
starting and finishing temperature of phase transformation from martensite to auste-
nite in the same setting. The sloping solid line indicates the critical phase transforma-
tion stress in different temperature. The phase transformation from austenite to
martensite occurs when stress is increasing while it is between σ and σ . The
phase transformation from martensite to austenite occurs when stress is decreasing
Fig. 2. Equivalent stress vs. strain curve (left); Mechanical property of SMAs (right)
Dynamic and Cyclic Response Simulation of Shape Memory Alloy Devices 501
while it is between σ and σ . As examples, the vertical solid line shows a typical
superelastic effect route, while the vertical dotted line shows typical shape memory
effect route, which correspond to the solid line and the dotted line in Fig. 2(left)
respectively.
ξ ξ ξ ξ ξ . (7)
ξ ξ ξ . (8)
502 Y. Toi and J. He
ξ 1 cos π . (9)
ξ ξ . (10)
ξ ξ . (11)
For simulations of this section, the same SMA bar is used. Material constants of this
SMA bar are listed in Table 1.
Fig. 3. One-dimensional superelastic free vibration response simulation: Displacement vs. time
(upper), Stress vs. displacement (down)
The martensite phase transformation completes only in the first cycle and almost
stops after the 5th cycle.
side, which is caused by the higher requirement of stress for the phase transformation
in asymmetric setting.
σ D ε ξ Ω Tθ . (12)
Ω D R ε . (13)
θ D α . (14)
where α is the coefficient of thermal expansion. By using Euler's formula for nu-
merical time integration, we can derive the stress-strain relation equation as
∆σ D ∆ε ξ D ∆R ε ∆σ ∆t D ε ξ D R ε
(15)
ξ D R ε TD α TD .
or
∆σ X D ∆ε X Y . (16)
or
∆σ D ∆ε Θ . (17)
where
X I ξ D ∆R ε . (18)
Y ∆t D ε ξ D R ε ξ D R ε TD α TD . (19)
The forced vibration simulation in superelastic setting is described in this section. The
simulation object is listed in Fig. 5. It is a highway-bridge-like concrete structure with
506 Y. Toi and J. He
SMA dampers installed in their upper part (in black). The bottom of the structure is
fixed in all directions. Therefore not only the dynamic behavior but also the damping
capacity of SMAs will be shown in following subsection.
The constitutive models described in section 3 and section 4 are also applicable for
one-dimensional and three-dimensional cyclic dynamic simulation. After careful
analysis of experiment data from Tobushi et al. [8], Dolce and Cardone [9], Gall and
Maier [10] and Sun and Rajapakse [11], it is found that the cyclic behavior of SMA
devices can also be described by making a connection between the accumulative
strain and the martensite phase transformation start and finish temperature M and
508 Y. Toi and J. He
M , the reverse phase transformation start and finish temperature A and A , as well
as the maximum residual strain ε . The accumulative strain is expressed as follows
e ε dt . (21)
where e is the accumulative strain, ε is the equivalent strain rate. The connection
between the accumulative strain e and the martensite phase transformation start and
finish temperature M and M , as well as the reverse phase transformation start and
finish temperature A and A are expressed as
M e M 0 α 1 exp β e . (22)
M e M 0 α 1 exp β e . (23)
A e A 0 α 1 exp β e . (24)
A e A 0 α 1 exp β e . (25)
ε e ε 0 α 1 exp β e . (26)
ε e ε 0 α 1 exp β e . (27)
where ε 0 is called the initial maximum residual strain; α , α , β and β are the
coefficients which have to be determined by experiment data. By substituting the equa-
tions above into the martensite phase transformation equation which was described in
section 2, we will be able to obtain the cyclic dynamic behavior of SMA devices.
According to the experimental data provided by Tobushi el al. [8], the simulation
object (SMA bar) is similar to the one which is described in the section 2. Unloading
starts as soon as the strain reaches 9%. Detailed simulation results are listed in Fig. 7.
The experimental data shows the stress and strain relationship during No. 1, 2, 5,
10, 20, 50, 100 cycles. The initial maximum residual strain is set at 3.5%, the
temperature is set at 353K. The experimental data shows that critical stresses of start
Dynamic and Cyclic Response Simulation of Shape Memory Alloy Devices 509
and finish martensite phase transformation as well as the reverse transformation are
getting lower and lower, while the maximum residual strain is getting larger and
larger. The simulation result which is presented in lower graph of Fig. 7 shows a
similar behavior. The critical stresses and the maximum residual strain are changing
faster at the beginning then slowing down, because of the term (1-exp(-X)) in
coefficients revolution function. Only the cycles No. 1, 2, 5, 10, 20, 50 and 100 are
presented in Fig. 7 to show a better comparison with the experimental data.
Fig. 7. Experimental data and simulation results under cyclic loading: simulation data (upper),
experimental data (down)
6 Concluding Remarks
been done by introducing cyclic effect and asymmetric behavior of SMA devices into
the phase transformation mechanism. According to the simulations performed in this
paper, several conclusions can be obtained as follows:
1. The validity of the dynamic response simulation model as well as the improved
phase transformation mechanism to describe SMAs' superelastic and quasi-plastic
behavior has been illustrated.
2. An improved SMA model by considering the cyclic effect has been developed and
tested.
3. The potential of SMAs being used as damping devices has been demonstrated.
The future works will be focused on enhancing the simulation model's stability,
especially for high frequency vibration simulations in three-dimensional scales.
References
1. Sugiyama, Y., Hirai, S.: Crawling and jumping by a deformable robot. International Jour-
nal of Robotics Research 25(5), 603 (2006)
2. Brinson, L.: One-dimensional constitutive behavior of shape memory alloys: thermome-
chanical derivation with non-constant material functions and redefined martensite internal
variable. Journal of Intelligent Material Systems and Structures 4(2), 229 (1993)
3. Toi, Y., Lee, J., Taya, M.: Finite element analysis of superelastic, large deformation beha-
vior of shape memory alloy helical springs. Computers and Structures 82(20-21), 1685–
1693 (2004)
4. Seelecke, S.: Modeling the dynamic behavior of shape memory alloys. International Jour-
nal of Non-Linear Mechanics 37(8), 1363–1374 (2002)
5. Achenbach, M.: A model for an alloy with shape memory. International Journal of Plastici-
ty 5(4), 371–395 (1989)
6. Toi, Y., Choi, D.: Constitutive Modeling of Porous Shape Memory Alloys Considering
Strain Rate Effect. Journal of Computational Science and Technology 2(4), 511–522
(2008)
7. Toi, Y., Choi, D.: Computational Modeling of Superelastic Behaviors of Shape Memory
Alloy Devices under Combined Stresses. Journal of Computational Science and Technolo-
gy 2(4), 535–546 (2008)
8. Tobushi, H., Iwanaga, H., Tanaka, K., Hori, T., Sawada, T.: Stress-strain-temperature rela-
tionships of TiNi shape memory alloy suitable for thermomechanical cycling. JSME Inter-
national Journal. Series 1, Solid Mechanics, Strenght of Materials 35(3), 271–277 (1992)
9. Dolce, M., Cardone, D.: Mechanical behaviour of shape memory alloys for seismic appli-
cations 2. Austenite NiTi wires subjected to tension. International Journal of Mechanical
Sciences 43(11), 2657–2677 (2001)
10. Gall, K., Maier, H.: Cyclic deformation mechanisms in precipitated NiTi shape memory
alloys. Acta Materialia 50(18), 4643–4657 (2002)
11. Sun, S., Rajapakse, R.: Simulation of pseudoelastic behaviour of SMA under cyclic load-
ing. Computational Materials Science 28(3-4), 663–674 (2003)
Adaptive Fuzzy Filter for Speech Enhancement
Abstract. In this paper an adaptive fuzzy filter, based on fuzzy system, is pro-
posed for speech signal enhancement and automatic speech recognition accu-
racy. In the past two decades the basic wavelet thresholding-algorithm has been
widely studied and is common applied to filter noise. In the proposed system
adaptive wavelet thresholds are generated and controlled by fuzzy rules con-
cerning the presence of speech in noise. First an amplified voice activity detec-
tor is designed to improve performance on SNR lower than 5dB. Then an
adaptive threshold decision module based on fuzzy inference system is pro-
posed. In this fuzzy inference system overall relations between speech and
noise are summarized into seven fuzzy rules and four linguistic variables, which
are used to detect the state of signals. The adaptive threshold and membership
functions are optimally obtained by particle swarm optimization algorithm so
the SNR of the filter output for training signal data can be maximized. Experi-
mental results reveal that our proposed system effectively increases the SNR
and the recognition rate of speech.
1 Introduction
Speech enhancement remains an important problem within the field of speech and
signal processing, such as mobile communication, speech recognition and hearing
aids [1]. Speech enhancement algorithms have undergone numerous studies through-
out the last two decades yet issues such as distortions to the original speech signal and
residual noise, sometimes in the form of musical tones created by the enhancement
algorithms, remain unsolved [2, 3, 4].
The main objective of speech enhancement is to reduce corrupting noise while pre-
serving as much original clean speech quality as possible. The objectives of speech
enhancement algorithms can be classified into increasing the recognition rate and the
signal-to-noise ratio (SNR) of the distorted speech [5, 6]. These can be roughly classi-
fied as digital signal processing and statistical analysis. The former usually remove an
estimate of the distortion from the noisy features, e.g. spectral subtractive [7]. The
latter utilizes a statistical model to describe predictable structures and expected pat-
terns in the signal process [8, 9].
D. Taniar et al. (Eds.): ICCSA 2010, Part III, LNCS 6018, pp. 511–525, 2010.
© Springer-Verlag Berlin Heidelberg 2010
512 C.-C. Yao and M.-H. Tsai
The principle of a spectral subtractive algorithm is to obtain the best possible esti-
mations of the short time spectra of a speech signal from a given contaminated
speech. The main attractions of a spectral subtractive algorithm are: (1) Easy and
simple implementation. (2) High flexibility against subtraction parameter's variation.
On the other hand, spectral subtractive algorithms suffer a number of deficiencies
including an imprecise estimation of signal and noise parameters and a mismatch of
the probability distribution models of speech or noise [10].
The wavelet thresholding method is often used to shrink the signal and remove the
noise. Using an adaptive thresholding algorithm could provide better performance
[11, 12]. In this paper we propose an adaptive fuzzy filter, based on wavelet thresh-
olding, to improve performance in speech signal enhancement and automatic speech
recognition accuracy. In this developed adaptive fuzzy filter the signals are classified
into speech and non-speech segments by an amplified voice activity detector. This
amplified voice activity detector employs the full wavelet packet transform (instead of
the perceptual wavelet packet transform) to decompose the input speech signal into
critical sub-band signals [13]. In each critical sub-band signal the mask construction is
obtained by smoothing the Teager energy operator and the entropy of corresponding
wavelet coefficients. Thus, the amplified voice activity detector enhances the ability
of distinction when the SNR is lower than 5dB.
Although some algorithms have been proposed to generate the adaptive threshold
the model between speech and noise is still not clear and the rules judging the exis-
tence of noise is usually expressed in an ambiguous style. A new-type of speech
enhancement system based on fuzzy rules is proposed here. In the design overall
relations between speech and noise are summarized into seven fuzzy rules. Adaptive
thresholds are then inferred by those fuzzy rules. Moreover, this filter is optimally
designed by particle swarm optimization, so that the SNR of the filter output is maxi-
mum; membership functions to represent the rules are not required to be obtained
beforehand.
As is well know, using critical parameters is helpful on detecting the presence of
speech in noisy environment. Some researchers have focused on energy, zero crossing
rate and time duration to find the boundary between the word signal and background
noise. Other parameters have been proposed such as linear prediction coefficient,
linear prediction error energy, pitch information and time-frequency parameter. How-
ever these parameters cannot yet be well adapted to variable-level background noise
even if more complex decision strategies are used. In this paper four parameters used
as linguistic variables are incorporated in order to detect the presence of speech in
noisy environment. The four parameters are energy, zero crossing rate, entropy and
standard deviation. Examples of processing actual speech with eight types of noises
are shown to demonstrate the high performance of this filter.
The remainder of this paper is organized as follows. In Section 2 we review the ba-
sic concept of wavelet transform and wavelet packet decomposition. In Section 3 the
scheme of the speech enhancement system is introduced. In Sections 4 we introduce
the fuzzy inference model for speech enhancement. In Section 5 a performance
evaluation of the proposed system is presented and comparisons with other protocols
are made. Our conclusions are made in Section 6.
Adaptive Fuzzy Filter for Speech Enhancement 513
2 Basic Concepts
∞
ψ 2j +p1+1 (k ) = ∑ g[n]ψ
n = −∞
p
j ( k − 2 j n) (3)
where h[n] and g[n] denote the low-pass and high-pass filters, respectively. ψ (n) is
the wavelet function and parameters j and p are the number of decomposition levels
and nodes, respectively.
In 1995, Kennedy and Eberhart introduced the particle swarm optimization algorithm
(PSO) into the field of social and cognitive behavior [14]. Traditionally the main prob-
lem in designing a neural fuzzy system is training the parameters. Backpropagation
514 C.-C. Yao and M.-H. Tsai
training is commonly adopted to solve this problem. However the steepest descent
approach, commonly used in backpropagation training to minimize the error function,
may reach the local minima very quickly and never find the global solution. Accord-
ingly, a new optimization algorithm, called particle swarm optimization (PSO), appears
to provide better performance than the backpropagation algorithm.
Like other population-based optimization approaches PSO is initialized with a
swarm of random solutions, each swarm consists of many particles. Each particle is
r
characterized by its current position xi = [ x i1 , x i2 , L , xiD ] and current velocity
r
vi = [v i1 , vi2 , L , viD ] , where D stands for the dimensions of the solution space. In the
PSO the trajectory of each particle in the search space is adjusted by dynamically
altering the velocity of each particle. Then the particles rapidly search the solution
space using the moving velocity of each particle. Each of these particle positions is
scored to obtain a fitness value based on how to define the solution of the problem.
During the evolutionary process the velocity and position of particle i are updated as
r r r r
vi = ω × vi + ϕ1 × rand1 ( ) × (NBesti − xi ) + ϕ 2 × rand 2 ( ) × (GBesti − xi ) (4)
r r r
x i = xi + vi (5)
where ω is the inertia weight, ϕ1 and ϕ 2 are the acceleration coefficients, respec-
tively. The second term in Eq. (4), called the cognitive component, reflects the experi-
ence of a particle since it is dependent on the best position of the respective particle.
The third term is referred to as the social component and contains the information of a
social group due to the dependence on the neighborhood best position. The random
numbers rand1 ( ) and rand 2 ( ) are chosen from the interval U(0,1). In Eq. (4), GBest
is the position with the best fitness found so far for the ith particle, and NBest is the
r r
best position in the neighborhood. The term vi is limited to the range ± v max . If the
velocity violates this limit, then it is set to the actual limit. Changing the velocity en-
ables each particle to search around its individual best position and global best posi-
tion. After initialization of the positions and velocities of the particles update equations
are applied to every particle in each iteration until a stopping criterion is fulfilled.
where T [ y (n)] is called the TEO coefficient of y(n). However a Teager energy opera-
tor is insensitive when the signal-to-noise ratio is low. As an example, when the SNR
is lower than 0dB the difference between noisy energy and speech energy is not obvi-
ous and the performance of TEO is not satisfactory. Fig. 2 shows an example of using
TEO to distinguish speech from noise signals contaminated by -5dB Gaussian white
noise. To overcome this problem, in our proposed module, a Teager energy operator
is cooperated with entropy to improve the ability of distinction. The formula of en-
tropy is shown as follow:
E ( S ) = ∑ p ( si ) log p ( si ) (7)
i
Entropy represents the degree of variation. When the noise model is stationary or
slight non-stationary the corresponding entropy is kept stable or slight changed. Fig. 3
shows the comparison of using TEO and entropy on VAD to distinguish signals con-
taminated by -5dB Gaussian white noise.
The proposed VAD algorithm computes signals wm ( x i ) that have been produced
by the wavelet package transform on each input frame to produce 2J sub-band wavelet
packets, where J is the number of levels for the wavelet packet decomposition tree
and 1 ≤ m ≤ 2 J . Then, a set of T [ wm ( xi )] can be derived from Eq. (6). The scheme of
the voice activity detection is designed as follows:
⎧ 0, if var(T [ wm ( xi )]) < λ J or var( E[ wm ( xi )]) < ς (8)
Vm ( x i ) = ⎨
⎩ xi , otherwise
where ς is user-defined and λ J = σ J 2 log( N J ) , proposed by Johnston and Silverman
[16]. var(⋅) denotes the variance.
516 C.-C. Yao and M.-H. Tsai
Fig. 3. The comparison of VAD based on TEO and entropy with -5db Gaussian white noise
generated by the fuzzy inference system and the noise components were removed by
thresholding the wavelet coefficients.
Let s(t) be clean speech with finite length and n(t) be noise, then contaminated
speech y(t) can be expressed as
y(t ) = s(t ) + n(t ) (9)
If W denotes the wavelet transform matrix, Eq. (9) can be written in the wavelet
domain as
Y (t ) = S (t ) + N (t ) (10)
Sˆ (t ) = FT (Y , T ) (11)
where FT (Y , T ) denotes the thresholding function and T is the threshold. The standard
thresholding function includes the soft thresholding function which is defined as
⎧sign(Y )( Y − T ), Y ≥T (12)
FTS (Y , T ) = ⎨
⎩ 0, Y <T
In this paper an adaptive threshold decision module based on a fuzzy inference system
was proposed. In the fuzzy inference system overall relations between speech and
noise are summarized to seven fuzzy rules. Moreover, four linguistic variables are
used to detect the state of signals. They are energy, zero crossing rate, standard devia-
tion and average residual. The reasons for adopting those four parameters are listed as
follows.
-Energy:
Energy is an effective factor on measuring the degree of noise when the SNR is big-
ger than 0dB. Studies have shown that energy provides excellent performance on
voice activity detection. However, rapid variation of energy in the speech model
causes implementation difficulties.
-Zero crossing rate:
Zero crossing rate (ZCR) is a basic acoustic feature that can be computed easily. It is
equal to the number of zero crossing of the waveform within a given frame. In general
the ZCR of both unvoiced sounds and environment noise are larger than voiced
sounds. ZCR is often used in conjunction with volume for end-point detection. It is
hard to distinguish unvoiced sounds from environment noise by using ZCR alone
since they have similar ZCR values.
-Standard deviation:
Standard deviation is a measure of how wide any given numbers are spread. It is use-
ful in comparing sets of data, which may have the same mean but different range. In a
given frame standard deviation is helpful to distinguish the signal's mode. The
standard deviation of a stable noise is different to voice sound. The observation of
standard deviation is useful in the threshold decision level.
518 C.-C. Yao and M.-H. Tsai
-Average residual:
The statistical property of average residual is similar to standard deviation. In this
paper standard deviation is used to detect the variation within a frame but average
residual is used to compare the difference between frames. By calculating the average
residual across several frames the signal's mode can be distinguished, whether the
variation is temporal or not, in a very short time interval. In this paper the average
residual is calculated based on the standard deviations of previous frame, current
frame and next frame.
In the fuzzy inference system, Eq. (13) is used to obtain the optimal solution.
T (k + 1) = T (k ) + α (k )THS (k ) (13)
Here, α(k) is a step length for threshold’s variation and THS(k) is the kth threshold.
α(k) and THS(k) are decided by above four parameters. First the range of threshold is
decided by energy and zero crossing rate. Studies have shown that energy is an effec-
tive factor in measuring the degree of noise greater than 0dB. However when the
threshold is only decided by energy any rapid variation of energy causes difficulty
and irrational implementation. On top of which, most mechanisms generally use an 8k
sample rate to record the speech signal and, when the frequency of the noise is bigger
than the sample rate, energy is ineffective. To overcome above problem zero crossing
rate is introduced to cooperate with energy on deciding the threshold.
As described in previous statements the ZCR of unvoiced sounds and environment
noise are larger than the ZCR of voiced sounds. In many studies a zero crossing rate is
used on voice activity detection. The signal's mode can be easily distinguished by
using zero crossing rate . When the noise is high frequency and the speech is low
frequency the zero crossing rate can be taken as the auxiliary and be cooperated with
energy. When the frequency of noise is low the threshold decision is highly dependent
on energy. As the frequency of noise becomes higher zero crossing rate is taken into
consideration. In short the importance of zero crossing rate on the threshold decision
is proportional to the frequency of noise.
Standard deviation and average residual are then introduced to control the degree
of threshold variation. The degree of variation on signal's modes between two adja-
cent frames is reflected in the difference between standard deviations. As the differ-
ence between standard deviations increases so the amount of threshold variation in-
creases. However this does not apply when speech is contaminated by instantaneous
impulse signal. An instantaneous impulse signal in a given frame will usually increase
the standard deviation and often causes a violent vibration of threshold. To avoid this
problem, the average residual of any adjacent three frames is calculated to smooth the
threshold's variation. The average residual is calculated from standard deviations of
any adjacent three frames and can effectively reduce the impact of instantaneous
impulse signals. Should the value of the average residual increase there is a high prob-
ability that it is neither temporal nor instantaneous.
The threshold can be set by the local characteristics of the input signals but, as previ-
ously mentioned, there are many ambiguous cases where constructing the noise model
Adaptive Fuzzy Filter for Speech Enhancement 519
from the contaminated signal is hard to achieve. In the proposed speech enhancement
system a fuzzy inference system is adopted to overcome these problems and the adap-
tive threshold is generated in order to suppress noise. To begin with various models of
noises and the relation between noise and speech signals are summarized into several
cases. The fuzzy sets of linguistic variables and their corresponding linguistic terms
are then defined according to the analysis. The relations between linguistic variables
and their fuzzy sets are shown in Table 1.
After eliminating the cases where the noise is impossible to distinguish from
speech signal, seven fuzzy rules are concluded, listed as follows.
Rule 1: If Energy is High and Zero Crossing Rate is Low, then Threshold is Medium.
Energy is an effective factor on measuring the degree of noise when the SNR is big-
ger than 0dB. In most cases speech signal has the property of high energy as opposed
to noise. In attempt to retain speech and suppress noise, the threshold can be set lower
than the speech signal. This rule is applicable in places like airports or streets with
heavy traffic. In this model the zero crossing rate is classified to low when the fre-
quency is between 50Hz and 5 KHz.
Rule 2: If Energy is High and Zero Crossing Rate is High, then Threshold is High.
As the zero crossing rate becomes higher, the probability of noise existence increases.
In rule 2 the probability of the existence of noise is higher than in rule 1, so that the
threshold can be set higher than rule 1. This rule is applicable when the energy is
large and the frequency is high such as noise generated by high frequency electric
products or machines. In this model the zero crossing rate is classified to high when
the frequency is bigger than 5K Hz.
Rule 3: If Energy is Low and Zero Crossing Rate is Low, then Threshold is Low.
Low zero crossing rate usually means the probability of noise existence is not high. In
most case the max energy is detected from the speech signal. Therefore, the threshold
…
is set to low when the energy of contaminated speech signal is low. The type of noises
is generated such as by the low frequency electric products (indoor fan, , etc).
Rule 4: If Energy is Low and Zero Crossing Rate is High, then Threshold is Medium.
On the basis of rule 4, high zero crossing rate denotes a high probability of noise. The
threshold should be bigger than in rule 3 but not exceed the energy of speech signal.
This rule is applicable on the environments such as the electromagnetic wave.
520 C.-C. Yao and M.-H. Tsai
Rule 5: If Standard deviation is Low and Average residual is Low, then Step length is
Low.
This rule is applicable when the model of noise is stable, whether in a very short time
interval or in a longer time interval. The property of this noise is that the standard
deviation is small and the average residual calculated from three successive frame
blocks' standard deviations is also small. Since the standard deviation is small, the
low threshold is necessary. The waveform under this situation is shown in Fig. 4 In
Fig. 4, vertical axis represents the amplitude and horizontal axis represents the time.
Besides, in Fig. 4, each frame is bounded by the rectangle. The standard deviation of
each frame is small as the waveform within each frame is stable. The average residual
of the successive three frame's standard deviation is also small since the standard
deviations are all small.
Fig. 4. The property of the noise is low standard deviation and low average residual
Rule 6: If Standard deviation is High and Average residual is Low, then Step length is
Medium.
This rule is applicable when noise is stable but variant in a very short time interval. The
property of this noise is that the standard deviation is high but the residual that is calcu-
lated from current and neighbor frame's standard deviations is small. In rule 6 high
standard deviation means the amplitude of noise is higher than the noise in rule 5.
Hence, the threshold should be set higher than in rule 5. Under the situation with multi-
ple sources of noise, this rule can bring the function into full play. The waveform under
this situation is shown in Fig. 5. In Fig. 5, a big variation of the waveform within the
middle frame causes the standard deviation to be big. The average residual of the suc-
cessive three frame's standard deviation is small since the standard deviations are all big.
Fig. 5. The property of the noise is high standard deviation and low average residual
Adaptive Fuzzy Filter for Speech Enhancement 521
Rule 7: If Standard deviation is High and Average residual is High, then Step length is
High.
This rule is applicable when noise is unstable, whether in a very short time interval or
in a long time interval. In rule 7 the current frame's standard deviation is high and is
higher than neighbor frames so that the average residual calculated from current and
neighbor frame's standard deviations is high. The waveform under this situation is
shown in Fig. 6.
Fig. 6. The property of the noise is high standard deviation and high average residual
The next step is to design this function “Threshold” and “Step length” for Energy,
Zero crossing rate, Standard deviation and Average residual. In most fuzzy systems
membership function is obtained by fuzzy approximate reasoning using the member-
ship function of the Low and the High of Energy, Zero crossing rate, Standard devia-
tion, Average residual and its relationship to Threshold and Step length. In this paper
the Gaussian type and the sigmoid function are used to approximate the membership
function. The benefit is efficient calculation. In the field of signal processing the goal
is to obtain the optimum filter, in which the output is as close to the original signal as
possible; thus some optimization is required in designing this filter. The particle
swarm optimization algorithm (PSO) is applied to optimize the parameters in these
nonlinear function. As we know, particle swarm optimization algorithm provides
better performance on finding the global solution while optimizing the overall struc-
ture. Experimental results demonstrate that our proposed system provides excellent
performance on speech enhancement.
5 Experimental Results
This section demonstrates the effectiveness of our proposed system. The experimental
results that pertain to our proposed speech enhancement system are compared to spec-
trum subtraction and signal subspace approach.
First, tests for the enhancement and recognition task were carried out on the Aurora
2 databasen [17]. In the database the speech files include recording of the 10 English
digits and 26 English letters spoken by males and females. The Aurora 2 database
does not only provide the uncontaminated speech of 4004 sentences it also includes
the speech of 48048 sentences contaminated with eight types of noise (Subway, Bab-
ble, Car, Exhibition, Restaurant, Street, Airport and Station) and are mixed with -5, 0,
5, 10, 15 and 20dB.
522 C.-C. Yao and M.-H. Tsai
Fig. 7. (a) Speech signal contaminated by 10dB car noise. (b) Enhancement signals denoised by
our proposed system.
(a)
(b)
Fig. 8. (a) The frequency of speech signal contaminated by 10dB car noise. (b) The frequency
of enhancement signals denoised by our proposed system.
Adaptive Fuzzy Filter for Speech Enhancement 523
The evaluation of the proposed system includes two parts. First the signal-to-noise
ratios on before and after enhanced signals are evaluated. Next applying those en-
hanced speech signals into the HTK classifier compares the recognition rates of our
proposed system with other methods. The experimental results are listed as follows.
Fig. 7 shows the waveform before and after signal denoising by our proposed sys-
tem. The signal is mixed with 10dB car noise. The corresponding time-frequency
diagrams are reveals in Fig. 8.
The comparisons of SNR are included to demonstrate the effectiveness of our pro-
posed system. Tables 2 and 3 show the SNR before and after signal denoising by
our proposed system. The experimental results show that our proposed system can
effectively remove noise.
Finally, the recognition rates of our proposed system with other methods are listed
in Tables 4 - 6. Table 4 shows the rate of recognizing the signal without denoising.
Tables 5 and 6 show the rate of recognizing the signal after denoising by our
Table 3. The SNR of the signals after enhancing by our proposed algorithm
Recognition rate %
Method Subway Babble Car Exhib. Rest. Street Airport Station AVE.
subspace 99.26 99.06 99.14 99.51 99.26 99.24 99.14 99.51 99.26
clean
Our 99.10 98.90 99.16 99.31 99.36 99.26 99.28 99.57 99.24
subspace 99.28 97.40 98.09 97.81 97.97 97.67 97.55 97.55 97.79
20dB
Our 98.10 97.96 98.56 97.57 98.6 97.93 98.26 98.84 98.22
subspace 95.52 92.87 96.69 94.48 94.32 93.23 94.18 94.18 94.43
15dB
Our 95.65 94.20 96.10 95.43 95.24 94.85 95.6 95.91 95.37
subspace 87.69 81.89 90.3 85.34 84.07 83.92 84.46 84.46 85.26
10dB
Our 87.90 83.60 82.9 87.8 88.3 84.95 84.9 86.7 85.88
subspace 74.70 64.06 74.59 62.94 65.18 67.02 67.52 67.52 67.94
5dB
Our 71.23 65.40 72.3 67.1 66.8 64.25 69.82 63.9 67.6
subspace 52.16 39.84 45.45 38.29 41.23 39.02 43.66 43.66 42.91
0dB
Our 53.33 40.21 44.33 39.1 42.3 41.36 45.1 43.89 43.70
subspace 28.31 21.49 20.34 16.11 17.93 16.84 21.47 21.47 20.49
-5dB
Our 30.20 24.60 22.3 16.9 18.3 17.77 20.6 19.6 21.28
Method Subway Babble Car Exhib. Rest. Street Airport Station AVE.
20dB spectrum 84.81 89.99 89.47 80.61 91.18 91.77 91.43 97.55 93.15
15dB spectrum 77.22 79.11 81.76 67.78 79.34 85.96 84.81 94.18 90.56
10dB spectrum 65.71 55.01 66.86 47.64 56.96 67.84 64.9 84.46 77.85
5dB spectrum 45.52 27.27 38.72 25.09 29.52 43.5 36.08 67.52 47.53
proposed method, the signal subspace approach and spectrum subtraction method,
respectively. Experimental results demonstrate that our proposed method outperforms
other methods.
6 Discussions
In this paper a novel speech enhancement system controlled by fuzzy rules is pro-
posed for removing additive noise in a contaminated speech signal. In this speech
enhancement system the rules about how to set the system's parameters depending on
the local characteristics of signals are expressed in an ambiguous style and the system
is designed optimally by using partial swarm optimization, so that the mean square
error of the system's output can be minimized. In the system four linguistic variables
and seven fuzzy rules are used to infer the adaptive threshold. Different kinds of
noises can be effectively represented and distinguished by the four linguistic vari-
ables. Experimental results demonstrate that our proposed method effectively re-
moves noise and outperform other methods.
Concerning further researches. Searching for more critical parameters to identify
the noise and finding more rules to add to this system is under considered. Although
some studies have demonstrated algorithms can automatically generate inference
rules, inference time increases as fuzzy rules increase. Finding the balance between
precision and time complexity is an open issue.
Adaptive Fuzzy Filter for Speech Enhancement 525
References
[1] Takahashi, Y., Takatani, T., Osako, K., Saruwatari, H., Shikano, K.: Blind spatial sub-
traction array for speech enhancement in noisy environment. IEEE Trans. on Audio,
Speech, and Language Processing 17(4), 650–664 (2009)
[2] Abramson, A., Cohen, I.: Simultaneous detection and estimation approach for speech
enhancement. IEEE Trans. on Audio, Speech and Language Processing 15(8), 2348–
2359 (2007)
[3] Wu, S.-N., Wang, J.-S.: An adaptive recurrent neuro-fuzzy filter for noisy speech en-
hancement. In: International Joint Conference on Neural Networks, pp. 3083–3088 (2004)
[4] Er, M.J., Li, Z.: Adaptive noise cancellation using online self-enhanced fuzzy filters with
applications to multimedia processing. In: Intelligent Multimedia Processing with Soft
Computing. Springer, Berlin (2006)
[5] Ma, N., Bouchard, M., Goubran, R.A.: Speech enhancement using a masking threshold
constrained kalman filter and its heuristic implementations. IEEE Trans. on Audio,
Speech, and Language Processing 14(1), 19–32 (2006)
[6] Shao, Y., Chang, C.-H.: A generalized time–frequency subtraction method for robust
speech enhancement based on wavelet filter banks modeling of human auditory system.
IEEE Trans. on Systems, Man, and Cybernetic–Part B: Cybernetics 37(4), 877–889 (2007)
[7] Furuya, K., Kataoka, A.: Robust speech dereverberation using multichannel blind decon-
volution with spectral subtraction. IEEE Trans. on Audio, Speech and Language Process-
ing 15(5), 1579–1591 (2007)
[8] Hao, J., Attias, H., Nagarajan, S., Lee, T.-W., Sejnowski, T.J.: Speech enhancement,
gain, and noise spectrum adaptation using approximate bayesian estimation. IEEE Trans.
on Audio, Speech and Language Processing 17(1), 24–37 (2009)
[9] Chang, J.-H., Jo, Q.-H., Kim, D.-K., Kim, N.-S.: Global soft decision employing support
vector machine for speech enhancement. IEEE Signal Processing Letters 16(1), 57–60
(2007)
[10] Hasan, M.K., Salahuddin, S., Khan, M.R.: A modified a priori SNR for speech enhance-
ment using spectral subtraction rules. IEEE Signal Processing Letters 11(4), 450–453
(2004)
[11] Tahmasbi, R., Rezaei, S.: A soft voice activity detection using GARCH filter and vari-
ance gamma distribution. IEEE Trans. on Audio, Speech, and Language Process-
ing 15(4), 1129–1134 (2007)
[12] Davis, A., Nordholm, S., Togneri, R.: Statistical voice activity detection using low-
variance spectrum estimation and an adaptive threshold. IEEE Trans. on Audio, Speech,
and Language Processing 14(2), 412–424 (2006)
[13] Johnson, M.T., Yuan, X., Ren, Y.: Speech signal enhancement through adaptive wavelet
thresholding. Speech Communication 49, 123–133 (2007)
[14] Kennedy, J., Eberhart, R.: Particle swarm optimization. In: Proc. IEEE Int. Conf. Neural
Network, pp. 1942–1948 (1995)
[15] Chen, S.-H., Wu, H.-T., Chang, Y., Truong, T.K.: Robust voice activity detection using
perceptual wavelet-packet transform and Teager energy operator. Pattern Recognition
Letters 28, 1327–1332 (2007)
[16] Johnson, I.M., Silverman, B.W.: Wavelet threshold estimators for data with correlated
noise. J. Roy. Statist. Soc., Ser. B 59, 319–351 (1997)
[17] Hirsch, H.G., Pearce, D.: The aurora experimental framework for the performance
evaluations of speech recognition systems under noisy conditions. In: ISCA ITRW ASR
2000: Challenges for the New Millenium, Paris, France (September 2000)
[18] Speech Vision and Robotics Group, http://htk.eng.cam.ac.uk
Risk Prediction for Postoperative Morbidity of
Endovascular Aneurysm Repair Using Ensemble Model
1
Department of Information Management, National Taipei College of Nursing
No. 365, Min-Ten Road 11257, Taipei, Taiwan, ROC
[email protected]
2
Department of Computer Science and Information Engineering, Tamkang University
No. 151, Ying-Chuan Road, 25137 Tamsui, Taipei County, Taiwan, ROC
1 Introduction
Cardiac surgery is a complex surgical operation that is indicated for patients with
severe insufficiency in cardiac function. Major cardiac surgical interventions include
coronary artery bypass grafting (CABG), repair of congenital heart defects, surgical
treatment of atrial fibrillation, heart transplantation, repair or replacement of heart
valves, aortic surgery, aneurysm repair or a combination of these surgical procedures.
During the operation and the postoperative stay at the ICU and nursing ward, there is
considerable morbidity for cardiac surgery patients with postoperative complications,
which results in increased hospital mortality and postoperative morbidity. Many
prediction models for cardiac surgical outcome apply logistic or multivariable regres-
sion to assess preoperative risk [1-3]. EuroSCORE is a predominant model [4]. Most
of the risk prediction models in current use were derived for patients undergoing open
D. Taniar et al. (Eds.): ICCSA 2010, Part III, LNCS 6018, pp. 526–540, 2010.
© Springer-Verlag Berlin Heidelberg 2010
Risk Prediction for Postoperative Morbidity of EVAR Using Ensemble Model 527
abdominal aortic aneurysm (AAA) repair and appear to lack utility when applied to
EVAR patients. EVAR is an advanced minimally invasive surgical technology that
helps reduce patients’ recovery time as well as postoperative mortality and morbidity;
it is especially helpful in the treatment of patients judged to be high surgical risk for
conventional surgery. EVAR benefits patients with medical co-morbidities, and post-
operative complications highly significantly influence longer-term postoperative out-
comes in EVAR patients.
The use of machine learning models has become widely accepted in medical appli-
cations. Various machine learning models including BNs, NNs, and SVMs have been
tested in a wide variety of clinical and medical applications [5]. Soft-computing, in-
cluding fuzzy set and rough set techniques that work well in descriptive data mining,
is also a promising technique. BN is a probability-based inference model that has a
wide range of applications and is increasingly used medically as a prediction and
knowledge representation modeling technique. Verduijn, M., et al. [6, 7] presented
the prognostic BN as a new type of prognostic model that builds on the BN method-
ology and implements a dynamic, process-oriented view of cardiac surgical progno-
sis. Lin and Haug [8] proposed BN suitable for exploiting missing clinical data for
medical prediction. NNs have featured in a wide range of medical applications, often
with promising results. Rowan et al. [9] used NNs to stratify the length of stay of
cardiac patients based on preoperative and initial postoperative factors. Eom et al.
[10] developed a classifier ensemble-based, including NNs, DTs, and SVMs, clinical
decision support system for predicting cardiovascular disease level. SVMs have been
successfully used in a wide variety of medical applications. Polat and Güne [11] used
a least square support vector machine to assist breast cancer diagnosis. Babaog˘lu et
al. [12] first used principle component analysis method to reduce data features, and
acquired an optimum support vector machine model for the diagnosis of coronary
artery disease. Choi [13] proposed the detection of valvular heart disorder (VHD) by
wavelet packet decomposition and SVM techniques.
Traditional medical data modeling tends to use statistical techniques to provide re-
ports. However, most diagnoses and treatments in medical decision making are im-
precise and are accompanied by errors. RS is able to circumvent this limitation with
its ability to handle imprecise and uncertain data [14]. Theoretically, statistical
clinical models use predetermined hypotheses and a priori assumptions in initial hy-
pothesis test design. These models also usually require a perfectly random statistical
distribution; furthermore, statistical formulae or equations usually prove poor in char-
acterizing medical data and solving for the characteristics or relationships of that data.
The RSs approach takes advantage of a correct proven philosophy to work with medi-
cal data without a strong a priori reasoning. As an example, Mitra et al. [15] imple-
mented a rule-based RS decision system for the development of a disease inference
engine for ECG classification. Podraza et al. [16] presented a complex data analysis
and decision support system for medical staff based on rough set theory. Wakulicz-
Deja and Paszek [17] implemented an application of RS theory to decision making for
diagnosing mitochondrial encephalomyopathies in children.
This study describes the development of an informative ensemble prediction model
consisting of BNs, NNs and SVMs and argument using an RS model for the predic-
tion of postoperative morbidity and the description of causal effects between preop-
erative variables and complication outcomes in EVAR patients. The development of
528 N.-C. Hsieh, C.-H. Chan, and H.-C. Tsai
Section 2 of this paper begins with an overview of study background and research
materials in general. Section 3 describes the methods and procedures used in this
study. Section 3 also empirically tests the proposed method using an EVAR clinical
dataset. Section 4 discusses the experimental findings and offers observations about
practical applications and directions for future research.
Most studies dealing with cardiac surgery prediction models have applied logistic or
multivariable regression to assess the preoperative risk. Few studies have utilized
machine learning algorithms i.e., decision trees, Bayesian networks or neural net-
works in analyzing clinical data. These state-of-the-art machine learning algorithms
are often informative and can represent more knowledge in the clinical data. Gener-
ally, clinical datasets involve numerical variables. To satisfy the requirements of
prediction models, the employment of a discretization approach is necessary. Discre-
tization is defined as a process that divides numerical values into states of discrete
categorical values, leading to informative expressed categorical values. For example,
the CART model originally was not designed to handle numerical attributes. During
the construction of a CART model, numerical attributes were divided into discrete
categorical values. Aside from CART, discretization techniques were frequently
adopted in other popular learning paradigms, such as C4.5, BNs, NNs, and genetic
algorithms.
This study employed entropy/MDL to discretize numerical domains into intervals.
Incorporation investigates fuzzy discretization; the discretized value is characterized
by a triple of interval (i,m,a), where i denotes interval name, m denotes membership
value of the value, and a denotes affinity of the value. The steps for automatically
finding fuzzy sets from a given dataset are described herein. Assuming that the do-
main of a numerical attribute ranges from v1 to v2, and that {c1, c2,…,ck} denote the k
cut-points obtained by the entropy/MDL algorithm, using these k midpoints, k/2 + 1
linguistic terms or membership functions can be determined for a trapezoidal fuzzy
set. The first membership function is computed as:
⎧ 1 if v1 ≤ x ≤ c1
⎪ .
f first ( x) = ⎨(c2 − x) (c2 − c1 ) if c1 < x < c2
⎪ x ≥ c2
⎩ 0 if
Generally, the middle pth, p =2,.., k/2, membership function is computed as:
⎧ 0 if x ≤ c2 p −3
⎪( x − c ) (c2 p −2 − c2 p −3 ) if c2 p −3 < x < c2 p −2 .
⎪⎪ 2 p −3
f p ( x) = ⎨ 1 if c2 p −2 ≤ x ≤ c2 p −1
⎪ (c − x ) (c2 p − c2 p −1 ) if c2 p −1 < x < c2 p
⎪ 2p
⎪⎩ 0 if x ≥ c2 p
⎧ 0 if x ≤ ck −1
⎪
f final ( x) = ⎨(ck − x) (ck − ck −1 ) if ck −1 < x < ck .
⎪ if ck ≤ x ≤ v2
⎩ 1
532 N.-C. Hsieh, C.-H. Chan, and H.-C. Tsai
For example, AAA_Size is a numerical attribute obtained from the dataset. The
domain of AAA_Size ranges from 5.0 to 9.6. Six cut-points, (5.5, 5.7, 5.8, 7.3, 8.1,
9.6), were obtained using the entropy/MDL algorithm. As depicted in Fig. 2, a trape-
zoidal fuzzy set was obtained with four membership functions corresponding to the
six intervals. If AAA_Size={5.65cm}, then it can be transformed to a triple value
[i2,max(round(0.75,0.8),round(0.3,0.4),i3] = [i2,0.8,i3], meaning that this value belongs
to interval i2 with membership value 0.8, and adheres to interval i3. If AAA_Size =
{6.55cm}, it can be transformed to a triple value [i4,0.5,i4]. Within each analysis, the
laboratory system should contain an interval that delimits life-compatible values.
Therefore, if cut points fall within the interval, they should be eliminated. For exam-
ple, one cut point {1.4 mg/dl} of creatine lay within an interval, i.e., [0.75~1.5] for
male, [0.5~1.2] for female; therefore the cut point {1.4 mg/dl} was eliminated.
BNs, NNs, SVMs and RSs were chosen because they represented different ap-
proaches, each of which is simple to implement and has been shown to perform well
in medical applications. The rationale of employing these models is that BNs can
easily model complex relationships among variables, NNs are generally superior to
conventional models, SVMs can be used as a benchmark technique, and RSs can be
used to augment the interpretability of the derived knowledge. The detailed configura-
tions of each individual model are as follows: BNs are Bayesian networks with K2
search algorithm, NNs are RBF neural network with radial basis functions as activa-
tion functions, SVMs use John Platt's sequential minimal optimization algorithm with
logistic to the outputs for proper probability estimates, and RS is Pawlak’s model with
Johnson search algorithm for generating the singleton reduce.
In this study, the model selection scheme is a mixture of bagging and cross-
validation (CV-bagging) that aims to improve the classification by combining models
trained on randomly generated subsets of the entire training set. We first applied a
cross validation scheme for model selection on each subset; subsequently, for the sake
of simplicity and so as not to run into over-fitting problems, we combined the selected
models in a uniform weights approach to reach a final decision. The concept of
Risk Prediction for Postoperative Morbidity of EVAR Using Ensemble Model 533
uniform weights voting is both appealing and simple to implement; it can generally be
applied to any type of classifier without relying on specific interpretations of the out-
put. In order to validate models, we used random subsets as cross-validation folds for
reasons of simplicity. In k-fold cross validation, the dataset is partitioned into k sub-
sets. Of these k subsets, a single subset is retained as the validation dataset, and the
remaining k-1 subsets are used for training datasets. The cross validation process is
then repeated k times with each of the k subsets used only once as the validation data.
The k results from the folds can then be averaged to produce a single estimation. In
cases where specific inside medical knowledge is not available, such a cross valida-
tion method can be used to select a classification method empirically, because it
seems to be obvious that no classification method is uniformly superior. For each
model, we train several individual classifiers and select the best performing model as
final model. If the trained models of each model differently with different initial con-
ditions, we then find proper values for the free parameters of the model. We extend
these results by combining models from different classes in order to increase the
model diversity. The result, a heterogeneous ensemble, allows classification methods
to be used more effectively.
The area under the ROC curve is based on a non-parametric statistical sign test, the
purpose of which is to estimate the probability of survival for each of a pair of pa-
tients, one of whom did not survive. The surviving patient gains greater probability of
survival. In this study, the area under the ROC metric was assessed through stratified
10-fold CV-bagging. The detailed accuracy of individual models and of the ensemble
TP FP Precisi F- ROC
Models Recall
Rate Rate on Measure Area
10 folds CV- bagging weighted average
BN 0.736 0.270 0.737 0.736 0.734 0.834
NN 0.675 0.330 0.657 0.657 0.674 0.755
SVM 0.757 0.247 0.758 0.757 0.756 0.755
Ensemble 0.757 0.249 0.760 0.757 0.755 0.813
model is shown in Figure 3. The results of the experiment show that individual mod-
els’ performance and improvements in performance were achieved by applying the
ensemble of models. This indicates that model combination techniques indeed yield
stable performance improvements for ensemble models.
3.3 BN’s Markov Blanket for the Causal Interpretability of the Prediction
Model
BN models have the ability to reason under conditions of uncertainty on the basis of
probabilistic inference mechanisms; therefore, they exhibit more resistance than other
machine learning models to noise that commonly occurs in medical datasets. That is,
if a dataset contains some samples with missing values, it can still be used to train or
test BN models. Because medical datasets tend to have missing values, this property
is useful for intelligent data analysis. Unlike most regression techniques, BN models
can easily model complex relationships among variables without any requirements for
the underlying distributions of variables and are more transparent and intuitive be-
cause relationships among variables are explicitly represented by a directed acyclic
graph (DAG) for modeling experts’ knowledge. The DAG consists of a set of nodes
linked by directed arcs, where the nodes represent the variables and the arcs indicate
the direct causal relationships among the connected attributes. The strengths of these
relationships are expressed by conditional probabilities. By using BNs, we can find a
given network’s structure and its associated parameters; the resulting model can pro-
vide diagnostic, predictive, and inter-causal reasoning. Therefore, the DAG of BN can
serve as a causal analysis tool to help surgeons arrive at valuable decision rules gen-
erated by RS. Through the use of causal-effect analysis, medical causal knowledge
can be exhibited through a causal DAG graph. Such knowledge can help surgeons
stimulate the hidden knowledge behind the medical dataset, communicate complex
causal concepts, and transform an individual’s tacit knowledge into a team’s explicit
knowledge.
A novel idea of BN for significant feature selection is the Markov blanket (MB)
[18], which is defined as the set of input features such that all other features are prob-
abilistically independent of the target features. In this case, based on a general BN
classifier, we can obtain a set of features that are present in the MB of the target fea-
ture, and features provided by the MB are sufficient for perfectly estimating the dis-
tribution of and for classifying the target feature classes. As an example, suppose that
we build a BN classifier with a complete input training data set. The MB of the target
feature then forms the feature selection mechanism and all features outside the MB
are ignored by the BN. The remaining features are then used to induce constrained
decision rules. In this study, RS is used to augment the interpretability of ensemble
model; all decision rules derived by RS are in the form, Y←X, where Y is the target
variable and X is a subset of variables from the MB. The rationale of this method is
that for any feature xi∉{MB∪xj}, xi≠xj and xi, xj∈MB, we have P(xi|MB,xj) =
P(xi|MB). That is, the MB of a variable xi is a minimal set of features MB such that xi
is conditionally independent of any other feature xj in MB. The MB structure to
be used for BN explores not only the relationship between target and predictive
variables, but also the relationships among the predictive variables themselves.
Risk Prediction for Postoperative Morbidity of EVAR Using Ensemble Model 535
In recent years, RSs and granular computing have been proven to be good soft-
computing techniques when used in various synergistic combinations with fuzzy
logic, neural networks and genetic algorithms. Since RSs do not require prior parame-
ter settings and extract knowledge from the data itself by means of indiscernibility
536 N.-C. Hsieh, C.-H. Chan, and H.-C. Tsai
where S(ti) denotes the degree to which ti satisfies the summary S, and the func-
tion μ EQ ( ak , bk ) = 0 if and only if (ak ≠ bk ) ∨ (ak = bk ∧ μ (bk ) = 0) ;
μ EQ (ak , bk ) = 1 − μ (ak ) − μ (bk ) if and only if ( ak = bk ∧ μ (bk ) ≠ 0) . Then, the individ-
ual truth value of “{t1,…,tn} are S” for a variable A over DS is computed as:
Truth({t1,…,tn} are S) = ⎛ 1 ∑ n S (t ) ⎞
⎜ i ⎟
⎝ n i =1 ⎠
Moreover, when the linguistic summaries are distributed over m variables with
ANDed conditions, that is, S = S1 ∧ … ∧ Sm, then the total truth value
Truth({t1,…,tn} are S)= min mj=1 (Truth({t1 ,..., tn } are S j ) . Finally, T=Q(Truth({t1,…,tn}
are S)) denotes the truth value of the linguistically quantified statement “Q {t1,…,tn}
are S” to the fuzzy quantifier Q in agreement. We gave an example for the procedure
in obtaining truth value using linguistic summaries. For example, let COPD be the
observation target; then {COPD, Seg, AAA_Size, AAA_Shape, Complications} is a
MB of partial causality. By using granular computing of RS, several decision rules
were obtained; we list three relevant decision rules as follows, where “support” means
number of cases supporting this rule:
IF THEN SUPPORT
COPD(A0) and Seg(i2,0.9,i3) and AAA_size(i1,1.0,i2) and Complications(Y) 7
AAA_Shape(A0)
COPD(A0) and Seg(i1,0.7,i2) and AAA_size(i1,0.8,i2) and Complications(Y) 3
AAA_Shape(A0)
COPD(A0) and Seg(i2,0.8,i3) and AAA_size(i4,0.6,i3) and Complications(Y) 1
AAA_Shape(A0)
Next, we defined the fuzzy quantifiers, Q, as shown in Fig. 6, which are repre-
sented as 7 membership functions. Then linguistically quantified statement for vali-
dating these decision rules is therefore given by:
“Q of the patients of postoperative morbidity have
COPD(A0) and Seg(i2) and AAA_size(i1) and AAA_Shape(A0).”
method of computing the individual truth value for the variable Seg is described as
follows. Let the linguistic summary, S1(Seg(i2))=(i2,1.0,i2) and a decision rule compo-
nent ti[Seg]=(i2,0.9,i3). Then, the membership value of ti to S1 of Seg is given by
538 N.-C. Hsieh, C.-H. Chan, and H.-C. Tsai
C o m p lic a tio n
No
Yes
Seg COPD
i4 0 .0 0 .1 0 .2 0 .3 0 .4 0 .5 0 .6 A999
P e rc e n ta g e %
i3 A2
i2 A1
i1 A0
0 .0 0 .1 0 .2 0 .3 0 .4 0 .5 0 .6 0 .0 0 .1 0 .2 0 .3 0 .4 0 .5
P e r c e n ta g e % P e rc e n ta g e %
A A A _ S iz e A AA_Shape
i7
i6 A12
i5
A2
i4
i3 A1
i2
A0
i1
0 .0 0 .1 0 .2 0 .3 0 .4 0 .5 0 .6 0 .0 0 .2 0 .4 0 .6 0 .8 1 .0
P e rc e n ta g e % P e rc e n ta g e %
4 Conclusions
Although several risk prediction systems have been proposed for patients undergoing
open aneurysm repair (Bohm et al., 2008), they basically rely on traditional statistical
methods and provide scant accuracy and utility when applied to EVAR patients. We
have proposed an ensemble model to predict postoperative morbidity after EVAR and
support clinical decision-making. The proposed ensemble model is constructed by
incorporating fuzzy discretization of categorical numerical values; BNs, NNs, and
SVMs were used to augment the ensemble model and the dataset was processed by
CV-bagging, showing moderate performance. The experimental result shows that the
proposed ensemble model predicts postoperative morbidity with relatively satisfactory
accuracy, even when data is missing and/or sparse, showing its usefulness in support of
clinical decision-making. In particular, the Markov blankets allow for a natural form of
causal-effect feature selection, which itself provides a basis for screening decision rules
generated by granular computing and generates supplementary information for risk
prediction. The learned knowledge is represented in multiple forms, including causal-
effect diagrams and causal-effect decision rules. The supplementary nature of multi-
models distinguish the proposed model from existing risk scoring systems that are
based on conventional statistical methods and from various machine learning models.
To summarize, the advantage of using the proposed ensemble model is that it can pro-
vide surgeons with practical, relatively accurate aid in their daily diagnostic tasks and
enable them to extract meaningful relationships among features of medical datasets
through the use of causality graphs and constrained decision rules.
Acknowledgement
This research was partially supported by National Science Council of Taiwan (NSC
97-2410-H-227-002-MY2).
References
1. Barnes, M., Boult, M., Maddern, G., Fitridge, R.: A model to predict outcomes for En-
dovascular aneurysm repair using preoperative variables. European Journal of Vascular
and Endovascular Surgery 35, 571–579 (2008)
540 N.-C. Hsieh, C.-H. Chan, and H.-C. Tsai
2. Bohm, N., Wales, L., Dunckley, M., Morgan, R., Loftus, I., Thompson, M.: Objective risk-
scoring systems for repair of abdominal aortic aneurysms: applicability in Endovascular
repair. European Journal of Vascular and Endovascular Surgery 36, 172–177 (2008)
3. Stijn, C.W., Wouters, L.N., Freek, W.A.V., Rene, M.H.J.B.: Preoperative prediction of
early mortality and morbidity in coronary bypass surgery. Cardiovascular Surgery 10,
500–505 (2002)
4. Roques, F., Michel, P., Goldstone, A.R., Nashef, S.A.: The logistic EuroSCORE. Euro-
pean Heart Journal 24, 1–2 (2003)
5. Bellazzi, R., Zupan, B.: Predictive data mining in clinical medicine: Current issues and
guidelines. International Journal of Medical Informatics 77, 81–97 (2008)
6. Verduijn, M., Peek, N., Rosseel, P.M., de Jonge, E., de Mol, B.A.J.M.: Prognostic Bayes-
ian networks I: Rationale, learning procedure, and clinical use. Journal of Biomedical In-
formatics 40, 609–618 (2007)
7. Verduijn, M., Peek, N., Rosseel, P.M., de Jonge, E., de Mol, B.A.J.M.: Prognostic Bayes-
ian networks II: An application in the domain of cardiac surgery. Journal of Biomedical In-
formatics 40, 619–630 (2007)
8. Lin, J.-H., Haug, J.: Exploiting missing clinical data in Bayesian network modeling for
predicting medical problems. Journal of Biomedical Informatics 41, 1–14 (2008)
9. Rowan, M., Ryan, T., Hegarty, F., O’Hare, N.: The use of artificial neural networks to
stratify the length of stay of cardiac patients based on preoperative and initial postoperative
factors. Artificial Intelligence in Medicine 40, 211–221 (2007)
10. Eom, J.-H., Kim, S.-C., Zhang, B.-T.: AptaCDSS-E: A classifier ensemble-based clinical
decision support system for cardiovascular disease level prediction. Expert Systems with
Applications 34, 2465–2479 (2008)
11. Polat, K., Güne, S.: Breast cancer diagnosis using least square support vector machine.
Digital Signal Processing 17, 694–701 (2007)
12. Babaoğlu, I., Fındık, O., Bayrak, M.: Effects of principle component analysis on assess-
ment of coronary artery diseases using support vector machine. Expert Systems with Ap-
plications (in Press)
13. Choi, S.: Detection of valvular heart disorders using wavelet packet decomposition and
support vector machine. Expert Systems with Applications 35, 1679–1687 (2008)
14. Pattaraintakorn, P., Cercone, N.: Integrating rough set theory and medical applications.
Applied Mathematics Letters 21, 400–403 (2008)
15. Mitra, S., Mitra, M., Chaudhuri, B.B.: A rough-set-based inference engine for ECG classi-
fication. IEEE Transactions on Instrumentation and Measurement 55, 2198–2206 (2006)
16. Podraza, R., Dominik, A., Walkiewicz, M.: Decision support system for medical applica-
tions. In: Hamza, M.H. (ed.) Applied Simulation and Modelling, Marbella, Spain (2003)
17. Wakulicz-Deja, A., Paszek, P.: Applying rough set theory to multi stage medical diagnos-
ing. Fundamenta Informaticae 54, 387–408 (2003)
18. Cheng, J., Greiner, R., Kelly, J., Bell, D.A., Liu, W.: Learning Bayesian networks from
data: an information-theory based approach. The Artificial Intelligence Journal 137, 43–90
(2002)
19. Pawlak, Z.: Rough sets. International Journal of Computer and Information Sciences 11,
341–356 (1982)
20. Pal, S.K.: Soft data mining, computational theory of perceptions, and rough-fuzzy ap-
proach. Information Sciences 163, 5–12 (2004)
21. Yager, R.R.: Database discovery using fuzzy sets. International Journal of Intelligent
Systems 11, 691–712 (1996)
Further Results on Swarms Solving Graph
Coloring
1 Introduction
D. Taniar et al. (Eds.): ICCSA 2010, Part III, LNCS 6018, pp. 541–551, 2010.
c Springer-Verlag Berlin Heidelberg 2010
542 M. Graña et al.
There are some other efforts to find other swarm inspired ways to solve the
GCP. A k-coloration of a undirected graph with vertex set V and E the set of
edges G = (V, E) is a partition of the graph in k classes of equivalence such that
the number of connected nodes belonging to the same class is minimized [6]. A
recent approach by ACO [5,4] to solve the GCP is [7] identifying ants to colors. In
[2] we have shown that our approach improved over some greedy algorithms over
a collection of hard coloring problems. Here we are more focused in identifying
the minimal swarm dynamics that allow to solve the coloring problem.
Section 2 gives some basic definitions of swarms for the GCP. Section 3 in-
troduces the Supervised Territorial Flock Coloring (SFTC). Section 4 reports
the experiments with the extreme version of the SFTC. Finally we gather some
conclusions and directions for further work in section 5.
v = Vm (αs vs + αc vc + αa va + αn vn ) (1)
board becomes continuous, and cells are agents that can change their positions
in the world. Each individual exhibits a very simple behavior that is specified by
a few simple rules that guide them to get along with the collective motion of the
flock. The global behavior of the flock emerges from these individual decisions.
We will stick to the birds metaphor, so that in the following, we call boids to
the agents that compose a flock.
The angle of vision, denoted ρ, comprises all the positions in the plane p = pi +v
where γvi − ρ2 γv γvi + ρ2 . In this article, the value ρ = 360o , meaning that
individuals can even observe what happens behind their backs. Let R be the
radius of vision: the cone of vision is the disk sector of radius R and angle ρ
around the boid. The private radius Z is a zone that separates the boid from
others.
Each boid is aware of an spatial region around it, its neighborhood. Let ∂iX
denote the population in the neighborhood of radius X. The set of boids lying
inside the neighborhood of the i-th individual of radius X is denoted:
where dist is the euclidean distance, R is the ratio of vision, and conei represents
the population in the cone of vision of the individual i. The population in goal
g is ∂gΓ = ∂ Γ (g) = {bj : dist (pi , pj ) < Γ }.
Further Results on Swarms Solving Graph Coloring 545
pi (t + 1) = pi (t) + vi (t + 1) (7)
where x+̇y = min {Σ, x + y} and x−̇y = max {0, x − y}. Predicates in− goal −
without − enemies and in − goal − with − enemies are defines as follows:
in − goal − without − enemies (bi ) ⇔ ∃g bi ∈ ∂gΓ ∩ Ēi
in − goal − with − enemies (bi ) ⇔ ∃g bi ∈ ∂gΓ ∩ Ei
enemies − in − all − goals (bi ) ⇔ ∀g ∂gΓ ∩ Ei = ∅
Next-goal is a criterion that allows the selection of the goal toward which a
boid navigates, g (t + 1) = next − goal (bi (t)) and can be for example: current
goal, nearest goal, nearest free goal,...
A TFC system is a pair (F , c̄) where F is a flock scheme and c̄ = c̄ (0) =
c1 (0) , . . . , c|B| (0) is the vector of initial configurations. A computation of
length k of a TFC system is the sequence c̄ (0) . . . c̄ (k).
experiment, the one that has minimum errors as well as the step at which the
minimum is reached. Each run ends either when a 3-coloring solution is reached
(success). or when 5000 iterations are completed in cascade. Figure 1 shows the
typical plot of the number of correctly colored nodes along a cascade coloration
process of a hard graph.
Table 1. Average succesful runs for the main topological graph classes and graph
topological measures
3-colored nodes per run), we calculated the Pearson coefficient over the rows of
1. The result is given in table
The Hostility quotient shows a very high negative correlation to the percentage
of successful runs per experiment (runs that find a 3-coloration) and to the
percentage of well 3-colored nodes per run. From this result we select Hostility
H as the best graph feature that explains tidily the results regardless of the
topology of the graph.
show poor results in the case of graphs that are not 3-colorable since not all
planar graphs are 3-colorable. The graphic shows also (dots) the worst cases per
experiment: even in this case, hard graphs obtain results above 90%.
5 Conclusions
We propose in this paper a framework to study GCP in terms of flocking birds
that show territorial behavior, calling the algorithm Supervised Territorial Flock
Coloring (STFC). We studied separately territoriality and flocking, finding that
Extreme Territorial Flock Coloring (ETFC), based in goal velocity and attack
is able to solve GCP. ETFC algorithm is able to color Mizuno’s hard graphs for
3-coloration while Brélaz algorithm fail for all them.
We found that a global graph parameter, the ratio of nodes at distance one
(enemies) to nodes at distance greater than one (friends), which we have called
hostility, is a good characterization of the difficulty of solving GCP for each
graph. It has negative strong correlation with the success and time to solve the
problem.
References
1. Brélaz, D.: New methods to color vertices of a graph. Communications of ACM 22,
251–256 (1979)
2. Cases, B., Hernandez, C., Graña, M., D’anjou, A.: On the ability of swarms to
compute the 3-coloring of graphs. In: Bullock, S., Noble, J., Watson, R., Bedau,
M.A. (eds.) Artificial Life XI: Proceedings of the Eleventh International Confer-
ence on the Simulation and Synthesis of Living Systems, pp. 102–109. MIT Press,
Cambridge (2008)
Further Results on Swarms Solving Graph Coloring 551
1 Introduction
In Japan, the population of aged people is increasing [1]. Accordingly, wheelchair
users are expected to increase. Government regulation has lessened barriers that hin-
der them. However, depending on the physical power of wheelchair users, there are
places where they are unable to move using their own physical power. Consequently,
many wheelchair users do not move around by themselves.
For wheelchair users to live independently, they should be supported to move
anywhere. For example, the system which helps movement of wheelchair [5] and the
system which guides a movable course of the wheelchair [2, 6] would be necessary.
However, almost all such studies investigate systems to show routes. No system has
been designed to collect information for a navigation database.
This research is undertaken to develop a system that navigates wheelchair users
who want to go to the destination and give them the possibility to venture freely out-
doors. Regarding the research, we first developed a system to measure the force to
move a wheelchair, the size of the space in which a wheelchair might turn back, and
position information. These data of forces for wheelchair movement and space sizes
can be matched with building plans and roads. As described, we report this system
developed to collect data for wheelchair user navigation.
D. Taniar et al. (Eds.): ICCSA 2010, Part III, LNCS 6018, pp. 552–563, 2010.
© Springer-Verlag Berlin Heidelberg 2010
Data Collection System for the Navigation of Wheelchair Users: A Preliminary Report 553
Force to move the wheelchair from the place of departure to the destination should be
corresponding to the map of the building and the road. The wheelchair positions are
identifiable using a Global Positioning System (GPS). We use the RMC format of GPS
to obtain the position information. The RMC format includes time, latitude, longitude,
ground speed, and direction of movement: such location data are sent as serial data.
The serial data are input to a built-in PC board via an AVR microcontroller with other
measured data. It is thereby possible to record all measured data with location informa-
tion. However, it is difficult to identify places within the buildings using GPS. We
attempted to identify the places in the buildings by the moving distances measured by
frequencies of wheelchair rotation from anchor points differentiated using QR codes
assigned to floor plans.
554 Y. Sumida, K. Goshi, and K. Matsunaga
The direction in which the wheelchair moves is identified using the QR code. In
Fig. 2, sheets on which QR codes are printed are attached on the ceiling at crossing
points of the passageways. The QR codes can indicate the wheelchair’s location and
direction of movement. The wheelchair position in the building is obtained according
to the moving direction identified by QR codes and moving distance measured using
the frequencies of rotation of the wheel from the position of certain QR codes at-
tached on the ceiling. The QR code includes position information in the building. In
Fig. 3, an example of QR code is shown. This QR code contains information of
"33.669835,130.444867,11" which means the degrees of latitude, the degree of longi-
tude and the altitude. The QR code includes position information in the building. In
Fig. 3, a QR code is shown: it contains information of "33.669835,130.444867,11",
denoting the degrees of latitude, the degree of longitude and the altitude. It is thereby
possible to obtain information about direction within buildings.
Fig. 3. A QR code
Wheelchair users are not always able to move on any path. For example, it might be
difficult to move on steep slopes, roads with different levels, and pathways covered
with deep-pile carpets. However, no map shows road surface states in detail. There-
fore, the force necessary to move wheelchair must be inferred.
The force necessary to move the wheelchair is determined by the current to the mo-
tor of the electric wheelchair. The current to the motor is obtained by conversion of
voltage by the resistor connected between the motor and the batteries. The voltage
between edges of the resistor is amplified and converted into a digital value by the
microcontroller’s (AVR, ATMEGA644P: Atmel Corp.) AD converter. The obtained
value is calibrated into the value of the drive torque and into the force.
Data Collection System for the Navigation of Wheelchair Users: A Preliminary Report 555
Wheelchair users want to know the width of spaces on the way to the destination.
They cannot turn back in narrow spaces. Therefore, for route navigation, space infor-
mation must be given that the space is sufficiently wide or not. To turn back, the pas-
sage width should be more than 3.6 m on many hand-driven wheelchairs and many
electricity-driven wheelchairs [4]. Width is measured using distance measuring de-
vices of ultrasonic type (EZ0 and EZ1, LV-MaxSonar: MaxBotix Inc.). The distance
measurement range is 0–6.5 m. The ultrasonic device outputs a pulse width that is
proportional to the distance to the object. The distance is obtainable by measuring the
pulse width by the counter of the AVR microcontroller. Equation (1) is used for com-
putation to make one count equal to 1 cm.
Count cycle = {1 [s] / (0.01 [m] / acoustic velocity [m/s])} /2 (1)
The acoustic velocity is altered by temperature, humidity, and air density. For that rea-
son, it is necessary to calibrate it according to these values.
As described previously, it is difficult for the GPS to obtain location data in build-
ings. Therefore, moving distances are measured using the wheelchair wheel’s
rate of rotation. It is obtained by counting by a device of counter a circular
pattern of monochrome alternation inside the wheel. The counter using a photo-
graph interrupter outputs pulses when it detects a white pattern. The rate of wheel
rotation was obtained by counting the standing up of the pulse every second using
an AVR microcontroller. The movement distance was obtained using eq. (2). The
movement velocity is calculable according to the movement distance and movement
time.
Movement distance each one second [m] = {wheel circumference [m] / (2)
number of white pattern} × pulse count each second
4.3.1 Design of the Software in the Interface and the Embedded PC Board
Data from all sensors and GPS are input into the AVR microcontroller in the interface
board. Then the inputted data are related with the GPS location data and sent to an
embedded PC board as serial data. The data transmission is done every second when
the data of GPS are complete. The data are transmitted to the data server and the web
server. Fig. 6 portrays a block diagram of the system showing the data flow.
Data Collection System for the Navigation of Wheelchair Users: A Preliminary Report 557
In-car unit
Built-in PC board
Picture Image
Serial data
with Picture
Mobile Phone
Memory
Video
Camera
We designed and developed two software, log-viewer software for checking log
data, and the live monitor software for a rear-time check. Fig. 7 shows the design of
the log-viewer software. Fig. 8 shows the design of the live-monitoring software.
5 Results
5.1 Results of System Performance
The measured data is stored in a flash memory by the character string by the recorder
program. The log-viewer software obtains each data by dividing this character string.
The stored data can be checked by the log-viewer software. Fig. 9 shows a method of
data processing in log-viewer software. The data is stored in order of the position
information from the GPS, measurement data of a microcomputer, data for debug-
ging. The Log-viewer software displays texts and graphs by reading this character
string, dividing it, and obtaining each data. Fig. 10 presents a screen display drawn
from data logged in the W-ASSIST hardware using the log-viewer software. The
hardware and the software of the system worked well to get signals from the GPS,
the camera, the current measurement device, the distance measurement devices, and
the wheel rotation-counting device. The log-viewer software is developed using Java.
The live-monitoring software consists of a recording program, a transmitting pro-
gram, a W-ASSIST server program and a web interface. Fig. 11 shows the flow of
information of the live-monitoring software. The recording program sends data to
summing program through UDP socket. The transmitting program sends live data to
the server of ASSIST through TCP socket on the mobile network. The live data con-
sists of times, wheelchair’s IDs, measurer’s IDs, latitudes, longitudes, measurement
data and front view images. The graphs are drawn using measured data for 20seconds
Fig. 10. Screen image displayed by the live-monitoring software through the web
Fig. 12. Screen image displayed by the live-monitoring software through the web
560 Y. Sumida, K. Goshi, and K. Matsunaga
each. Fig. 12 shows a screen display drawn from the data obtained by the W-ASSIST
hardware using the live monitor software. Data were transmitted to the servers using
the mobile phone system via the internet system. Software for live monitoring of the
driving behavior worked well (Fig. 12). The live-monitoring software is developed
using PHP and javascript.
Fig. 14 shows the results of measured aisle width which were gotten by the dis-
tance measuring devices. The differences could be thought to be small.
Fig. 15 shows the result of the measured voltage to analyze the currents to the mo-
tors. Although there are some differences, the reproducibility could be thought to be
high enough.
Data Collection System for the Navigation of Wheelchair Users: A Preliminary Report 561
Fig. 15. The result of measured voltage to analyze the driving force
5.2.1 Verification of the Relation between Drive Torque and Actual Power
To confirm the correlation of the voltages and the forces to move wheelchair.
y The method of experiment
The measurements were done five times by running the wheelchair on the slopes of
the same asphalt pavement. Those slope's gradient are different. And the forces to
move wheelchair were measured by the spring balance.
y The result of experiment
Fig. 16 shows the result of the measured forces to move wheelchair using the spring
balance and the voltage while the wheelchair movement for the measurement. Both
inclinations could be thought to be proportional to the slope angles. In addition, we
conducted multiple linear regression analysis about the starting powers and the run-
ning powers, and got the multiple regression equations (eq. (3, 4)).
562 Y. Sumida, K. Goshi, and K. Matsunaga
The needed starting power [kg] = 5.2054 + 0.0032 * Max voltage[mV] (3)
We developed the QRcode reading system using the opencv1.1 library and the libde-
codeqr library by C language. QR Codes are read in the following procedure. First,
the still pictures were gotten using camera. If a QR code in a still picture image was
found, it was clipped out. Next, it’s image was expanded into double and the QR code
was recognized. Finally, the location information in the QR code was analyzed. Fig.
19 shows an example of the screen display where the system was executed actually.
References
1. Statistics Bureau in Japan,
http://www.stat.go.jp/data/jinsui/tsuki/index.htm
2. Yang, A., Mackworth, A.K.:
http://people.cs.ubc.ca/~mack/Publications/CCAI07.pdf
3. Goshi, K., Matsunaga, K., Kuroki, D., Shidoji, K., Matsuki, Y.: Educational Intelligent
Transport System ASSIST. In: Proceedings of the Fourth IASTED International Confer-
ence Computers and Advanced Technology in Education, Banff, pp. 150–154 (2001)
4. Epoch House, http://www.epochhouse.com/care05.htm
5. Wang, H., Salatin, B., Grindle, G., Ding, D., Cooper, R.: Real-time model based electrical
powered wheelchair control. Medical Engineering & Physics 31(10), 1244–1254
6. Kasemsuppakorn, P., Karimi, H.A.: Personalised routing for wheelchair navigation. Jour-
nal of Location Based Services 3(1), 24–54 (2009)
Author Index