Agile Software Development Methods and Its Advantages
Agile Software Development Methods and Its Advantages
Technological Developments
in Networking, Education
and Automation
123
Editors
Prof. Khaled Elleithy Dr. Tarek Sobh
University of Bridgeport University of Bridgeport
School of Engineering School of Engineering
University Avenue 221 University Avenue 221
06604 Bridgeport Connecticut 06604 Bridgeport Connecticut
USA USA
[email protected] [email protected]
CISEE 09 was a high-caliber research conference that was conducted online. CISSE 09 received
432 paper submissions and the final program included 220 accepted papers from more than 80
countries, representing the six continents. Each paper received at least two reviews, and authors
were required to address review comments prior to presentation and publication.
The conference organizers and we are confident that you will find the papers included in this
book interesting and useful.
Khaled Elleithy
Tarek Sobh
Magued Iskander
Vikram Kapila
Mohammad Karim
Ausif Mahmood
Bridgeport, Connecticut
January 2010
vii
Acknowledgements
The technical co-sponsorship provided by the Institute of Electrical and Electronics Engineers
(IEEE) and the University of Bridgeport is gratefully appreciated. We would like to express our
thanks to Prof. Toshio Fukuda, Chair of the International Advisory Committee and the members
of the TENE, EIAE and IETA Technical Program Committees.
The excellent contributions of the authors made these world-class conference proceedings
possible. Each paper received two to four reviews. The reviewers worked tirelessly under a tight
schedule and their important work is gratefully appreciated. A complete list of reviewers is
provided in the book.
Khaled Elleithy
Tarek Sobh
Magued Iskander
Vikram Kapila
Mohammad Karim
Ausif Mahmood
Bridgeport, Connecticut
January 2010
ix
Contents
1. Education in the Clouds: How Colleges and Universities are Leveraging Cloud Computing................ 1
David C. Wyld and Rusty L. Juban
3. Problem Based Learning: Obtaining Enzyme Kinetics Parameters Integrating Linear Algebra,
Computer Programming and Biochemistry Curriculum ....................................................................... 13
A. Regalado-Méndez et al.
7. Following the Paper Trail: Measuring the Economic and Environmental Impact of Digital
Content Delivery .................................................................................................................................. 37
Vanessa K. Cox et al.
10. Faculty Driven Assessment of Critical Thinking: National Dissemination of the CAT Instrument..... 55
Barry Stein et al.
11. Enhancing the Instruction of Introductory Electric Circuit Courses using MATLAB ......................... 59
Thomas Yang et al.
xi
xii CONTENTS
17. Use of Ontology for Reusing Web Repositories for eLearning ............................................................ 97
Punam Bedi et al.
18. Actual and Future Trends of the E-education and Its Costs Specification in the Czech Republic ..... 103
M. Hedvicakova and P. Maresova
23. Wikibooks as Tools for Promoting Constructivist Learning in Higher Education: Findings
from a Case Study .............................................................................................................................. 133
Ilias Karasavvidis
25. Can Mobile Computing Help Reverse the Decline in Informatics Studies? ....................................... 145
D.Vitkute-Adzgauskiene and V. V. Fomin
26. Towards an Academic Social Network for Bologna Process ............................................................. 151
João C. Ferreira and Porfírio P. Filipe
27. Supporting Knowledge Discovery in an eLearning Environment Having Social Components ......... 157
Paola Monachesi et al.
CONTENTS xiii
29. Application of Project Based Study in the Learning Process: A Case Study ..................................... 169
Ashfaque A Chowdhury et al.
31. Stereo Spectral Imaging System for Plant Health Characterization ................................................... 181
S. C. Yoon and C. N. Thai
32. Transition Network for Automation Modeling: A Case Study ........................................................... 187
Yuval Cohen et al.
34. Design of An Intelligent Solar Tracking System Using PIC18F452 Micro Controller ...................... 197
Mostefa Ghassoul
35. Adaptive Two Layers Neural Network Frequency Controller for Isolated Thermal
Power System ..................................................................................................................................... 203
Ognjen Kuljaca et al.
38. One Approach for Training of Recurrent Neural Network Model of IIR Digital Filter ..................... 219
Stefanova S. A.
39. Classification of the Power Transformers using Dissolved Gas Analysis .......................................... 225
T. C. B. N. Assunção et al.
40. Modelling Control of Pore Number and Radii Distribution in Single-Cell Electroporation .............. 231
Sadhana Talele and Paul Gaynor
41. Single Allocation P-Hub Median Problem to Monitor Land Borders by Using
Unmanned Aircraft ............................................................................................................................ 237
Yasemin Akkuş and İnci Sarıçiçek
xiv CONTENTS
44. Depth Calculation using Computer Vision and Sift .......................................................................... 253
Edson CAVALCANTI Neto et al.
45. Dynamic of New Technologies in Switching Power Devices IGBTs ............................................... 257
Christian Ayala Esquivel et al.
46. Anisotropic Percolating Pathways in the Thin Films of Polymeric PEDT/PSS Complex
and their Relation to the Electrical Conductivity as Revealed by the Mesoscale Simulation............. 263
Toomas Kaevand et al.
47. Predicting Trading Signals of Sri Lankan Stock Market Using Genetic Algorithms
and Neural Networks .......................................................................................................................... 269
M. M. K. Dassanayake and Chandima Tilakarathne
51. Generalized Minimum Variance Controller with Dynamic Pole Assignment to Improve
Performance in Industrial Applications .............................................................................................. 291
M.A. Paz Ramos et al.
52. Automated Detection of Sunspots and Sunspot Groups in Full-Disk Solar Images ........................... 297
Sarab F. Majed et al.
53. Stochastic Model Based Approach for Biometric Identification ....................................................... 303
Mofakharul Islam et al.
54. A New Stochastic Model Based Approach for Object Identification and Segmentation in
Textured Color Image......................................................................................................................... 309
Mofakharul Islam and Paul A. Watters
55. An Unsupervised Stochastic Model for Detection and Identification of Objects in Textured
Color Images Using Segmentation Technique ................................................................................... 315
Mofakharul Islam and Paul A. Watters
CONTENTS xv
56. Automated Classification of Sunspot Groups with Support Vector Machines ................................... 321
Mehmmood A. Abd et al.
57. A Novel Ballistics Imaging System for Firearm Identification .......................................................... 327
Dongguang Li
60. Architecture and Initial Performance Indicators of New OPCUA Automation Protocol
Stack Java Implementation ................................................................................................................. 345
Mikko Salmenperä and Mikko Salonen
61. Automatic Segmentation of Cardiac Cavity Images Using Collinear and Triangle Equation ............ 351
Riyanto Sigit et al.
63. Intelligent Plasma Monitoring Systems Based on Optical Methods and Chromatic Modulation....... 361
Dimitrios Tomtsis
64. A Neoteric Chaotic & Adaptive Watermarking Scheme for Colored Images ................................... 367
Navita Srivastava and Achintya Singhal
66. Regression Analysis about Humidity Elimination from Diesel Fuel Via Bioorganic
Compounds to Increase Antifouling Action ....................................................................................... 377
C. G. Tsanaktsidis et al.
67. Optimal Threshold for Locating Targets Within a Surveillance Region Using a Binary
Sensor Network .................................................................................................................................. 381
Zhenxing Luo and Thomas C. Jannett
70. An IT Change Management Tool to Optimize Scheduling and Planning .......................................... 399
Denílson Cursino Oliveira and Raimir Holanda Filho
xvi CONTENTS
71. A Low-Power Content-Addressable Memory (CAM) using Pipelined Search Scheme .................... 405
Yibo Song et al.
76. Modular Design and Structure for a Mobile Sensory Platform ......................................................... 433
Ayssam Elkady et al.
77. wradvs: A Didactic Server for IPv6 Stateless Autoconfiguration ..................................................... 443
Gabriel Abadi et al.
79. Integrating Information Security into Quality Management Systems ................................................ 455
Margareth Stoll
82. A New Algorithm and its Implementations for RC Filters ................................................................. 473
Guoliang Zeng
84. A General Method for Synthesis of Families of Orthogonal Complementary Codes ........................ 487
B. Y. Bedzhev and M. P. Iliev
85. E-Speed Start- A Window Based Additive Increase, Multiplicative Decrease Network
Congestion Control Technique ........................................................................................................... 493
I. K. Oyeyinka et al.
CONTENTS xvii
89. GA based Neural Network Retraining using Adaptive Genetic Operations ....................................... 511
C. I. Bauer et al.
90. Decision Support for IP Uplink and Routing Choice ......................................................................... 517
Peng Su and Zhengping Wu
93. Vertical Handover between WiFi Network and WiMAX Network According to IEEE 802.21
Standard .............................................................................................................................................. 533
Mariem Thaalbi and Nabil Tabbane
94. Study on the Impact of Node Density and Sink Location in WSN..................................................... 539
Plamen Z. Zahariev et al.
95. An Algorithm for Synthesis of Families of Generalized Orthogonal Complementary Pairs .............. 543
B. Y. Bedzhev and M. P. Iliev
96. Low Message Overhead Clustering Algorithm for Wireless Sensor Networks.................................. 549
Batista et al.
99. Reduction of Energy Consumption in WSN using an Approach Geocast .......................................... 567
Harilton da S. Araújo et al.
101. Developing a Feedback Model for Managing Congestion in Packet Switched Wide Area
Networks ............................................................................................................................................ 579
E. F. Ogbimi et al.
102. Voice Gateway to IP-PSTN Integration and NGN Architecture ........................................................ 585
Geneflides Laureno da Silva and Raimir Holanda Filho
103. An Algorithm for Route Selection on Multi-sink Wireless Sensor Network Using Fuzzy Logic ...... 591
Líliam Barroso Leal et al.
104. Mining Mobile Sequential Patterns in Wireless Cellular Networks ................................................... 597
Joshua G. Bradley and Sherif S. Rashad
105. Agile Software Development Methods and its Advantages ............................................................... 603
Soly Mathew Biju
106. Toward a High Performance Grid Site RO-14-ITIM and the Contribution to WLCG ....................... 609
F. Farcas and G. Popeneciu
109. EvalTool for Evaluate the Partitioning Scheme of Distributed Databases ......................................... 627
Adrian Runceanu and Marian Popescu
110. Automatic Creation of Daily News Digests from Internet News Sites............................................... 633
David B. Bracewell
xix
Education in the Clouds: How Colleges and
Universities are Leveraging Cloud Computing
David C. Wyld and Rusty L. Juban
Southeastern Louisiana University
Department of Management – Box 10350
Hammond, LA 70402-0350 USA
Abstract- Cloud computing has been touted as a revolutionary and other smart, Web-connected devices in the past 10 years
concept in computing. In this article, we introduce the cloud (see Figure I).
computing concept, and then discus the implications for the use
of the cloud model in higher education. Due to budget
constraints and the power of computing “on-demand,” many
believe that colleges and universities will be at the forefront of
the shift to using cloud-based resources. We examine the impact
of cloud computing on both research and university computing
operations. We see that globally, this shift is already making a
difference in how universities procure and use IT resources. In
conclusion, we advance a six-stage Cloud Migration Strategy for
introducing and implementing cloud computing in institutions.
We also discuss the necessary changes in policies and
procurement, as well as security and reliability concerns for
institutions implementing cloud computing.
I. INTRODUCTION
1
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_1, © Springer Science+Business Media B.V. 2010
2 WYLD AND JUBAN
depicted in Figure II (The Cost Associativity Principle in been devoted to the concept of “freeconomics,” most notably
Cloud Computing). the recent book by Wired magazine editor Chris Anderson
[19] entitled, Free: The future of a radical price. Most
consumer-level cloud offerings would be labeled a
“freemium,” which is a free version that is supported by a
paid, premium version. Such freemiums are becoming an
emergent business model, as they are particularly popular
among online service and software companies [19]. And,
when faced with competing against “free” alternatives, older,
more established companies have seen users migrate to the
gratis alternative. Indeed, some see an entire “Culture of
free” emerging, where from music to entertainment to news
to software, people are coming to expect that free is the price
Fig. II. The Cost Associativity Principle in Cloud Computing they should pay [20].
Source: Based on Armbrust, et. al. [13]; Reynolds and Best [14].
In the corporate computing market, as software, hardware
and processing power, and storage capacity become more and
Additionally, where in the past only the largest universities
more commoditized, cloud computing becomes a free – or
have had supercomputing capabilities cloud computing, with
lower cost – alternative to the way things have been done for
number-crunching capabilities available on an on-demand
decades. As DiMaio [21] remarked: “Why should I bother
basis, affords researchers anywhere to scale their computing
looking for an email client to replace Outlook and coexist
power to match the scale of their research question - bringing
with my newly installed OpenOffice, if I can get email and
supercomputing to the mainstream of research. As Delic and
office suite as a service with somebody like Google at a
Walker [15] characterized it, cloud computing might just
fraction of the cost and - most importantly - giving up the IT
“enable new insights into challenging engineering, medical
management burden too? Why are we talking about moving
and social problems,” as researchers will now have newfound
servers from Windows to Linux when the real question is
capabilities “to tackle peta-scale type(s) of problems” and to
why do we need to have our own servers in the first place?”
“carry out mega-scale simulations.” Craig A. Stewart,
Already, there have been many campuses that have
Associate Dean for Research Technologies at Indiana
switched to Google or Microsoft-hosted email. Google and
University, recently remarked that with cloud computing,
Microsoft host email for over four thousand colleges and
“You reduce the barrier to use advanced computing
universities, not just in the U.S., but in over 80 countries
facilities” [16].
worldwide. In fact, almost half of all campuses are now
We have seen the reduction of barriers already paying
making use of hosted email services [13]. The switch to
dividends in research. At pharmaceutical giant Eli Lilly,
hosted services is paying significant dividends for the early
researchers needed to queue their projects to run in Lilly’s
adopting institutions. By switching to Gmail, Notre Dame
internal data center. This process to provision enough server
reports that it saved $1.5 million in storage and other tech
capacity for their respective projects often meant a delay of
costs, while at the same time, finding that their students’
up to two months waiting on their data run. Today however,
satisfaction with the campus’ email rose by over a third!
with cloud computing, research scientists can today provision
Likewise, institutions (such as Arizona State and Washington
the necessary processing capacity for their projects in five
State) are consistently reporting at least six figure annual
minutes. This allows researchers at Lilly and other research
savings from switching to Google or Microsoft hosted
organizations to crunch data and test theories in ways that
systems [22]. Even more importantly, by switching to hosted
may have gone unexplored in the prior era where they would
email and productivity software, the job and focus of college
have been dependent solely on in-house computing resources
IT staff can be changed. As Pepperdine University’s CIO
[17]! Similar experiences are being reported at universities,
Timothy Chester observed, his smaller IT staff can be more
both in the U.S. and abroad. For instance, at the International
productive. He recently stated: “We want our staff working
Institute of Information Technology in Hyderabad, India,
more with students and faculty and less on the nuts and bolts
Associate Professor Vasudeva Varma reports that the ability
of delivering technology” [22].
to run more data from more experiments more quickly has
Certainly, as challenging budgetary times have been
resulted in more publications for faculty in the Institute’s
forecast to persist across higher education for the next few
Search and Information Extraction Lab, which he heads [18].
years – at least, there will likely be even greater pressures on
B. Economics of Computing colleges and universities to replace “paid” software and
There is much discussion about the whole concept of computing resources with “free” or low-cost cloud
“free” pricing for products and services today – and many of alternatives. From the cloud provider standpoint, Google has
the email, storage, hosting, and applications that are at the stated that its incentive in providing such free services to
forefront of cloud computing today are indeed free. The most universities is to create “relationships for life” with students
notable of these are the product offerings of Google (Gmail, and faculty [23].
GoogleApps, GoogleDocs, and others). Much attention has
4 WYLD AND JUBAN
III. ANALYSIS faculty could have access to the scale and type of computing
power enjoyed only by elite institutions [26].
Many in higher education are coming to concur with the
opinion of Keahey [24], who plainly stated: “I believe that IV. CONCLUSION
this (cloud computing) will be the model of the future. This
will absolutely drive cost savings for the universities.” A. A Cloud Migration Strategy for Higher Education
Across higher education, the cloud computing landscape It is important to bear in mind the advice of Higginbotham
should be quite active over the next few years, as we will see [27] that “cloud computing is a tool, not a strategy.” IT
both coordinated efforts and “rogue” operations that will test leaders in higher education will thus be well-advised to take a
how and where cloud computing can be effectively applied. programmed, assessment of how cloud computing can fit into
As we have seen, colleges and universities will in many their overall IT strategy, in support of the mission and overall
instances lead the way. These entities will continue to do so, strategy of their institution. This should take a 6-step process,
based on their need for computing power on demand and for as shown in Figure III.
providing the types of ready – and in many cases free – IT
resources – to their faculty and students. With pressure to
reduce the fixed costs of higher education – and IT being a
very rich target – the shift to cloud may be more forced in
some cases than may be dictated by the on-the-ground
circumstances. Indeed, some of the most exciting uses and
best practices for cloud computing could well come from the
world of higher education.
We have seen predictions that due to the cost and
operational benefits of cloud computing, more and more
companies will find themselves outsourcing most – if not all
– of their IT to cloud providers, creating what has been
termed as “serverless organizations” [25]. Indeed, it has been Fig. III. From the Internet to Cloud Computing.
predicted that organizations of all sizes will find it beneficial
to concentrate on and optimize their business processes by The Cloud Migration Strategy begins with learning about
outsourcing the IT function. So, why not “serverless the basics of cloud computing – through attending seminars,
universities”? By outsourcing almost all of IT and all data networking, talking with vendors, and reading articles such
storage/handling – this may be a viable proposition for as this one. Given that cloud computing represents a new
colleges and universities, particularly as cloud offerings paradigm in computing technology, it will be important for
expand and are made more secure and reliable. technology transfer to occur – the “techies” in and outside of
As we have seen in this article, there are certainly the institution will need to go the extra mile to educate and
discussions and embryonic efforts underway – both in the inform the “non-techie” amongst their ranks and
U.S. and abroad - as public and private universities examine constituencies as to the merits and value of cloud computing.
how to best made the cloud-concept work for they and their It will be especially important to devote sufficient funding for
students and faculty. Universities are beginning to work research to establish how cloud computing is working – or
collaboratively in the cloud to pool their IT resources. not working – in various areas in the university and across
Already, this has occurred in Virginia and North Carolina. In institutions, so as to ground policies and develop best
the Commonwealth, a dozen colleges and universities have practices in regards to the use of cloud computing.
come together to form the Virginia Virtual Computing Lab Then, IT executives should conduct an honest assessment
[16]. Such efforts allow institutions to cut their IT costs by of their institution’s present IT needs, structure, and capacity
reducing their need for software licensing, for upgrade utilization. In a cloud computing environment, where
capabilities, and for perhaps maintaining their own data resources can be added – or subtracted – based on needs and
centers, all while improving the IT resources for their faculty demand, it will be critical for IT managers to honestly assess
and students. Already, by shifting to cloud offerings, North their institution’s IT baseline for faculty, students and
Carolina State University has been able to dramatically lower operations. In looking at data center utilization, it will be
expenditures on software licenses and simultaneously, reduce vital to look at what resources are used all the time and are
the campus’ IT staff from 15 to 3 full-time employees [18]. necessary for day-to-day operations to establish a baseline for
Additionally, there have been calls for the federal internally-hosted operations. Only then can one look at
government to take the lead to create a cloud computing whether to continue to host “excess” capacity in the data
environment, to be available for use by all colleges and center or to contract for cloud services as needed to scale-up
universities nationwide. In doing so, proponents argue for the to meet demands for greater amounts of computing resources.
economic and educational benefits that such a resource University IT leaders should then pick one area – even one
would provide, as it would democratize computing specific project - to “cloud pilot” and assess their ability to
technology and “level the playing field” so all students and manage and bring such a project to fruition. As with any new
technology, we are seeing a great deal of pure
EDUCATION IN THE CLOUDS 5
experimentation with cloud computing – “science project” examine how cloud computing will secure and procure their
like work for the most part up till now. All of us who use the computing environment.
Internet are experimenting with cloud applications in our Indeed, one of the principal concerns about cloud
daily lives – from Twittering to Gmail to using photo-sharing computing whether it is secure and reliable. Unfortunately,
sites. In the same way, we are seeing organizations worries over cloud reliability and availability – or
conducting cloud computing trials - what Woody [17] termed specifically, the lack thereof when such instances arise - are
as “science experiments” in the use of the technology. Such not just theoretical, as there have been well-publicized
efforts that are far away from their core IT operations and outages of many of the most popular public cloud services
many times on (or trying to connect) the periphery of the [30, 31]. And, as Schwartz [32] astutely pointed-out, when
organization. . Many times – even in the public sector and cloud service outages or inaccessibility occur, “most of the
especially on campuses, these experiments may be “rogue” risk and blame if something goes wrong will fall directly on
operations – taken on by individuals and units to test the the shoulders of IT -- and not on the cloud computing service
utility of the technology. These are important efforts, and providers.”
they should be supported – and reported within and outside Security concerns may indeed impede the shift to cloud-
the institution – so that others in the IT and the wider based models. As with prior shifts in information technology
community can learn of the successes – and the downsides – with the advent of the Internet and the Web, the introduction
of operating in the clouds. Thus, it will be vitally important to of e-mail, and the explosion of social media, their growth and
share both “best practices” and “lessons learned” in cloud adoption rates have been slowed by initial fears – some
computing. Indeed, many predict that such “science projects” justified and some very unjustified – over security concerns
in large and small organizations will drive the eventual and the loss of control over data and operations [33].
acceptance and adoption of cloud computing [28]. Certainly, privacy and security questions will need to be
After the internal assessment and external outreach addressed as institutional data and applications move into a
stemming from the pilot effort, they should then conduct an cloud environment. Indeed, analogies have been drawn
overall IT cloud-readiness assessment to determine if they between the advent of cloud computing today with the
have data and applications that could readily move to a cloud introduction of wireless technologies a decade ago [34].
environment and if a public/private/hybrid cloud would be Finally, security is undoubtedly a hard metric to quantify.
suitable or useable for these purposes and rank-order And, all too often, from the perspective of Golden [35] and
potential projects. Finally, it is time to begin a cloud rollout other observers, the IT community has a somewhat damaging
strategy – gaining buy-in from both institutional leadership tendency to treating all risks – whatever the real nature of
and IT staffers and communicating with both internal and them – as the very worst case scenario and not judging the
external stakeholders as to the goals, progress, and true impact – and likelihood – of their occurrence.
costs/benefits of each cloud project. This is where the cloud Finally, universities’ often outdated and byzantine
goes from being a test effort to become more mainstream in procurement rules and regulations, some of which may even
the way the university manages its data, its operations and its preclude the use of cloud computing in select instances, will
people. It becomes part of “normal” operations, just as other need to be changed to be more cloud-friendly and encourage
prior tech innovations (from telephony to fax to the Internet the savings and efficiencies that can come from this new
to email and to social media) have become IT tools, used in model of IT [36]. There will also need to be changes made in
support of the institution’s IT strategy and more importantly, not just the language, but in the mindset of contracting for
its overall strategy. computing services. For while IT administrators look at
At this point, the process enters the final stage – call it capacity and systems, end users look to performance. As
“continuous cloud improvement” – to where the institution Jackson [37] recently put it, the key metric will now become:
continues to move appropriate data and applications to the “When I sit down at that computer, do I see the functionality
cloud – and perhaps even back from the cloud to internally- I need?”
hosted operations, if necessary, based on a thorough and In time, we may look back on the latter portion of this first
continuous assessment of the appropriate use of cloud decade of the new millennium as a true turning point in the
technologies for their particular university. history of computing. The transition however will take years,
perhaps even decades, and “we're not going to wake up
B. Implications for Higher Education tomorrow and get all our computing requirements through a
The shift to more cloud-based applications will indeed socket in the wall” [38]. However, all signs point to a true,
bring newfound capabilities to communicate, collaborate and campus-led revolution in computing.
conduct research to university faculty, staff and students.
However, it will also necessitate a flurry of policy decisions REFERENCES
that will need to be made and operational rules that will need
to be implemented. For instance, there will have to be IT [1] The Economist Editorial Staff, “Let it rise,” The Economist, October
23, 2008. [Online]. Available:
policy decisions made as to who can access what files and http://www.economist.com/specialreports/displayStory.cfm?story_id=1
what type of access they will have (i.e. read-only, editing 2411882 [Accessed: March 3, 2009].
access) [29]. The shift will also necessitate institutions to [2] Gartner (2008), Press Release: Gartner says cloud computing will be
as influential as e-business, June 26, 2008. [Online]. Available:
6 WYLD AND JUBAN
E
An intelligent tutoring system is defined in [25] as a
stablishing the knowledge state of a student in a "computer system" based on "artificial intelligence" designed
subject is crucial for providing him/her with individual help. to deliver content and to provide feedback to its users. In [26]
Subsets of relevant examination questions and certain skills an intelligent tutoring system is considered to be a software
from a branch of knowledge are listed as examples of agent that provides customized and adaptable instructions and
knowledge states in [12]. It is quite important to underline that feedback to learners. The topic received a lot of attention from
not all possible subsets of such items turn out to be knowledge the research community since the 1980's. However, it seems
states. that the process of providing adequate help to users based on
Outcomes of diagnostic tests usually give an indication their individual responses is still open.
about the general status of students’ knowledge. At the same A cognitive model represents the knowledge that an ideal
time, they do not address the problem of classification of student would possess about a particular subject. Developing
knowledge states. The latter is strongly related to a cognitive tutor involves creating a cognitive model of
identification of existing misconceptions and lack of adequate students’ problem solving by writing production rules that
knowledge. This work presents an approach that can be used characterize the variety of strategies and misconceptions
in the process of solving these problems. students may acquire, [16, 17, and 18]. Cognitive tutors have
Based on students' responses a content provider can follow been successful in raising students' math test scores in high
the process of response clusterings and make appropriate school and middle-school classrooms, but their development
adjustments in contents of a problem or the related materials has traditionally required considerable time and expertise, [8].
or both. Response clusterings can be used to track down the An investigation of whether a cognitive tutor can be made
effect of assistance provided to each student and thus more effective can be found in [15].
contribute to a better individualized assistance. The process of delivering hints in an intelligent tutoring
The rest of the paper is organized as follows. Section 2 system has been discussed in [24]. Taxonomy for automated
contains information about related work. Section 3 is devoted hinting is developed in [20]. The role of hints in a Web based
to ordered sets. Section 4 relates various personal responses learning systems is considered in [7] and [8]. Usability issues
and Section 5 is devoted to a system description. Section 6 of a Web-based assessment system are discussed in [13] and
contains the conclusion of this work. [21].
A model for detecting students’ misuse of help in intelligent
tutoring systems is presented in [3]. A proliferation of hint
abuse (e.g., using hints to find answers rather than trying to
II. BACKGROUND
understand) was found in [1] and [15]. However, evidence
that when used appropriately, on-demand help can have a
Navigational structures in an automated tutoring system positive impact on learning was found in [19] and [23].
should prevent students from becoming overwhelmed with
information and losing track of where they are going, while
7
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_2, © Springer Science+Business Media B.V. 2010
8 ENCHEVA AND TUMIN
III. ORDERED SETS Theorem 1 along with several axioms, [5] provides an
existing social welfare function.
Two very interesting problems about orderings are
considered in [5], namely the problem of determining a Theorem 1 There is a unique social welfare function
consensus from a group of orderings and the problem of satisfying Axioms 0-4. It is given by: i is preferred to j if
making statistically significant statements about ordering. and only if the sum of the row of A corresponding to j is
An ordered set (or partially ordered set or poset) is an
larger than the sum of the row of A corresponding to i , and
ordered pair ( P, ≤) of a set P and a binary relation ‘ ≤ ’
otherwise the social ordering is indifferent between i and j .
contained in P × P called the order (or the partial order) on
Two elements a and b where a ≠ b and a, b ∈ P are
P such that ‘ ≤ ’ is reflexive, antisymmetric and transitive.
comparable if a ≤ b or b ≤ a , and incomparable otherwise.
A relation I is an indifference relation when given AIB
neither A > B nor A < B has place in the componentwise If ∀a, b where a, b ∈ P are comparable, then P is a chain.
ordering. A partial ordering whose indifference relation is If ∀a, b where a, b ∈ P are incomparable, then P is an
transitive is called a weak ordering. antichain.
Let ω1, ω2, ω3 be weak orderings. The ω2 is between ω1
and ω3 if each decision made by ω2 is made either by ω1 or
B. Knowledge spaces
ω3 and any decision made by both ω1 and ω3 is made by ω2,
i.e. ω1 ∩ ω3 ⊆ ω2 ⊆ ω1 ∩ ω3.
The distance is defined as Within the theory of knowledge spaces, [11], [12] a
d(ω1, ω2) + d(ω2, ω3) = d(ω1, ω3 ). knowledge domain is described by a set of problems Q. Then
The distance is a metric in the usual sense, it is invariant under the knowledge state of a student is a set Q’ ⊆ Q that the
permutation of alternatives, and the minimum positive student can solve. A knowledge structure of a subject is a set
distance is 1. If two weak orderings agree except for a set of of all knowledge states.
alternatives that is an open interval in both, then the distance A knowledge space is a closure system, [11] where closure
may be computed as if the alternatives in this set were the system on a finite set M is a set F of subsets of M such that
only objects being ranked. M ∈ F and C , C 1 ∈ F C ∩ C 1 ∈ F
If given two alternatives, a person is finally choosing only one.
The natural extension to more than two elements is known as the
'majority rule' or the 'Condorcet Principle'. A relation
IV. GEOMETRIC INTERPRETATION OF KNOWLEDGE STATES
R ( L1 , L2 ,..., Lk ) is constructed by saying that the pair
(a, b) ∈ R if (a, b) belongs to the majority of relations Li . Initially we are trying to establish the status of students’
knowledge in a topic with respect to three knowledge states
The following three linear orderings a b c , b c a , and
called a, b and c. This is realized by considering students’
c a b , leading to R = {( a, b), (b, c), (c, a)} (three-way answers to three questions in a multiple choice test that is a
tie), illustrate the 'paradox of voting', [4]. part of a Web-based system. Students' responses are saved in
The probability of the voting paradox for weak orderings is a database. Depending on whether a provided answer is
calculated analytically for the three-voter-three-alternative correct, partially correct or incorrect, the student is suggested
case. It appears that the probability obtained this way is appropriate reading materials, hints or examples. Afterwards
considerably smaller than the corresponding case for linear she can repeat the test. The consecutive test contains new
orderings, [22]. questions but related to the same knowledge states. This
allows the course/test designer to follow each student’s
progress within any knowledge state and to obtain a clear
A. Social Welfare Function
picture about the degrees to which these knowledge states are
reached by a particular student, and by a group.
A 'social welfare function' maps k -tuples of the set of If for example one or more of the knowledge states
linear orderings of any b ⊂ A to single linear orderings repeatedly appear to be difficult to be reached by a single
of B , where A is a set of at least three alternatives, [2]. student or by a group then the corresponding course materials
and the related questions, have to be redesigned.
A social welfare function acts on k -tuples representing
A knowledge state is assumed to be reached by a student if
either weak or linear orderings of k individuals for m the student provides a correct answer to the question
alternatives. The representation of an individual's ordering can presented to her/him. The question is of course related to that
be thought of as a column vector in which an integer in knowledge state. A knowledge state is assumed to be partially
position i represents the preference level the individual reached by a student if the student provides a partially correct
assigns to alternative i . Thus k individuals presented with answer to the presented to her/him question, and knowledge
m alternatives can be illustrated by k -tuples of orderings in state is assumed not to be reached by a student if the provided
a m × k matrix of integers of preference levels. answer is incorrect.
WEAK ORDERINGS OF KNOWLEDGE STATES 9
Using nested parentheses one can represent the partial With this distribution, one can safely conclude that
ordering structure internally in a given application. Let the knowledge state 'a' has been reached by the majority of
alphabets in this definition be '( ), a, b, c'. The use of students. Another interesting observation is that knowledge
parentheses signifies the position of ordering of a particular state 'c' is reached by 20% of the students. Since the distance
element 'a', 'b', or 'c', where ‘( )' implies correct answer, '(( ))’ from (ab(c)) or (a(bc)) to (c(ab)) is the same, in this particular
implies partially correct answer, '((( )))' implies incorrect case, we can safely conclude that there are two types of
answer. Parentheses are not used for the elements in the center students with different knowledge background in relation to
of the graph. the learning process of this particular study unit. Such
Note that the set {a, b, c} is neither a chain no an antichain, observations are very useful for future tuning of course
i.e. a pair of different elements can either be comparable or materials.
incomparable. When a pattern of clustering does not show any significant
value but is evenly distributed amongst the twelve ordering
patterns (see f. ex. Fig. 5) then one can say that none of the
provided materials was particularly helpful for the students’
study progress. Both the course materials and the questions in
diagnostic tests need to be redesigned.
Fig. 1 Partial ordering (a(b(c)))
Discussions on distances between different orderings are
enclosed in the next subsections.
The ordering ( a (b(c ))) in Fig.1 implies that knowledge state
‘a’ is reached by that student, knowledge state ‘b’ is partially
A. Orderings Where Two of the Knowledge States Either
reached, and knowledge state ‘c’ is not reached by that
Reached or Partially Reached
student.
Two of the knowledge states in Fig. 8, ‘a’ and ‘c’ are being
respectively reached and partially reached in ordering
(a(c(b))) and are partially reached and reached in ordering
(c(a(b)), where knowledge state ‘b’ is not reached in both
orderings. This indicates that the course materials, hints,
questions, etc. related to knowledge state ‘c’ do not currently
WEAK ORDERINGS OF KNOWLEDGE STATES 11
V. SYSTEM DESCRIPTION
12 ENCHEVA AND TUMIN
A system prototype is build as a Web-based application Intelligent Tutoring Systems, ITS 2000 Berlin: Springer Verlag (2000)
292—303.
using Apache HTTP server, mod_python module and SQLite
[2] Arrow, K. J. (1951). Social Choice and Individual Values. Wiley, New
database. The mod_python module provides programmable York, (1951).
runtime support to the HTTP server using Python [3] Baker, R.S., Corbett, A.T., and Koedinger, K.R.: Detecting student
programming language. The application components are Web- misuse of intelligent tutoring systems. Lecture Notes in Computer
based users interface, application logic and application Science, Vol. 3220. Springer-Verlag, Berlin Heidelberg New Jork
(2004) 531-540
interfaces written in Python, and a relational database. The [4] Berg, S.: Condorcet's Jury Theorem and the Reliability of Majority
system architecture is presented in Fig. 12. Voting, Group Decision and Negotiation, 5(3), (1996), 229-238
The system contains also an interactive theoretical part [5] Bogart, K. P.: Some social sciences applications of ordered sets. In: I.
(lecture notes), problems to be solved at home, tests assessing Rival, Editor, Ordered Sets, Reidel, Dordrecht (1982), 759–787
[6] Breuker, J., Components of Problem Solving and Types of Problems. In
recall of facts, an ability to map the relationship between two EKAW '94, (1994), 118-136
items into a different context, and conceptual thinking, asking [7] Brunstein, A. and Krems, J.: Helps and Hints for Learning with Web
students to evaluate consequences and draw conclusions, and Based Learning Systems: The Role of Instructions, Intelligent Tutoring
various quizzes. Systems, Lecture Notes in Computer Science 3220, (2004), 794-796
[8] http://www.carnegielearning.com/web_docs/Whitepaper_Edweek.pdf
The users, i.e. expert tutors, teachers, and students interact [9] http://ctat.pact.cs.cmu.edu/
with the system using Web forms. Experts and teachers can [10] http://www.learning-styles-online.com/overview/
submit and update data, while students can only view [11] Doignon, J.-P. and Falmagne, J.-C.: Knowledge Spaces. Springer-
information. Verlag, 1999.
[12] Dowling C.E.: On the irredundant generation of knowledge spaces.
The system provides recommendations on whether or not a Journal of Mathematical Psychology, 37, 1993, pp. 49-62
student needs to take additional classes (courses) based on [13] Jan, N., Shun-Chieh Lin, Nancy P. Lin, Chang, C., A Trend-based
fuzzy dependencies. Prediction System for Web User Behavior, WSEAS Transactions on
Advances in Engineering Education, 5, 2008
[14] Keefe, J. W., Learning style: An overview. In NASSP's Student learning
styles: Diagnosing and proscribing programs, (1979) 1-17. Reston, VA.
National Association of Secondary School Principles.
[15] Koedinger K.R., McLaren B.M. and Roll I.: A help-seeking tutor agent.
Proceedings of the Seventh International Conference on Intelligent
Tutoring Systems, ITS 2004, Berlin, Springer-Verlag, (2004) 227-239
[16] Mayo, M. and Mitrovic, A.: Optimising ITS behaviour with Bayesian
networks and decision theory. International Journal of Artificial
Intelligence in Education 12 (2001) 124-153.
[17] Pecheanu E., Segal C. and Stefanescu D.: Content modeling in
Intelligent Instructional Environment. Lecture Notes in Artificial
Intelligence, Vol. 3190. Springer-Verlag, Berlin Heidelberg New Jork
(2003) 1229-1234.
[18] Renkl, A.: Learning from worked-out examples: Instructional
explanations supplement self- explanations. Learning and Instruction,
12, (2002) 529-556
[19] Stewart, K. L., and Felicetti, L. A., Learning styles of marketing majors.
Educational Research Quarterly, 15(2), (1992) 15-23
[20] Tsovaltzi, D., Fiedler, A., and Horacek, H.: A Multi-dimensional
Taxonomy for Automating Hinting, Intelligent Tutoring Systems,
Lecture Notes in Computer Science 3220, (2004), 772-781
Fig. 12: System architecture [21] Yeum, Y., Lee W., Usability Issues of Web-Based Assessment System
for K-12 and Implementation of PAS System, WSEAS Transactions on
Advances in Engineering Education, 9(4), (2007)
[22] Van Deemen, A., The Probability of the Paradox of Voting for Weak
V. CONCLUSION Preference Orderings. Social Choice and Welfare, 16(2), (1999)
[23] Wood, D.: Scaffolding, contingent tutoring, and computer-supported
learning. International Journal of Artificial Intelligence in Education, 12
The work described in this article provides an approach (2001) 280-292
that helps to locate significant gaps in students’ knowledge. In [24] Zhou, Y., Freedman R., Glass, M., Michael, J. A., Allen A., Rovick, A.
A., and Evens, M. W.: Delivering hints in a dialogue-based intelligent
addition, it gives an opportunity not only to follow the tutoring system, Proceedings of the sixteenth national conference on
learning progress of individuals but also to recognize pattern Artificial intelligence and the eleventh Innovative applications of
clustering’s within a study unit. The course developer can artificial intelligence conference innovative applications of artificial
then easier find out which of the involved reading materials, intelligence, (1999), 128-134
[25] http://design.test.olt.ubc.ca/Intelligent_Tutoring_System
hints and tests need further adjustments and what the future
[26] http://www.cs.ualberta.ca/ zaiane/courses/cmput605/2008-3.html
work should be concerned about.
REFERENCES
[1] Aleven, V., and Koedinger, K. R.: Limitations of Student Control: Do
Student Know when they need help? In G. Gauthier, C. Frasson, and K.
VanLehn (Eds.), Proceedings of the 5th International Conference on
Problem Based Learning: Obtaining Enzyme
Kinetics Parameters Integrating Linear
Algebra, Computer Programming and
Biochemistry Curriculum
Regalado-Méndez, A. (a), Báez-González, J., G. (b) Peralta-Reyes, E. (a), and Trujillo-Tapia, M., N. (a)
(a) Ingeniería Ambiental, Universidad del Mar-Puerto Ángel, Ciudad Universitaria S/N, Km. 1.5 Carretera
Puerto Ángel–Zipolite, Puerto Ángel, Oaxaca, 70902, México.
(b) Departamento de Alimentos, Facultad de Ciencias Biológicas, Universidad Autónoma de Nuevo León,
Av. Universidad S/N, Ciudad Universitaria San Nicolás de los Garza,
Nuevo León, 66451, México.
E-mail: [email protected]
I. INTRODUCTION
In Mexico, problem based learning (PBL) has not been
developed much, contrary to what is found in European
countries and United States. In the last few years, this type of
learning has become very popular [8], allowing the integration
of different disciplines in to the curriculum. Frequently,
students ask themselves about the usefulness of some courses, Fig. 1. Number of failed and accredited students each year.
as they don’t manage to see any concrete application. To
ensure that students take a primary part in their own academic To improve the Teaching-Learning process in Mexico, it is
formation, the lecturer should assume a facilitator role, and necessary that professors interconnect the knowledge obtained
encourage the students to motivate themselves in searching for by the program curricula not only in a horizontal way but also
possible solutions to a given problem, looking at it from a vertically, with the objective to improve the cognitive
multidisciplinary viewpoint. processes of students in the first semesters, in order to form
As shown in fig.1, students from the Universidad del Mar more critical and less procedural students [9].
have high failure statistics which leads teachers to seek The PBL structure was kept similar to the one adopted by
different pedagogic approaches. A project that allows Offman and Cadet, 2002, with 3 work sessions of 2-3 hours,
integrating courses of biochemistry, linear algebra, and the first with an introductory purpose. In these sessions,
programming is proposed herein. It is intended for use with discussion is encouraged with a minimum involvement of the
undergraduate students of the Universidad del Mar, Mexico. professor (facilitator) [1, 3, 6 and 7]. Each session is spaced by
Until now, it has been applied to students in the Environmental three days; in each a written report that answers specific
Engineering program, but it can be expanded to students of questions is expected.
natural science.
13
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_3, © Springer Science+Business Media B.V. 2010
14 REGALADOǦMÉNDEZ ET AL.
i =1 ⎣ i =1 ⎦
(8)
m
⎡m ⎤⎡ m ⎤
C. Part II (Session 2) m∑ f ( xi ) xi − ⎢ ∑ f ( xi ) ⎥ ⎢ ∑ xi ⎥
As equation (3) is hyperbolic and obtaining the Vmax and km a1 = i =1 ⎣ i =1 ⎦ ⎣ i =1 ⎦
2
parameters directly is complex, a first step is to make the m
⎡m ⎤
m∑ xi − ⎢ ∑ xi ⎥
2
⎛ m ⎞ ⎛ m ⎞⎛ m ⎞
m ⎜ ∑ xi yi ⎟ − ⎜ ∑ xi ⎟⎜ ∑ yi ⎟
According to the case, one or the other equation can be ⎝ i =1 ⎠ ⎝ i =1 ⎠⎝ i =1 ⎠
r2 =
adopted. The choice is basically guided by the experimental ⎡ m 2 ⎛ m ⎞2 ⎤ ⎡ m 2 ⎛ m ⎞2 ⎤ (10)
data that, once in the linear form have to be far enough from ⎢ m∑ xi − ⎜ ∑ xi ⎟ ⎥ ⎢ m∑ yi − ⎜ ∑ yi ⎟ ⎥
each other in order to obtain a confident linear curve. The ⎢⎣ i =1 ⎝ i =1 ⎠ ⎥⎦ ⎢⎣ i =1 ⎝ i =1 ⎠ ⎥⎦
following deviations to linearity can be related to impure
substrate or enzyme, to the fact that more than one enzyme Depending of the value of this correlation coefficient, it is
acts on the substrate, or to different types of inhibition of the accepted that:
system [4].
1) Perfect r 2 = 1.0
D. Determination of enzymatic parameters using minimum squares 2) Excellent 0.9 ≤ r 2 < 1.0
Once the Michaelis-Menten linear equation is used, the 3) Good 0.8 ≤ r 2 < 0.9 (11)
minimum squares method is applied, which tries to apply the 4) Medium 0.5 ≤ r 2 < 0.8
best curve upon experimental data. The method involves the 5) Bad r 2 < 0.5
sum of squares of the calculated distances between the ordinate
at an experimental abscissa (x) of the line are polynomial
function P (xi) and the real function f(x) of the experimental E. Part III (Session 3)
data is minimum, i.e. [5]: In this section, kinetic parameters are calculated.
m m
F. Program flux diagram
∑ ⎣⎡p ( x ) − f ( x )⎦⎤ =∑ d
2 2
(6) Before doing the program, a blocks diagram will be
i i i
i =1 i =1 designed, which will help to have a clear idea of the data
needed, of the following steps, and of the desired results. The
Substituting (4) in (6), we observe that there is a family of flux diagram is written in order to be translated in any
linear curves that pass through the points, nevertheless programming language. An example is given in fig. 2.
obtaining the minimum of (6) using the derivate of the Note: this section has greater complexity and the students
function respecting to each of the variables (ao and a1) and develop the cognitive process in accordance to the
setting to zero, a system of linear equations of the following mathematical sequential logic obtained in his/her academic
form is generated: trajectory. In this step, the student constructs new knowledge.
m m
G. Example of developed software
mao + ∑ xi a1 = ∑ f ( xi ) Once the flux diagram is obtained in order to calculate the
i =1 i =1
(7) kinetic parameters of Michaelis-Menten equation, it is
m m m
translated in MATLAB 7.6 language. An example of what can
∑xa
i =1
i o +a1 ∑ x = ∑ f ( xi ) xi
i =1
2
i
i =1
be achieved by students is given, in the format of an
16 REGALADOǦMÉNDEZ ET AL.
executable file, in order to be able to run the program without Square Minimun Method
the use of MATLAB 7.6, as shown in fig. 2. 35
Experimental Date
Linear Adjusment
30
Begin
1/V = 237.3456(1/[S]) + 7.0887
1/V(mM/minute)
20
Read the data from File
datosbioq.dat
15
Subprogram minimum
5
0.01 0.02 0.03 0.04 0.05 0.06 0.07 0.08 0.09 0.1
1/[S](mM)
Print Km, Vmax, r^2
In screen
program from MATLAB 7.6. A window opens that explains Experimental Data
what the program does, as observed in fig. 3. 0.1 Adjusment
V = (0.1414*[S])/(33.4821+[S])
0.09
0.08
V(mM/minute)
0.07
0.06
0.05
0.04
0.03
10 20 30 40 50 60 70 80 90 100
[S](mM)
iv. PBL improved their generic skills. educational role. This supports an updated commitment
v. It makes us feel more comfortable to work in a group to the search for learning alternatives, as well as a
to solve problems. responsibility of institutions to invest in teaching courses
vi. They said that “It really makes us work for ours for the improvement of the educational processes.
studies and we feel intelligent doing it.” ACKNOWLEDGMENT
vii. We feel like engineers when we are doing the group
assignments. The authors wish to thank to colleagues Kathleen Meredith
viii. The given problem excites the students’ curiosity to Brown, Kathleen Dooley, Nina Carol Buck and Aitor Aizpuru
know more. for theirs revisions.
ix. It can be described as a double edged sword, because REFERENCES
we are required to do more than with traditional [1] Dean, R., L., “Laboratory Exercises: Kinetic Studies with Alkaline
learning methods education. Phosphatase in the Presence and Absence of Inhibitors and Divalent
Cations”, Biochemistry and Molecular Biology Education, 2002, Vol. 30
VI. CONCLUSIONS No. 6, pp. 4021-407.
[2] Hemberg T. “Abscising II as an inhibitor of α-quimoptripsinae”, Act
• Once the students concluded each step of the work Chem. Scand., 1967, Vol. 21, pp. 1666.
program, the facilitator of this activity notes a cognitive [3] Magued, I., Innovate Techniques in Instruction Technology, E-learning,
improvement in the global understanding of the involved E-assessment and Education, 2008, pp. 602.
discipline, and the objective is satisfactorily attained. [4] Montgomery R. and Swenson C. A., “Problemas Cuantitativos de las
Ciencias Bioquímicas”, Acribia, España, 1975, pp. 179-186.
• The PBL method is a strong tool to develop the creativity [5] Nieves H. A. and Domínguez S. F., “Métodos Numéricos Aplicados a la
of students, as the prime objective of universities is to Ingeniería”, Continental, México, D.F., 2003, pp. 362-364.
produce critical students. [6] Offmann, B. and Cadet, F., “Problem-based Learning: Redox-active
Disulfides in a Plant Light Switch”, Biochemistry and Molecular Biology
• If the objective was attained for the students which had Education, 2002, Vol. 30 No. 4, pp. 249-254.
the PBL to solve, it is worth noting that the degree of [7] Prausnitz J. M., “Hacia el fomento de la creatividad de los estudiantes”
understanding depends as well of the solidity of the Educación en Química, 1997, Vol. 5 No. 3, pp. 156-159.
knowledge background and the abilities acquired during [8] Savin-Baden, M., “Problem-based Learning: Disciplinary Differences or
Modes of Curriculum Practice? Who promised to deliver what in
the academic curriculum, and of the availability of problem-based learning”, Biochemistry and Molecular Biology
specialists and the leadership of the facilitator involved. Education, 2003, Vol. 31 No. 5, pp. 338-343.
• The use of PBL methodology can make the difference in [9] White H., B., “Problem-based Learning: Commentary Problem-based
the preparation and formation of students, and this is why Learning and Becoming a Physician”, Biochemistry and Molecular
Biology Education, 2003, Vol. 31 No. 5, pp. 337.
it remains important that professors assume their
Adaptive Assessments using Open Standards
Héctor G. Barbosa León, Francisco J. García Peñalvo, María José Rodriguez-Conde
University of Salamanca, Science Faculty, Plaza de los Caidos s/n, Salamanca, Spain 37008, Technological Institute of Colima, Av. Tecnológico s/n, Villa de
Álvarez, Colima, México 28976
[email protected], [email protected], [email protected]
Abstract - This article presents the conclusions of a research According to the Australian Flexible Learning Framework
workpreparedin the University of Salamanca, Spain, including [3]assessment, especially when is included within a real
the investigation, definition, construction and test of a model to learning task or exercises, could be an essential part of the
define adaptive assessments considering the user´s presentation learning experience, giving to the entire Web site the
needs and also the traditional adaptation process using the level
of complexity for every item. We applied innovatively some
characteristic to adapt itself to the needs of the users.
common open standards and specifications. Conceptualizing the learning process to its basic elements,
we can identify at last the following elements:
• The educational material to be taught by the teacher in a
I. INTRODUCTION classroom.
Students has been adopted the Internet as their first source • The teaching and learning activities that take place in a
of information. The Web and its intrinsic characteristics allow classroom.
the content to evolve from a static to a dynamic organization • The assessment activity to measure the student learning
by including adaptive processes, making the user´s experience and,
to be improved and enhanced [1]. This content, developed in • The report of the score results given by the teachers to the
form of Units of Learning (UoL), include one or several students. This conception is well suitable for the
learning objects and their related multimedia material. traditional educative process.
But we must ensure that those developments have
characteristics like be standardized, compatible and
interchangeable, incrementing the characteristic of actual Web
applications that are cross-platform systems.To achieve those
characteristics, developers defined the Semantic Web, in
which the content is seen in a semantic view instead of been
seen in a syntactical one by using meta-data [2].
Among the huge amount of information in the Internet we
will focus ourselves in instructional and educational
institutions which have been using the Internet to incorporate
the information and communication technologies in the
learning and teaching processes in order to increase the
quality, efficiency and dissemination of education.
We want to emphasize the role of the assessment activity
inside the e-learning process, and how it can help to improve it
to all participants: students, teachers and content designers. FIGURE 1. IMPORTANCE OF ASSESSMENT IN THE E-LEARNING PROCESS
II. IMPORTANCE OF THE ASSESSMENT ACTIVITY IN THE
EDUCATIVE CONTENT In figure I, we characterize the importance of assessment
in the learning flow. As we can note, the tests and evaluations
Traditionally, assessment activity has been seen like task not only are an integral part of the learning process, but also is
aside of the e-learning process and there is a danger in an element that complete and close a circular activity,
focusing research on assessment specifically [3], as this tends contributing as a feedback source for: the users (giving the
to isolate assessment from teaching and learning in general. scores and feedback), for the instructors (by giving support
For us, the assessment activity after the student took the and feedback) and for the instructional designer (to update the
educative lesson, could be seen as an element that closes and contents of the learning system) as well. This circular
complete a circular activity, being an integral part of the conceptualization of the learning process allow us to see the
learning process as well [4]. In addition, the assessment significance of the assessment because it helps to the
activity inside the e-learning process could be used to adapt adaptation of the system by setting a new user knowledge
the system by setting a new user knowledge level, evaluate, level, evaluating and determining new learning profiles, assign
and setting new learning profiles, assign user grades and, in the user grades and, in consequence, performing user content
consequence, performing user content re-adaptation [1].
19
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_4, © Springer Science+Business Media B.V. 2010
20 LEÓN ET AL.
re adaptation. This is how we see the importance of the Those specifications allow the creation of many types of
assessment task for the adaptation process. educational designs following a consistent notation to develop
Conclusively, we would like to mention an assessment a system that could have a homogeneous implementation in
categorization: several courses or learning contexts, giving the systems extra
• Formative assessment: this is an assessment that helps to characteristics like compatibility and exchangeability. In
give a convenient feedback and motivation to the student addition, those specifications define an abstract data model
and do not have scores. Also brings convenient feedback using XML, supporting the deployment of item banks and
to designers of materials. well documented content format.
• Summative assessment: this is a scored assessment and The potential benefit of Computer Based Test (CBT) are
gives a result to measure outcomes and success of the that a test item can attach various multimedia, such as graphic,
learner. audio or video, and reuse the test item easily, mostly if those
• Norm assessment: use the achievements of a group to set are designed and developed with open specifications and
the standards for specific grades and is used in most standards, allowing to different information from test result
universities. from students can be collected and recorded by computers
• Criterion assessment: establish the standard (criterion) effortlessly and immediately in electronic portfolios (e-
and mark students against it. portfolios).
There is another kind of assessment called alternative But most of CBT are linear that makes all students take
assessment. Here, the integration of the assessment activity the same number of test items in a specified order during a
with learning processes and real-life performance as opposed testing section. Nevertheless, this approach seems to imitate
to display of “inert knowledge”. Know as authentic the paper-and-pencil (P&P) test because it merely presents the
assessment it is very much based on the constructivist test in a digital format.
approach that enables students to demonstrate knowledge by IV. CHARACTERISTICS OF ADAPTIVE SYSTEMS
working on authentic task, putting students undercontrol of
their own learning [3]. In contrast to the linear test with fixed test items,
Computer Adaptive Test (CAT), based on Item Response
Theory (IRT) the presentation of each test item and the
III. CHARACTERISTICS FOR INTERNET-BASED SYSTEMS
determination to finish a test are dynamically adopted based
Nowadays, learning systems in the Internet must be on the student´s answers.
flexible and efficient, and one way to accomplish that is to be One aim is to make those systems to work in adaptive
an open and standardized system. The following standards learning systems and, given the fact that the assessment
give to those systems some of the mentioned characteristics: activity is an important and integral part of the e-learning
• The LTS (Learning Technology Standard) process; it is desirable that this evaluation activity could be
specification is a group of agreements about the adaptable as well. If we want the assessment to be
characteristics that a learning element should have. interoperable, compatible, and shareable, we must have to
The use of standards ensure instructional develop a standardized tool [8].
technologies to work with other systems The implementation of the adaptability in the assessment
(interoperability), follow-up information about activity could be performed by giving the instructor:
learners and contents (manageability), generate • Tools to design the assessment with several types of
objects that are usable in other contexts (reusability) resources (text, graphics, video, animations) to improve
and avoid obsolescence (durability) [5]. the student´ understanding in the examination
• IMS Specifications1 are part of the LTS, developed in • Design different types of assessment
1997 as a project of the National Learning • Define content structures and groups of students
Infrastructure Initiative at Educause. Its mission is to • Manage the assessments and the questions linked to each
promote distributed learning environments. For the assessment
IMS many areas require interoperability when • Define the grades for each question and assessment
learning is distributed, thus it details a set of • Define and manage the schedule for the assessment that
specifications that constitute a framework to will be taken.
interchange educational elements [6]. The purpose of our work was to adapt the characteristics
• Inside IMS specification one with special interest for of the questions that are included in one individual’s exam to
our investigation is the IMS QTI (Question and Test their learning style [4],[5]. To do this, the IMS QTI
Interoperability), for sharing test and/or items specification allows the author to link the multimedia
between compatible LMS. It defines a data model for resources that could fit for three different user needs: visual,
the presentation of presentation of questions and test auditive or kinesthetic. That categorization fits with one of the
and the correspondent results. learning styles definition2.
1
http://www.imsglobal.org
2
http:///www.ldpride.net/learningstyles.MI.html
ADAPTIVE ASSESSMENTS USING OPEN STANDARDS 21
Our aim was to define the primary resource (that is, the author in the definition stage, accordingly to his/her personal
material presented to the user by default) and the equivalent criterion.
resource (presented to the user taking into account his/her One of the main characteristics of the IMS LD (Learning
learning styles, physical disabilities, or preferences). For users Design) specification is its potential to define adaptive
with disabilities the user interface must be capable to make a characteristics covering the student preferences, prior
transformation of the display using flexible controls. This knowledge and/or learning needs. To do this it is necessary to
adaptation is performed by matching the characteristics use this specification in the level B to define individual
defined in the assessment item meta-data with those defined in interactions because in this level we can use some elements
the user profile in the Learning Management System. like <properties> and <conditions>.
Adaptation is another way to adjust to the user needs. This
process is performed when the student is answering the
questions, following the next sequence [9]:
• At the authoring stage, all items are categorized to
estimate their level of difficulty.
• When the user starts the exam, he/she is presented with a
pretest question to estimate his/her ability.
• The “best” next item is administered and the student
responds.
• A new ability level is computed based on the responses of
the student.
• Steps 3 and 4 are repeated until a stopping criterion is
met.
Questions would be given randomly to eliminate cheating.
The new ability level is computed based on equations
described in the figure 1, used in the MANIC [10] system
during the quiz sessions to compute the new estimated ability
level.
The equation (1) is for correctly answered questions and
the equation (2) is for wrongly answered questions.
V. ADAPTATION PROCESSES
• As separate single items that could be referenced by this group into two: for the first group we applied the test with
compatible external tools using the IMS QTI questions and accompanying multimedia material that matches
specification. their preferences of presentation (audio, text or video); for the
• As complete test that could be used in a session by LMS second group we applied a test that did not match these
as a part of a learning activity. preferences. Also, we applied a test a test to determine the
• As a learning object referenced from a learning unit learning style of each student, prior to the adaptive test.
defined using the IMS LD specification in which some Please, recall that this test also perform an adaptation in
conditions are defined to select the testPart using the the level of complexity of each item, for this we organize the
value stored in the user model. items grouping them according this characteristic and, in a
The elements to perform the adaptation are: higher level, we set a classification according to the
• User preference: stored in the user model with values of multimedia material they are refereeing to.
“visual”, “audio” or “text. The aim of this activity is to evaluate our thesis
• Result of the first normal-level question stored in the hypothesis that, if a student is presented with an adaptive test
unifiedResult variable, with values of “Y” or “N”. This that matches its preferences of presentation of the
variable is special because look to unify the result for the accompanying multimedia material he/she could average
different kind of questions (simple choice, multiple better results in a test.
choice, etc.). The evaluation included the following statistical test:
• The result stored in the following questions in the variable • Final note: To check if the final note of the test is higher
unifiedResult, with values “Y” and “N”. for the students with adapted test.
In the authoring process the author select from the list, • Time to response: To check if the total time to answer the
those items that will be included in the test. The tool shows the test is lower when an adapted test is presented to the
items grouped by the multimedia files referenced by them. student.
Once the selection is made, the tool construct the XML-QTI • Level of satisfaction: To check if the level of satisfaction
file with the testPart, assessmentSection sections and the reported by the student is higher when the test is adapted
reference to the external QTI files for each question and to his/her preferences.
references to the external multimedia files as well. At the end, For the final note test the results were: the average final
a file with the IMS CP instruction is generated and packed as note obtained by the students with an adapted test was 6.6
.zip format; this file is the one that will be used by the LMS. points, whilst for students without adapted test, the final note
Also in the authoring process is possible to define the was of 5.9 points (t test = 0,836).
following kind of test: In the time of response the results were: the average time
• Sequential test: This test does not include branch rules, for an adapted test was 17 minutes whilst for an not adapted
presenting the questions in the sequential order they were test was 20 minutes ( t test=0.964).
selected from the list. The level of satisfaction was obtained from the students
• Random test: This test does not include branch rules, but themselves from a poll after they answer the test. It cover level
the ordering label is set to shuffle. of satisfaction from 1 (lower satisfaction) to 5 (highest
• Adapted test: This test includes the adaptation rules with satisfaction). The results were: for an adapted test 3.9 and for
the branchRule and preCondition labels to control the not adapted test 3.6 (t test=0.821).
items presented to the users. Bring to mind that, the statistical test was applied only to
verify the results for the adaptation process in the presentation
VIII. ADASAT TOOL of the multimedia material, but this test also performs the
adaptation in the complexity level for all participants.
We develop an authoring tool to construct items and test
Also, we think that these results must be taken with some
with adaptation characteristics, following open standards and considerations at the time to make generalizations. In other
specifications like the IMS QTI, CP and LD and the XML words the results obtained in this work may change in other
language. The aim is to allow to the author to define items and
learning environment and stated that this is an internal work.
test using a graphical interface starting from the definition of Upon this, we consider that our conclusions are a science
single questions to the construction of complete test packaged contribution for future works in this area.
using IMS CP and delivered in a .zip package [13], [14].
The application covers the following activities: (1) X. CONCLUSIONS
Question Construction management, (2) Test Construction
management, (3) User management, (4) Test management. Online assessment is an important step inside the e-
learning process, helping to improve the learning and teaching
IX. RESULTS OBTAINED experience.
According to the new developments in the area of e-
To verify that an adapted test is better for the students in learning, we can see that most of these developments look to
terms to ease their answer and understanding, we apply an be compliant with accepted standards like the IMS QTI. This
exam to evaluate the level of knowledge in English language gives the convenience to those works to be interoperable and
to a group of students in the University of Salamanca. We split adaptable to several learning platforms. In concordance,
24 LEÓN ET AL.
referring to the assessment activity, we can think that it must [3] Booth, R., Berwyn, C. The development of quality online assessment in
vocational education and training.Australian Flexible Learning
be interoperable as well, because it is one element of the Framework, Vol.1, 2003, 17.
e-learning process and play an important role inside this task.
Adaptability is another key factor in assessment. Given [4] Barbosa, H., Garcìa, F.,An Authoring Tool to Develop Adaptive
Assessments.In proceedings of 2006 Webist Congress, Setubal,
the fact that assessment is an important element of the e- Protugal, 2006.
learning process and the fact that this process look to be
interoperable, then we can think that the assessment tool could [5] García, F., Berlanga, A., Moreno, M, N., Carabias, J. “HyCo and
authoring tool to Create Semantic Learning Objects for Web-based E-
be used with different educative content administrators with Learning Systems”.In Web Engineering Fourth International
different conceptualizations and ways to design an ways of Conference, ICWE 2004. Published in N.Koch, P. Fraternali, M.
purposes to the students. To face this situation it is necessary Wirsing (Eds.). Series Lecture Notes in Computer Science. Springer
to develop a test with different types of resources, different Verlag. Vol. LNCS 3140, pp. 344-348.
kind of assessments, groups of students, kind of questions, [6] Berlanga A., Garcia, F., “Towards Semantic Metadata for Learning
etc. Elements”. In proceedings of the fifth. International Conference on
Information Technology Based Higher Education and Training ITHET
Under this conceptualization, we created and applied an 2004. Y. Akpinar (Ed.) (2004) pp. 572-577.
adaptive test to a group of students to evaluate the hypothesis
that, if an assessment tool is adapted to the user`s needs or [7] Barbosa, H., Garcia, F. “Setting and Sharing Adaptive Assessment
Assets”. In proceedings of the SIIE congres, Leon Spain, 2006.
preferences we can reduce the time to answer the test while
we increase the level of satisfaction of the user at the same [8] Barbosa, H., García, F. “An Authoring Tool for Adaptative Assessment
Items”. In proceedings of IAIRA Congress. Bucharest, Romania, 2006.
time.
[9] Barbosa, H., Garcia, F. “An Authoring Tool to Construct Adaptive
ACKNOWLEDGMENT Assessments”. In proceedings of the 2006 EIAE Conference.Online,
2006.
We want to thank to the group KEOPS (ref. TSI2005-
00960) of the University of Salamanca for their contributions [10] Mia,k., Stern, Woolf, B. “Curriculum sequencing in a Web-based tutor”.
Lecture Notes in Computer Science, 1452, 1998.
and ideas for the development of this work. In addition, we
want to thank to the Education and Science Ministry of Spain, [11] Barbosa, H., Garcia, F., Rodriguez, M. “Defining Adaptive
Assessments”.In proceedings of the Webist Conference, Barcelona,
National Program in Technologies and Services for the Spain, 2007.
Information Society, Ref.: TSI2005- 00960.
[12] Barbosa, H., Garcia, F. “Constructing Learning Objects for Adaptive
Assessments”. In proceedings of IASTED conference, Chamonix,
HéctorBarbosa thanks the National System of Technological France, 2007.
Institutes (SNIT–Mexico) for its financial support.
[13] Barbosa, H., Garcia, F., Rodriguez, M. “A Tool for Online Assessment in
REFERENCES Adaptive e-Learning Platform”.In proceedings of the eUniverSAL
conference. Virtual, Salamanca, Spain. 2008.
[1] Barbosa, H., Garcia F. Importance of the online assessment in the
elearning process. 6th International Conference on Information [14] Barbosa, H., Garcia F., Rodriguez, M. “Construction of Assessments
Technology-based Higher Education and Training ITHET & IEEE, with Double Adaptation Processes”. In proceedings of EIAE Congress,
Santo Domingo. Virtual, 2008.
[2] Barbosa, H., García, F., Adaptive Assessment Tool for the Hypertext
Composer”, ICTE2005. In proceedings of the 3th.Internatonal Conference
on Multimedia and Information and Communication Technologies in
Education: m-ICTE 2005, Cáceres, Spain, Vol.l 1, pp. 16-20.
Online Academic Advising Support
Fawzi Albalooshi1 and Safwan Shatnawi2
College of Applied Studies
University of Bahrain
Sakhir Campus, Bahrain Kingdom.
1
[email protected], [email protected]
Abstract- The paper addresses academic advising – an and graduation support system is explained in Section 4 and
important issue that is often not given enough attention by the timetabling support system is explained in Section 5.
students and advisors alike. An online advising system is
presented that can be utilized by students, advisors, and course 2. LITERATURE SURVEY
timetable planners. Students are given informative advice on
which courses to register for in the next semester and are
Although the system described in this paper was primarily
informed of their remaining graduation requirements; advisors motivated by our need to optimize advising, to determine the
are able to see students’ progress towards their graduation courses to be offered and to utilize existing information
requirements; and timetable planners are given statistics on available in the university registration system, there is no
courses and sections requirements for the coming semester. doubt that academics worldwide agree that proper advising is
an important factor for students’ successful progress in
higher education. Many studies have been conducted to
Index Terms- Academic Advising, Automated Advising, Course confirm this matter. Examples of such studies are reported in
Scheduling, Proactive Advising [1] and [3]. As a result, many academic institutes investigated
the use of computer technologies in academic advising to
1. INTRODUCTION overcome the difficulties experienced with traditional
Student advising is one of the important challenges facing methods of advising. A sample list is as follows:
academics, but direct consultation between instructors and
students is another problem. Many students do not take the Reference [1] present a systematic view identifying key
time and effort to see their advisors to plan their timetable subsystems grouped under basic study planning and high
before registration resulting in many registration issues and level advising that can be automated to the benefit of advisers
long queues for advising at registration time. Another and advisees. Under basic study planning, the authors suggest
problem facing academic institutes is planning ahead for the that a number of basic queries be available such as course
courses to be offered for the coming semester and availability, courses in which a student can register, degree
determining the number of sections for each course. This completion requirements, degree credit transfer possibilities,
difficulty is due to the fact that exam results for the courses and time constraints. Reference [2] reports a system that
in which students are registered in the current semester are offers advisees up-to-date online advising and related
not available until exams are over at the end of the semester. information including a recommended list of courses in
Determining students’ graduation status is another time which a student must register in the next semester in order to
consuming effort for academics worldwide. complete his degree requirements. The system has a web-
based main page through which system users such as
In this paper we present a system that accesses students’ students, faculty, and administrative staff are allowed access
transcript records and generates a report for each student to their respective sites. Reference [3] presents the
indicating the courses in which he must register in the motivations to develop FROSH [4] the automated advisor for
coming semester based on rules explained later in the paper. freshman. Reference [5] presents a collaborative course
Such details can only be exact towards the end of the recommender system that recommends elective modules to
semester once grades are input into the registration system. students based on the core modules they have selected. A
During the course of the semester, the proposed system can “More Like This” recommender offers students similar
generate the students’ requirements for courses including the modules to their first choice of electives in cases where no
number of sections required for the coming semester. This places are available or where timetable clashes occur.
will enable the timetable committee to decide on the courses Reference [6] presents InVEStA, an expert system for
and sections to be offered and prepare the most suitable advising students on the courses in which they must register
timetable for the students. Once grades are entered into the next. The recommender system also generates a semester
system at the end of the semester, a more accurate list of schedule based on the courses offered for the semester and
courses and sections can be generated before the start of the the student curricula. Reference [7] presents a web-based
new semester. The proposed system helps students in interactive student advising system for course plan selection
advising, graduation, and timetabling. The students’ advising and registration using Java framework.
25
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_5, © Springer Science+Business Media B.V. 2010
26 ALBALOOSHI AND SHATNAWI
TABLE 1
STUDENT INFORMATION TABLE EXTRACTED FROM THE REGISTRATION SYSTEM.
3.1 Text Processing Algorithm
Field Description Data type Domain Example The algorithm used to process the students' transcript file
follows the same steps used in third generation languages
Name Student name string A-Z FAISAL ALI
AHMED AJAJ
compilers. It scans the file character by character, identifies
tokens, and takes actions according to the recognized token.
ID Student's ID Number integer 20081234
Pseudo code
Degree Students string A-Z Office
program (major) management 1- Open the transcript's file for input
S_Cr_hrs Semester Credits number 0-21 9 2- while not end of transcript file do
Attended
3- read the file character by character and identify
C_Cr_hrs Cumulative number 0-180 30 tokens
Credits Attended
4- check if the token is one of the following identifiers:
S_GPA Semester GPA number 0-4 3.2
[course , GPA , name , status, degree, year,
GPA Cumulative GPA number 0-4 3.5
semester, passed]
TABLE 4
TABLE OF COURSES AND THEIR PRE-REQUISITES
31
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_6, © Springer Science+Business Media B.V. 2010
32 ARIFFIN ET AL.
software used was Game Maker, Game Maker Language scenario, the fire will not fade away and deteriorate. The
(GML) Anime Studio, Adobe Photoshop, Audacity, PHP, FireHazard game is purposely designed to simulate three
MySQL and Apache HTTP Server. different scenarios to maintain consistency. Scenario A
depicts a situation of a Class A type of fire which caused
by ordinary combustible materials such as rubbish and
A. Simulation game architecture paper. Player must decide what to do in this situation and
Figure 1 illustrates the architecture of FireHazard. It which fire extinguisher he must use to put off the fire
consists of four main components which are the game, based on the clue or hint given in the scenario. Scenario B
knowledge repository, scoring system and score database. of the FireHazard game depicts fire in a kitchen caused by
Class A and Class C type of fire. Player should be able to
identify and use fire extinguisher two different fire
extinguishers to put off the fire. Player must think
instantly as the fire is spreading fast and rapidly. Scenario
C illustrates another different scenario which in this case
in a chemical laboratory caused by Class D type of fire. If
player use incorrect type of fire extinguisher, the fire will
grow bigger and uncontrollable.
Component Description
Knowledge Consists of tutorials and video clips about how
Repository to put out fire using the correct techniques. The
players have to read the tutorials and watch the
videos before playing the game.
FireHazard game The main portion of the FireHazard application.
The game consists of three levels depicting three
different scenarios. FIGURE 2: Game flow of FireHazard
Scoring System The scoring system receives the scores from the
game and calculates the total score using various
pre-determined formulas.
Score Database Stores the score of all players which consists of
the subtotal score for each level and the total D. Players
score for all three levels.
25 players were chosen for this study. These players
were selected from Health and Safety department,
B. Game Description lecturers and students who have no experience in playing
FireHazard is a fire-fighting computer game with the the FireHazard game but have prior experience in playing
aim to educate players on how to react in the case of fire. other computer game.
Players should be able to identify the type of fire
extinguisher to be used to put off different type of fire. If
players failed to use the correct fire extinguisher for the
34 ARIFFIN ET AL.
V. CONCLUSION
REFERENCES
[1] Peter F. Drucker, Post Capitalist Society, HarperBusiness, p 8,
1993
[2] Yozgat, U. Yönetim Bilişim Sistemleri. İstanbul: Beta Basım
Yayım, 1998.
[3] Conner, K.R., Prahalad, C.K.. “A resource-based theory of the
firm: knowledge versus opportunism”. Organization Science. 7/5:
477-501. 1996
[4] Nahapiet, J., Groshal, S. “Social capital, intellectual capital, and
the organizational advantage”. Academy of Management Review.
40/2:242-266, 1998
USING COMPUTER GAME TO SHARE KNOWLEDGE 35
[5] Flaherty, M. Global Operations Management, McGraw-Hill, [18] P.S. Goodman, E.D. Darr, "Computer-aided systems and
Singapore 1996 communities: Mechanisms for organizational learning in
[6] Yip, G. Total Global Strategy, Prentice Hall, New Jersey. 1992 distributed environments", Management Information Systems
Quarterly, Vol. 22 pp.417 - 440. 1998
[7] Tiwana, A. The Knowledge Management Toolkit, Prentice-Hall: [19] Ruben, B.D, Simulations, Game and Experience Learning,: the
Saddle River, 2000. quest for a new paradigm for teaching and learning, Simulation
[8] Elias M. Awad, Hassan M. Ghaziri, Knowledge Management and Game 30(4) 498-505,1999
Upper Saddler River, New Jersey : Pearson Education, Inc., 2004 [20] Abt, C.A Serious games. New York: Viking, 1970.
[9] Applehans, Globe, Laguero. Managing Knowledge: A Practical [21] Gredler, M.E, Educational games and simulations: A technology
Web-Based Approach. 1999 in search of a (research) paradigm, In D Jonassen (Ed.)
[10] Polanyi, M. The Tacit Dimension, Routledge & Kegan Paul, Handbook of Research for Educational Communications and
London. 1966. Technology (pp521-540). New York: Macmillan 1996
[11] Nonaka, I. and Takeuchi, H. The Knowledge-Creating Company, [22] Driskell, J.E & Dwyer, D.J Microcomputer videogame based
Oxford,University Press, New York. 1995 training. Educational Technology, pp11-16. 1984
[12] Snowden D “ A Framework for creating a sustainable [23] Christopher, E. M. Simulation and games as subversive activities,
programme” CBI Guide to Knowledge Management, London Simulation and Gaming, pp 441-455.1999.
Caspian Publishing/Confederation of British Industry 1998 [24] Schank, R. & Neaman, A Motivation and failure in educational
(Republish in Knowledge Management Year Book 1998 simulation design. In K. D. Forbus & P.J Feltovich (Eds), Smart
Butterworth Heinemann, April 1999 machines in education: The coming revolution in educational
[13] Argote, L., Ingram, P., Levine, J. & Moreland, R. “Knowledge technology. Cambridge, MA, pp. 37-69, US: the MIT Press.2001
transfer in organizations: learning from the experience of [25] Washburn, J., Gosen, J. An exploration of game-derived in total
others,” Organizational Behavior and Human Decision Processes, enterprise simulations. Simulation & Gaming, pp 281-296, 2001.
82(1), 1-8. 2000
[26] Baker, E.L, & O’ Neil, H. F. Jr. Evaluation and research for
[14] Goh, S.C “Toward a learning organization: the strategic building technology: Not just playing around Evaluation and Program
blocks”, SAM –advanced Management Journal, Vol. 63 No. 2, pp. Planning, pp 169-176. 2003
15-22. 1998
[27] O’ Neil, H. F. Jr. , & Andrews, D. (Eds). Aircrew training and
[15] Garvin, D., “Building learning organizations”, Harvard Business assessment. Mahwah, NJ: Lawrence Eribaum Associates. 2000
Review, July-August, pp. 78-91.1993
[28] Mayer, R. E., Mautone, P.m & Prothero, W. Pictorial aids for
[16] Chai, K H, Y S Shi and M J Gregory, “Bridging islands of learning by doing in a multimedia geology simulation game.
knowledge: A framework of knowledge sharing in international Journal of Educational Psychology, pp 171-185,2002.
manufacturing networks In Managing Operations Networks”,
Proceedings of the EurOMA 1999 Annual Conference, 7-8 June [29] Rosenorn, T., & Kofoed, L. B. Reflection in Learning Processes
1999, Venice, Italy, edited by E Bartezzaghi et al., pp. 853-860, through simulation/Gaming. Simulation & Gaming, pp432-
1999 440,1998
[17] Goh, S.C, “Managing Effective Knowledge transfer:an integrative [30] Westbrook, J.I & Braithwaite, J. The health care game: an
framework and some practice implications”, Journal Knowledge evaluation of heuristic, web based simulation, Journal of
Management, Vol 6 No 1. pp23-30. 2003 Interactive Learning Research , pp89-104,2001
Following the Paper Trail:
Measuring the Economic and Environmental Impact
of Digital Content Delivery
Vanessa K. Cox
Ruth C. May
Stanley L. Kroder
Geralyn McClure Franklin
University of Dallas
College of Business
1845 E. Northgate Drive
Irving, TX 75062
37
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_7, © Springer Science+Business Media B.V. 2010
38 COX ET AL.
cannot be managed effectively, much less used for strategic digital content for the COB consisted of 92 online course
advantage. sections for spring and summer 2009 combined.
The Graduate School of Management at UD, now part of 1) Digital Content: Calculating the total number of virtual
the College of Business, has been involved in distance pages required establishing a definition of what constitutes a
learning since 1969 beginning with a regional microwave virtual page as well as specific guidelines to ensure a
video network. As the Internet emerged in the 1990s, the consistent, accurate page count. We defined a virtual page
school leveraged the new technology by launching three as either a) actual pages in a paginated document or b) the
online credit-bearing courses. This initial offering has equivalent number of pages for scrolling digital content in a
evolved over the past decade to include the full Masters of Microsoft Word document. The following items were
Business Administration (MBA) core curriculum and a considered when counting the virtual pages in each online
dozen specialized MBA concentrations, with no residency course.
requirement. The expansion of the college’s online delivery
has provided an opportunity to reduce costs associated with • Definitions: The total number of virtual pages includes
traditional paper and printing content production and to obvious printable downloads, nonobvious printable
reduce the carbon footprint of the institution. In early 2009, digital content and nonprintable digital content.
a pilot study was initiated to document these savings which Examples of obvious printable downloads include
are a natural byproduct of online education. The following syllabi and XanEdu (vendor) course packs.
section describes the process used to account for digital Nonobvious printable digital content includes items
content not printed and the resulting savings in dollars and such as chat sessions as well as directives to students in
carbon emissions. digital format. Nonprintable digital content includes
grade book and exam items protected by plug-ins that
III. METHODOLOGY prohibit the printing of designated content.
Students were also asked “What type of printer do you B. Economic and Environmental Consequences
typically use for printing course related materials?”. They Providing digital content versus traditional paper and
could select one of the following: small desktop printer, printing delivery results in trees not destroyed to produce
shared workgroup within their company or work group, or a paper, energy not used to print content on paper, and carbon
business center printer such as Fed-Ex Kinkos or a hotel. dioxide not emitted in the printing process. Collectively,
these outcomes result in economic savings and
IV. RESULTS environmental benefits for the planet. To quantify these
benefits, we relied on the Hewlett-Packard (HP) carbon
A. Digital Content footprint calculator.
Table I summarizes the virtual page count in all categories:
obvious printable, nonobvious printable, nonprintable and HP has a strong focus on global citizenship and
total virtual pages. For both terms combined, the college sustainability with corporate goals that pertain specifically
provided over 1.5 million pages of digital content. to environmental outcomes [6] [7]. Its printer division is a
However, the net number of virtual pages that remained in leader in the imaging market and has a “green” focus in its
digital format (i.e. were not printed by students) could only product and delivery. The HP Carbon Footprint Calculator
be derived by applying the data captured on student printing is designed to assess printer energy consumption and
behaviors. Specifically, students in the spring 2009 term associated carbon output as well as quantity of paper used
printed an average of 53% of obvious downloadable pages and related monetary costs. The calculator generates
and 30% of nonobvious printable content. Students in the estimates of energy consumption during use of a printer and
summer 2009 term printed an average of 49% and 32% in emissions of carbon dioxide from production of estimated
the same categories, respectively. volumes of paper consumed during printing [8]. The
savings derived by the calculator provided explicit figures
The net results for the institution are illustrated in Fig. 2 for the study with the exception of the CO2 emissions which
where 543,275 pages remained virtual in the spring semester had to be converted from pounds of CO2 to a Metric Tonne
and 320,028 remained virtual in the summer term. Thus, a Carbon Equivalent (MTCE), the standard unit of CO2
total of 863,303 pages of digital content provided by the emissions that we adopted for this study [9].
institution remained in digital format, with their
corresponding economic and environmental consequences, Table II provides the gross institutional savings in
even after students’ printing behaviors were taken into kilowatts, tonnes of CO2 emissions, pounds of paper and
account for both semesters. dollars generated across all online courses in spring and
summer 2009. While the savings depicted in Table II are
TABLE I real savings from an institutional viewpoint, the printing
RUBRIC SUMMARY
FOR ALL ONLINE COURSES
habits of students must be taken into account in order to
determine the net real environmental benefits of the digital
content delivery. After accounting for both printing habits
University of Obvious Nonobvious Nonprintable Total
Dallas Printable Printable Pages Virtual
and types of printers used by students, the net results for the
Term Pages Pages Pages planet remain positive as illustrated in Table III.
Spring 2009 544,151 399,593 7,808 951,552
Summer 2009 421,757 147,722 5,381 574,860
Year to Date 965,908 547,315 13,189 1,526,412
TABLE II
CARBON FOOTPRINT CALCULATOR RESULTS
FOR TOTAL DIGITAL CONTENT
Term Year
Savings In: to
Spring Summer
2009 2009 Date
This paper reported on the results to date of a pilot study Despite these inherent limitations, the pilot program
currently underway in the College of Business at the demonstrates the value of considering economic and
University of Dallas to develop a framework for measuring environmental consequences of digital content delivery for
the economic and environmental impact of digital content the institution and the planet. Moreover, while the pilot
delivery. While many universities have stepped up efforts framework described here is only one approach to
to integrate sustainability concepts into their curriculum and measuring the sustainable impact of online learning, our
their on-ground campus operations, few [11], if any, have intention is to provide an initial blueprint for other
attempted to quantify the positive environmental effects that educational institutions to follow in:
their eLearning operations are generating while producing
cost savings as well. • Determining the volume of digital content they provide
and the corresponding economic benefit.
Using survey data from 1,039 online MBA students in the • Determining the offsetting effects of student printing
spring and summer semesters of 2009 and archival data behaviors on the institution’s digital content delivery
created via virtual page counts, we found positive economic strategy.
and environmental benefits from digital delivery for the first • Determining the combined effect of institutional and
eight months of the year. Cumulatively, digital delivery of student choices on the broader environment.
content resulted in 298,645 kilowatts not used, 196.47
tonnes of CO2 not emitted and 16,789 pounds of paper not University administrators, faculty and students have the
printed for a total of $37,341 saved by the University of opportunity to positively influence the economic and
Dallas. environmental dimensions of the TBL with regard to digital
content. Specifically, administrators can be more proactive
After taking into account the printing behaviors of online in adopting institutional strategies to increase online
students, which is a more conservative measurement offerings in various academic units and can mandate the
approach, the online operations of the COB still produced continuous tracking of benefits and costs associated with
environmental benefits totaling 65,487 kilowatts saved, digital delivery. Administrators can also consider
47.05 tonnes of CO2 not emitted and 9,509 pounds of paper establishing green “stretch goals” for their online operations
not printed. The savings in carbon emissions alone equates on both a gross and per capita basis.
to the following:
Students have a significant role in determining how much
of an institution’s digital content remains virtual as a result
MEASURING THE ECONOMIC AND ENVIRONMENTAL IMPACT 41
of their printing behaviors. Therefore, it is important to The authors also thank Jeff Borden, Senior Director of
inform students of the power that they have to reduce carbon Teaching and Learning for Pearson-eCollege. Mr. Borden
emissions simply by becoming more aware of how much has led his company’s global efforts to deliver high-quality
they print and how their printing decisions have direct online courses for hundreds of higher educational
environmental consequences. A leading edge example of institutions since 2002 and he provided valuable advice to
institutional influence on student printing behaviors is a our team during the design of the survey and data collection
pilot program involving six colleges and Amazon.com. The process.
partnership provides designated students at the pilot
universities with a Kindle DX and access to required course Finally, the authors acknowledge with grateful appreciation
readings. The e-book reader is designed to meet the needs the contributions of Nestor Via, Graduate Assistant in the
of students by offering a larger screen and searchable online learning business unit of the University of Dallas,
content embedded within required reading materials while College of Business. Mr. Via conducted the count of all
motivating them to reduce their printing of reading materials virtual pages in both spring and summer terms to ensure the
[15]. In general, today’s students are likely to be more integrity and consistency of the data.
environmentally conscious than their predecessors [16], but
they cannot contribute to a solution if they are not made REFERENCES
fully aware of the problem and their potential to influence
its outcome. [1] J. Elkington, “Towards the sustainable corporation: Win-win-win
business strategies for sustainable development,” California Management
Review, 36, no.2, pp.90-100, 1994.
Likewise, instructors should be encouraged to focus on [2] D. Ferrer-Balas, J. Adachi, S. Banas, C.I. Davidson, A. Hoshikoshi,
their roles in the sustainable efforts of their institutions by A.Mishara, Y. Motodoa, M. Onga, M. Ostwald, “An international
rethinking their design of course content in keeping with the comparative analysis of sustainability transformation across seven
universities,” International Journal of Sustainability in Higher Education,
“less is more” proposition. For instance, instructors should vol. 9, no.3, pp.312, 2009.
consider the following: How effective is pure text content [3] M. Porter, M. Kramer, “Strategy & Society: The link between
and how is it likely to be perceived by students, particularly competitive advantage and corporate social responsibility,” Harvard
those in the Gen Y demographic? Recent studies show that Business Review, pp 1-2, December, 2006.
[4] W. Adams, “The future of sustainability: rethinking environment and
Gen Y individuals prefer learning through multi-media and development in the twenty-first century,” The world conservation union,,
multi-tasking designs rather than traditional text delivery online, last updated May 2006, http://www.iucn.org.
[17]. With this reality in mind, instructors should consider [5] B-school facilities resource center, online,
what might be more effective pedagogical options relative to http://www.aacsb.edu/resource_centers/fcilities/buildinggreen.asp .
[6]HP Global Citizenship Report: Environmental sustainability, online,
text for increasing student engagement and reducing the September 24, 2009,
temptation to print. These options might include relevant http://www.hp.com/hpinfo/globalcitizenship/gcreport/envsustainability.
videos and student-directed content “scavenger hunts” based html .
around a central theme or problem which prompts students [7]Letter from Mark Hurd, HP Chairman, CEO and President,
http://www.hp.com/hpinfo/globalcitizenship/gcreport/intro/ceoletter.html
to focus on collecting and analyzing virtual content as [8] HP Carbon footprint calculator for Printing,
opposed to printing and memorizing basic text. This http://www.hp.com/go/carbonfootprint .
approach to content design is consistent with John Dewey’s [9] MtCO2- Metric Tonne Carbon Equivalent, online, 10/06/2009,
philosophy which is, “Give the pupils something to do, not http://www.mtco2e.com .
[10] Tonne, online, http://en.wikipedia.org/wiki/tonne .
something to learn; and if the doing is such a nature as to [11] R. Roy, S. Potter, “Designing low carbon higher education systems:
demand thinking; learning will naturally result [18].” Environmental impacts of campus and distance systems,” International
Journal of Sustainability in Higher education, vol.9, No. 2, pp. 117-118,
In conclusion, our results demonstrate the value of not just 2008.
[12] C.Dunn, “Reduce your carbon footprint by half in three steps: Plant
following the paper trail, but in documenting and measuring Green”, online, July, 2009, http://planetgreen.discovery.com .
it in order to better guide organizations in achieving [13] Emission Facts: Greenhouse Gas Emissions from a Typical Passenger
sustainable strategies. We hope that these initial efforts will Vehicle- US EPA, online, July, 2009,
prompt greater consideration and discussion of the potential http://www.epa.gov/OMS/climate/420f05004.htm .
[14] How your pollution is calculated – Fight Global Warming-
benefits of digital content delivery and the corresponding Environmental Defense Fund, online, July, 2009,
challenges involved in measuring them. http://fightglobalwarming.com .
[15] S. Peters, “Universities turn to Kindle- Sometimes to save paper,”
ACKNOWLEDGMENTS New York Times, online, July 30, 2009,
http://greeninc.blogs.nytime.com/2009/07/30/universityies-turn-to-kindle-
sometimes-to-save-paper/ .
The authors of the paper thank John Frey, Americas [16] J. Pryor, S. Hurtado, J. Sharkness, W. Korn, L. DeAngelo, L.
Sustainability Executive with Hewlett-Packard Company. Romero, S.Tran, “Political engagement among college freshmen hits 40-
Under Frey’s leadership, Hewlett-Packard has been named year high,” Higher Education Research Institute (HERI): The American
Freshman: National Norms Fall 2008, online,
one of the 100 most sustainable corporations in the world by http://www.gsei.ucla.edu/heri/pr-display.php?prQry=28.
the Global 100 each year since the inception of the list. In [17] C. Dede, “Planning for neomillennial learning styles: Shifts in
addition, HP continues to be listed on key sustainability students’ learning style will prompt a shift to active construction of
indices including the Dow Jones World Sustainability Index knowledge through mediated immersion,” Educause Quarterly, pp. 8-9.
2005.
and the FTSE4Good. Mr. Frey provided important input [18]S.Khurana,About.com,online,
and clarification related to the methodology of the Hewlett- http://quotations.about.com/od/stillmorefamouspeople/a/JohnDewey1.htm .
Packard carbon calculator and the paper and printing aspects
of the study.
Context Driven Personalized e-Learning Environment
Provided as a State Service
Deniss Kumlander
Department of Informatics, TTU
Raja 15, 12618
Tallinn, Estonia
e-mail: [email protected]
Abstract—The main topic of this paper is the life-long education. delivering it to students. Therefore the university approach
The rapidly evolving world permanently challenge both doesn’t work very well in many cases of the modern, rapidly
individuals and governments as the productivity and educational changing world. Constant technology, science, best practices,
level of the state work-force is one of the most important factors standards and so forth changes cause various types of
of the long term success in both financial and social sectors. The unemployment, low productivity of work-force since skills and
paper describes an e-education system that should be funded by knowledge become outdated very quickly. This is a serious
the state and exposed to all citizens in order to constantly challenge for all involved parties: society, business and
upgrade their skills. The broad potential auditorium and targets individuals and therefore it should be solved together in a
of such system define certain unique requirements that should be
synergy.
met by the context driven personalized approach of delivering
information as proposed in the paper. It makes the process of The concept of the life-long education is the first step on
acquiring knowledge simpler for all categories of users and this road. Unfortunately in many cases it is left to individuals to
stabilizes the virtual learning environment from the data follow it, while others do minor investments into that mostly
protection point of view. motivating people to continue the education on constant basis.
In the result many people spent their time trying to find and
Keywords-e-education;artificial intelligence; life-long learning; acquire knowledge in parallel instead of following mentor
state services; context.
advices consuming it in a prepared form. Fortunately
everything started to change (so far slowly) and the importance
I. INTRODUCTION of an electronic learning as an alternate solution to the
There are two different approaches to the educational traditional education has been recognized. The quick
system in the modern society. One of those is university centric development of electronic channels and world globalization [6,
suggesting to concentrate only on educating young people. It 7] help us to organize now the re-education much better than it
proposes that correctly organized educational process gives sued to be in the past. Unfortunately those attempts mostly
enough skills to upgrade knowledge later on without mentors. local and lack the strategic, global power that could be provide
Here the state builds up centers of massive education for young by a state financing and driving the development of such virtual
people and fund those also as research facilities in order to education environment. At the same time, upgrading to the new
concentrate science in one place and produce this way highly level of virtual education means also facing completely new
skilled work-force, which is so much required by the modern challenges in process of building such systems. There are
world [1]. This approach has both positive and negative sides already millions of students attending e-courses in developed
[2, 3]. For example it was criticized as been mostly looking countries and this number is growing permanently. They come
into the past accumulating the knowledge in order to spread it from very different education levels, age and society groups.
without major updates of courses during sufficient periods of Therefore the online education based on average student
time [3]. Universities do argue that this is much more stable doesn’t work as well as it was hoped and requires
method of educating as proper understanding of fundamental personalization of the delivered content. Fortunately computers
science (basics, techniques) will give to students a good base and artificial intelligence give us just enough tools to deliver
for acquiring whatever additional knowledge in the future [5]. such personalized content as described in this paper.
Having agreed with it, we still like to highlight the sufficient
gap between the knowledge obtained in universities and the II. PERSONALIZED E-EDUCATION PROVIDED AS A STATE
knowledge students will need dealing with specific tasks at SERVICE
their future workplaces. Moreover the ability to acquire new
knowledge in the future doesn’t necessary mean that the person A. Auditorium
will do that. It could be a problematic or extremely time The kind of the system aimed in the paper has certain
consuming for him without having a good understanding of difference from other e-Education systems described a lot in
what is going on in the modern technology and having to parse the literature and the most important one is the targeted
a huge amount of information sources. Notice that the last auditorium. If the standard literature case is moving the
mentioned tasks are tasks of the teacher in normal educational process from offline to online aiming still the
circumstances, who should concentrate the knowledge before
43
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_8, © Springer Science+Business Media B.V. 2010
44 KUMLANDER
university courses and students, then the state service is mainly with the main participants’ activities at work or home.
directed to broader auditorium of citizens of different ages. So, Therefore we should foresee different level of education and
here e-learning participants can be those who completed the previous experience of users. Besides the ability of individuals
last study long time ago, people who spent a lot of time on to acquire knowledge is known to be sufficiently varying
other primary activities, individuals who wish to obtain modern dependently on their age and it makes students also different.
knowledge or acquire a completely new profession. Moreover Finally it should be possible to define any course on packets
we should still count with potential participants from suitable to students’ availability as highlighted earlier and be
universities or even schools. Therefore it is important to available online on demand.
personalize the delivered courses as much as possible it order
to increase a probability of successful acquisition of knowledge C. Responsibilities
by each participants. It is incorrect to assume that the organizer of the virtual
As a concluding note for the auditorium description we learning environment is responsible only for the content of
would like to mention that all this defines a challenging task for courses, i.e. quality of the learning process. Unfortunately there
the system to acquire the context of students in order to are a lot of other responsibilities, which should be taken into
personalize delivered information correctly. account building such a service.
First of all the service operates with information about the
B. Targets user and in many cases it is more just a username and a
The high level task of the proposed virtual educational password. Notice that here we don’t mean only the protection
system is to serve citizens of a state looking for possibilities to of the data in the communication channel between the e-
upgrade their knowledge. Besides, the system should be learning server and client computers. The personalization of the
intelligent enough to find a knowledge delivery workflow system means that it should collect a lot of context about the
suiting for a particular individual basing on her previous consumer and all this information is stored on the server in
experience, background and availability. forms of different logs, patterns and so forth. Moreover even
the information on which courses were taken, when and what
All this must be explained in more details in order to marks she got should be classified. Only the person (i.e. owner)
identify correctly the scope and complexity of the goal. First of could expose this information, for example temporary to the
all such a virtual environment should be a proactive way of a employer, and nobody else.
state to deal with a potential unemployment and prevent the
drop of workforce skills. The reactive approach is usually too Secondly the course authors are usually also highly
expensive due a need to pay social benefits and have a low interested in protecting the content of lessons as much as
level of technological development loosing competition to possible. So, there is a clear demand to build virtual
other states. All this expenses are carried by the entire society environments that will be both easy to consume by students
in form of different taxes and the government task is to avoid and hard to misuse or replicate without the author permission.
them if possible. That is why the paper proposes the system to It is quite a challenge task and the easiest way to protect the
be developed as a service provided by the state. It generally content is to set a copyright on this. Unfortunately it is not
means that it should be enough in many cases, so authors will be thankful if the system
will contain other protection mechanisms.
1. Supported by the state (financing);
The protection of such private data is the most important
2. Advertised by the state; challenge for the system developers. They need to ensure that
3. Integrated into the general state infrastructure and data neither can be stolen nor made available outside due any
made available basing on dedicated infrastructure kind of errors or backdoors of the system.
elements implemented by the stated.
If the first two elements are easy to understand then the last III. GENERAL MODEL
one can easily be missed. Unfortunately it is not possible to This chapter continues the description of the virtual
build such a system from scratch. It should be an evolutional learning environment introduced on our last paper [10] and
approach when the educational system follows other state proposes additional artificial intelligence elements supporting
services and relies on the same delivery infrastructure in order the personalization based on the acquired context. Therefore in
to make it available (in both meanings: ability to access and a the beginning an overview of the system is given to describe
constant availability). The educational content (data flow) is the topic defined so far: prerequisites for building such a
much bigger than other information service like registers and system and interfaces of then model.
so demands sometimes completely new infrastructure system to
be built up. Fortunately state usual can afford financing and A. Prerequisites
supporting that considering the social cost it will carry vice
The prerequisites will be described basing on a particular
versa.
example of one European country – Estonia. The country has
Moreover the system should not unify or demand too much an informational system available to citizens to request their
from consumers in terms of spent time and required knowledge information hold by different state registers. The system is
level in order to participate. An ideal life-long education orients developed centrally by a special agency and mostly
on gradual upgrade of work-force skills organized in parallel concentrates information available internally from different
CONTEXT DRIVEN PERSONALIZED E-LEARNING ENVIRONMENT 45
departments like Real-estate agency, Police, Automobiles should stimulate producing personalized courses trying
registry and so forth. Moreover the system does allow citizens to keep simple courses set on minimum level and make
to send requests and documents to the state in order to obtain a this route available only for distinguished lectors. The
new passport for example. All departments of the state should interface should contain also templates making the
process such electronic documents signed by so called “ID- course building process both simpler and standardized.
card” and there should be no difference between paper and
electronic channel of communication. The information system Notice that for the simplification we assume that all
has a single entry point. courses do not require a mentor to help students to
acquire the knowledge, although this element can also
This is an example of the system that normally should be in be included if required.
place before building the state-wide fully functional
educational system. Here we would like to list some reasons Student Author Business company
why we believe such system is required:
1. The process of delivering information to citizens is
verified and in use, so citizens know how to find and access it; Calls brokers
2. Citizens do trust to such system much more than to any
other especially if the state is taking seriously the protection of
private data; HTML UI Engine Security server
3. The security (log in) is well established and is normally
supported by some kind sophisticated mechanism of passwords Communication
distribution (in Estonian case: ID-cards issued by the state, server
which are equivalent to passports, i.e. well protected including Courses
warehouse Templates,
the law); standards and AI engine
4. The developers’ team is acquired and educated; referencies
relationship to external contractors (including the internet
providers) is well established. Context
Results: engine
B. Overview of model interfaces Statistic engine
Such a broad system like described here has several
complex interfaces designed to different type of actors.
Figure 1. General model
1. Common part.
First of all each interface does contain the security 3. Consumers (students) interface
elements that can be inherited from the pre-system like The nature of this interface is obvious: students do
the one described earlier. Secondly, there are a set of participate in courses by passing different topics of
databases on the back end that contain course step by step and (optionally) pass an exam in the end.
materials, results and other personal information on This is the interface we are going to personalized first
participants and so forth. Thirdly there are of all by using AI and context engines shown on the
infrastructure elements that do support the figure 1.
communication like a load-balancing server, which
should spread the load between a set of servers, since 4. Third-party interface
the state wide educational system is likely to be heavily The system should not be restricted to the course
used. Finally there should be a communication server (teacher) – student communication. There are other
that should allow communication between all involved groups of people including business community that
parties outside the learning process in order to provide may be interested in this process retrieving information
a) general feedback on courses, b) define interests from if it is exposed or providing requests. This is a way job
business community to design courses or stimulating seekers could expose their professional level to
participations – for example a job advertisement companies. It is also an interface for companies to
defining passing certain courses as a minimum stimulate creation of a work-force they do require by
requirement and so forth. motivating authors to create certain courses and
2. Courses’ authors interface. students to participate and obtain high marks.
IV. CONTEXT DRIVEN PERSONALIZATION The next important parameters the system could react on is
the available time slots. The minimum requirement would be to
A. Personalisation using context preserve the position of a lecture where the student had to stop
The context-aware system is a very popular topic nowadays last time as the delivered content is now unique and so could be
in different artificial intelligence systems starting from web hard to reproduce exactly. Ideally the delivered portions should
pages up-to automobile user interfaces for drives. be dependent on availability of a person. For example she
could define in advance time size of available time slots (for
Generally, context-aware system development means that example 15 minutes each) and the system will deliver lessons
the computer reaction can be organized using more than just of that size dividing those on topics and subtopics making the
the direct data input. Instead computer system should consider acquired knowledge tractable on the lowest grain (up to a
circumstances under which the input is provided sensing the lesson page).
environment in which the request is formulated and form the
response basing on those parameters [10]. Some authors say The progress of learning could also be measured during the
that it is nothing else that a system ability to adapt to specific education and delivered in different level of explanation
circumstances under which the system operates or reacts on the provided. The explanation could be short to keep the speed of
request and ability to collect and recognize the context moving over topics or could be very details if the person have
information [11, 12]. Therefore the context-aware approach is no ideas at all about the explained concepts (or previous
mainly used to increase relevancy of the response (service) experience with those), so will require slowing down in this
provided by the system and is targeted to bridge the gap particular concept or a slow delivery of the knowledge in total.
between formal and human to human type interaction [13]. The next important stage of the education process that can
Unfortunately majority of scientists do restrict themselves be a subject on context driven personalization is the
with the most obvious context they can acquire – the location. examination. Sometimes results of the education process are a
Fortunately for us it is not the only one and the virtual valuable achievement exposed to the business community.
education system is the perfect example of a system that could Those can be used to sort individuals basing on their skills and
obtain and process other types of context in order to make the therefore the system should prevent the misuse of the education
process of education very personal. process by persons attempting to get higher marks than they
actually are worth of. Therefore the study and examination
The previous experience and educational level can be process should be closely watched in order to identify the
identified as one context that can be used to build up personal person taking the exam. There are hard and soft security system
courses. It can be acquired in a form of tests and questionnaires based on biometrical patterns that can be recorded
where a student defines both her current knowledge and topics communicated to a person. The basic idea here is to learn
she wishes to learn. The standard environment would leave it computer system to identify humans as they do in their
up to users to select courses or lessons she likes to attend, but everyday life distinguishing one person from another. Hard
sometimes courses and lessons interconnection is too complex methods, like face recognition, require certain hardware to be
to understand or even hidden from the student. Therefore an used and therefore are not possible to be applied in many cases.
artificial system based on clearly defined rules will be a perfect Fortunately there are soft methods as well. One of the simplest
mentor existing inside the educational system. If courses are “soft” biometrical identification methods is keystroke patterns,
divided into isolated lessons, then it will be possible to skip where the system monitors the keystroke typing dynamic, used
certain, that the student doesn’t need, or replace them with symbols and words. It is well-known that writing/typing
others bridging the gap in certain areas of student knowledge. dynamics are assumed to be unique to a large degree among
This will ensure that all prerequisites for a successful different people [6]. The same is known for the selection of
understanding of next lessons are provided to the student. words individuals use or keys on the keyboard (consider
Moreover, if the previous case will do some minor changes “delete” or “backslash” keys, which are nearly identical in
in a course, then more sophisticated engine could combine functionality). Moreover the keystroke dynamic is not one
different lessons building a course on demand from available factor to be measured but is a set of characteristics [7, 8, 9]:
parts. Such artificial intelligence could easily produce non 1. Duration of the keystroke, hold time;
envisioned courses basing on personal characteristic either 2. Latency between consecutive keystrokes;
without authors supervision or minimizing it sufficiently. 3. Overall typing speed;
Obviously this is possibly only if the number of available
4. Habits by using additional keys on the keyboard like
lessons form a big enough database and those are well isolated.
typing numbers via number pad part;
The production of the delivered context basing on previous 5. Frequency errors and backspace vs. delete key using;
knowledge is not the only way to process a context available and some others
for the system. Another important property the system could
monitor and measure is how well the student acquires the
knowledge. For example the course could be a combination of The pattern captured during tests and learning process can
explanation and practical tasks. The number of practical tasks be reused during the examination phase and identify the student
could vary depending on how well previous were made in doing the exam. If there is any mismatching then the system
order not to slow down quick brains providing enough tasks to should either have some kind person-to-person interaction
slow catchers. possibility making possible to uniquely identify knowledge of
the person or should have a concept of reliability of exam
CONTEXT DRIVEN PERSONALIZED E-LEARNING ENVIRONMENT 47
marks, which is delivered together with those, so interested The more sophisticated protection mechanism can be build
parties could reevaluate those if needed. basing on the acquired context during the process of education.
The first problem that occurs quite often is to a) detect the node
Recent researches have reported that up to 65% of people the course leaked out and b) prove who the real author of the
over the middle age have certain disabilities and therefore course is. A so called “watermarking” could greatly help us
could like to modify the user interface via which the content is resolving those two issues. It is possible to make the delivered
delivered. This is another very important personal context dynamic even if quite static files are delivered to the
characteristic, which the system should react on. Notice that in student by embedding into those invisible digital marks stating
many cases this is ignored in systems directed to students as the owner of the content and the student it is delivered to. The
quite a minor portion of young people suffer from any kind of process is basically similar to adding watermarks into
disabilities. In the state wide system such ignorance is not banknotes except the fact that in the digital case we would like
possible any longer. The two elementary possibilities are to keep them as hidden as possible in order to prevent
required as a minimum. A possibility to increase the font size removing of those. Notice that those marks are added right
of the delivered content is one of those. Notice that although prior delivering the content to the student inside the virtual
many browsers nowadays offer such functionality, not all web environment web server.
pages react on this correctly. The second one is possibility to
deliver voice explanations. It will decrease the overall pressure Besides, the earlier described personalization of the
on eyes for the education process participants and allow delivered context means that all lessons are divided into very
consuming the content by persons with disabilities. Moreover it small chunks which are delivered on demand basing on
could detach the education process from the computer as there concrete person characteristics. In those small parts certain
are a lot of devices able to reproduce voice recording but not lessons can be omitted or replaced with others from completely
able to show web pages due the restricted screen size (or no other topics. All this could make the process of restoring the
screen at all) – consider as an example cell phones and mp3- original content and reusing it to be very complex. It will
players. require a lot of time and a certain level of knowledge making it
even more complex that creating a course from scratch.
The age or disabilities is not the only reasons for slowing Summarizing, the personalization divides the content into
down the educational process. The overall environment blocks making he course hard to acquire as a whole.
students are located in could also promote or suppress the
ability to understand information including the type of device
the content is delivered too. Obviously even the sophisticated V. CONCLUSION
PDA device need restructuring delivered information as it The virtual educational system described in the paper is an
doesn’t have the same screen as a usual computer. Besides important component ensuring the social and financial stability
writing a text using the same keyboard at home or in the train is of any state. The work-force productivity is one the most
sufficiently different. Therefore the monitored keystroke patter important factors defining the success of the state industry and
used so far and property of the remote device could and should innovation. The level of unemployment and professionalism
be used to adjust materials correspondingly. (and consequently of salaries) defines the health of society
including such as the level of crime or development of all state
B. Protecting the content institutions. Therefore it is important to fund and developed
Nowadays, when knowledge becomes more and more such systems centrally by the state. The broad goal of the
important once again, many teachers are concerned about system makes it hard to build as it should serve educational
protect of their intellectual product (the course) from needs of completely different people. Therefore the
unauthorized copying. This becomes extremely important in personalization of the delivered services becomes extremely
the virtual environment, which is easy to access and even easier important developing such systems. Fortunately the state either
to copy and reproduce materials. contains enough information about their citizens that can be
consumed or the state system is trusted enough to enable
There are several ways to avoid this issue. First of all a context-driven collection of such data. Later the context can be
copyright can be defined, which means a legal protection. converted into the concrete delivery of the education content to
Unfortunately it is not always good enough in virtual the concrete person during the concrete time-interval making
environments since the course can be used in other countries the lesson unique and completely suitable to the available
producing certain barriers for courts and laws. Secondly the moment of life-time of the student. All this is designed to
author rights can be acquired by the state and deliberately increase the knowledge acquisition level from the online
unmonitored. The primary goal is to educate citizens and if it is system and close the gap between online education and
achieved, then reuse of the course by anybody else can be teacher-student offline collaboration, which is usually much
ignored, as long as citizens do understand what courses they more intensive and so productive.
could take free of charge via the state system. The third and the
last elementary protection mechanism is the language barrier,
REFERENCES
i.e. the national language used to build up the course prevents
the same course to be transferred to countries with other [1] D. Mowery, R. Nelson, B. Sampat, and A. Ziedonis, “Ivory Tower and
Industrial Innovation,” Stanford Business Books, California: Stanford,
languages solving a problem with countries where the law 2004.
cannot be enforced to restore the course ownership.
48 KUMLANDER
[2] K. Fischer, “The university as economic savior: as higher education the 25th conference on Proceedings of the 25th IASTED International
replaces industry in some cities, colleges face both support and Multi-Conference: artificial intelligence and applications, 2007, pp.
unrealistic expectations,” Chronicle of Higher Education, 2007. 317-322.
[3] J. Youtie, and P. Shapira, “Building an innovation hub: A case study of [8] A. K. Jain, A. Ross, and S. Prabhakar, "An Introduction to Biometric
the transformation of university roles in regional technological and Recognition," IEEE Trans. on Circuits and Systems for Video
economic development,” Research policy, vol. 37 (8), pp. 1188-1204, Technology, vol. 14 (1), 2004, pp 4-19.
2008. [9] A. K. Jain, L. Hong, and S. Pankanti, “Biometric Identification,”
[4] D. S. Siegel, D. A. Waldman, L. E. Atwater, and A. N. Link, “Toward a Communications of ACM, vol. 43 (2), 2000, pp. 91-98.
model of the effective transfer of scientific knowledge from [10] D. Kumlander, “Soft Biometrical Students Identification Method for e-
academicians to practitioners: qualitative evidence from the Learning,” in Advances in Computer and Information Sciences and
commercialization of university technologies,” J. of Engineering and Engineering, Springer, Netherlands, 2008, p. 114 – 118.
Technology Management, vol. 21(1-2), pp. 115-142, 2004. [11] P. Chang, A. Gul, "Towards Context-Aware Web Applications," 7th
[5] B. Readings, Education / Teaching, Harvard University Press, 1997. IFIP International Conference on Distributed Applications and
[6] E. González-Agulla, E. Argones-Rúa, C. García-Mateo, and Ó W. M. Interoperable Systems (DAIS), Springer, LNCS 4531, 2007, pp. 239–
Flórez, “Development and Implementation of a Biometric Verification 252.
System for E-learning Platforms,” EDUTECH, Computer-Aided Design [12] W.Y. Lum, F.C.M. Lau, “A context-aware decision engine for content
Meets Computer-Aided Learning, IFIP 18th World Computer Congress, adaptation,” IEEE Pervasive Computing, vol. 1, no.3, 2002, pp. 41–49.
2004, pp. 155-164. [13] J. Pascoe, “Adding Generic Contextual Capabilities to Wearable
Computers,” Proc. 2nd Int’l Symp. Wearable Computers, IEEE CS
[7] O. Guven, S. Akyokus, M. Uysal, and A. Guven, “Enhanced password
Press, Los Alamitos, Calif., 1998, pp. 92–99.
authentication through keystroke typing characteristics,” Proceedings of
Exploring Teachers’ Beliefs about Digital Citizenship
and Responsibility
Harrison Hao Yang
State University of New York at Oswego
Oswego, NY 13126, USA
[email protected]
Pinde Chen
South China Normal University
Guangzhou, 510631, China
[email protected]
Abstract – Promoting and modeling digital citizenship and This paper reviews related studies connected with the digital
responsibility can be a challenge to today’s teachers. This paper citizenship movement and teachers’ beliefs gleaned from the
presents a review of related studies connected with the digital literature. It then provides a study which explores teachers’
citizenship movement and teachers’ beliefs gleaned from the beliefs about digital citizenship and responsibility.
literature. It then reports a descriptive study which explores
teachers’ beliefs about and opinions on digital citizenship and II. LITERATURE REVIEW
responsibility.
A. Digital Citizenship
The movement to address and characterize digital citizenship
I. INTRODUCTION started in the UK, where educators have been working toward
Today’s world has moved from the industrial age into the establishing protocols for good digital citizenship since the
information age. The rapid technological change and mid-1990s [5]. According to Digizen.org, owned and operated
proliferation of information resources are lineaments of our by London-based nonprofit Childnet International,
contemporary society. Multimedia on the Internet, Digital citizenship isn't just about recognizing and
telecommunications, wireless applications, electronic portable dealing with online hazards. It's about building safe
devices, social network software, Web 2.0, etc., are all spaces and communities, understanding how to manage
radically redefining the way people obtain information and the personal information, and about being internet savvy -
way to teach and learn [1]. Findings of a 2007 national-wide using your online presence to grow and shape your
survey study Teens and Social Media conducted by the Pew world in a safe, creative way, and inspiring others to do
Internet and American Life Project show that some 93% of the same [6].
teens use the internet, and more of them than ever are treating In the US, Ribble, Bailey, and Ross defined digital
it as a venue for social interaction – a place where they can citizenship as “the norms of behavior with regard to technology
share creations, tell stories, and interact with others. In use” [4]. They identified nine elements of digital citizenship:
addition, content creation by teenagers continues to grow, with digital access, digital commerce, digital communication, digital
64% of online teenagers ages 12 to 17 engaging in at least one literacy, digital etiquette, digital law, digital rights and
type of content creation, up from 57% of online teens in 2004 responsibilities, digital health and wellness, and digital
[2]. security. Exploring these elements has led an array of studies
While the emerging technologies provide more opportunities focusing on the implementation of good digital citizenship in
and possibilities to enhance existing teaching and learning schools, such as assessments, plans, strategies, activities, etc.
activities, there is considerable evidence that technology [3] [4] [5] [7] [8] [9] [10] [11] [12]. As Ribble, Bailey, and
misuse and abuse is widespread and can be found inside and Ross indicated “digital citizenship speaks to several levels of
outside the school today: Web sites to intimidate or threaten responsibility for technology. Some issues may be of more
students, downloading music illegally from the Internet, concern to technology leaders while others may be of more
plagiarizing information using the Internet, using cellular concern to teachers” [4]. The recent National Educational
phones during class time, and playing games on laptops or Technology Standards for Teachers 2008 (NETS·T 2008)
handhelds during class, etc [3] [4]. As a result, digital prepared by the International Society for Technology in
citizenship has been received the increased attention of many Education (ISTE) has attested the digital citizenship movement
educators and has been considered as one of top priorities in and specialized its Standard IV for teachers to “Promote and
numerous school districts. “Everyone – administrators, board Model Digital Citizenship and Responsibility” [13]. Standard
members, teachers, parents and students – need to be involved IV requests teachers “understand local and global societal
in the dialogue about the appropriate use of technology” [3]. issues and responsibilities in an evolving digital culture and
49
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_9, © Springer Science+Business Media B.V. 2010
50 YANG AND CHEN
exhibit legal and ethical behavior in their professional to investigate teachers’ beliefs about digital citizenship and
practices.” In order to guide teachers meeting this standard, responsibility in particular.
NETS·T 2008 Standard IV has four indicators covering
III. THE STUDY
different aspects of digital citizenship and responsibility:
• Indicator A: advocate, model, and teach safe, legal, The purpose of this study was to examine teachers’ beliefs
and ethical use of digital information and technology, about and opinions on NETS·T 2008 Standard IV - Promote
including respect for copyright, intellectual property, and Model Digital Citizenship and Responsibility. The
and the appropriate documentation of sources. following questions guided this study:
• Indicator B: address the diverse needs of all learners 1. What were teachers’ beliefs about and opinions on
by using learner-centered strategies providing each indicator of NETS·T 2008 Standard IV?
equitable access to appropriate digital tools and 2. How could teachers adopt and implement the four
resources. indicators of NETS·T 2008 Standard IV into their
• Indicator C: promote and model digital etiquette and teaching practices?
responsible social interactions related to the use of A. Participants and the Course
technology and information. The participants of this study came from four sections of
• Indicator D: develop and model cultural students (n = 87) who were enrolled in the asynchronous online
understanding and global awareness by engaging with course entitled Multimedia and Internet for Educators, offered
colleagues and students of other cultures using digital- at a university in the northeastern region of the United States
age communication and collaboration tools [13]. during the winter, spring, summer, and fall terms in 2009. Most
B. Teachers’ Beliefs of the participants were part-time in-service teachers located in
In describing a working conception of beliefs, Bryan the same geographical area, who were pursuing graduate level
summarized “beliefs are psychological constructions that (a) education programs in content areas of literacy, biology,
include understandings, assumptions, images, or propositions chemistry, earth science, mathematics, social studies,
that are felt to be true; (b) drive a person’s actions and support technology, etc.
decisions and judgments; …” [14]. It is important to realize The course Multimedia and Internet for Educators provides
that pre- and in-service teachers’ opinions and briefs shape an introduction and guide to pre- and in-service teachers to
their actions. As Clark and Peterson pointed out: current and emerging technology. It intends to help PK-12
…the research shows that thinking plays an important educators not only to use multimedia and Internet resources in
part in teaching, and that the image of a teacher as a their own educational endeavors, but also integrate technology
reflective professional…is not far-fetched. Teachers do into their work environments as teachers. Since spring 2001,
plan in a rich variety of ways, and these plans have real this course has been joined in the State University of New
consequences in the classroom. Teachers make York Learning Network (SLN) as one of its asynchronous
decisions frequently (on every two minutes) during learning network courses offered to both on and off-campus
interactive teaching. Teachers do have theories and students and is an elective graduate course. In each module of
belief systems that influence their perceptions, plan, and this course, students were required not only to complete and
actions [15]. submit the assignment, but also to collaborate on specific
Similarly, Cuban stated, “The knowledge, beliefs, and learning activities required during a certain time frame: (a)
attitudes that teachers have shape what they choose to do in reading each “Mini-Lecture” and supplemental materials, (b)
their classrooms and explain the core of instructional practices responding to the questions in each Mini-Lecture at the
that have endured over time” [16]. Pajares found there was a “Online Discussion” section, (c) responding publicly to some
“strong relationship between teachers’ educational beliefs and or all of the questions submitted by classmates, and (d)
their planning, instructional decisions, and classroom replying to classmates who responded to the question.
practices” and that “research on the entering beliefs of pre- Asynchronous messages posted by the participants were
service teachers would provide teacher educators with threaded and could be sorted by author or by date for the
important information to help determine curricula and program participants’ viewing and reviewing preference.
direction and interpretation of knowledge and subsequent B. Data Collection
teaching behavior” [17]. To investigate teachers’ beliefs about and then to promote
In sum, the literature suggests that it is critical to investigate their understanding of digital citizenship and responsibility, in
the pre- and in-service teachers’ beliefs about technology the discussion area of Module 1 - Introduction of Multimedia
innovation and integration. Specifically, the pre- and in-service and Internet the participants were asked to provide the written
teachers’ beliefs and opinions indicate their willingness to responses to the following statement:
adopt a new teaching behavior such as promoting and People are excited by the power and popularity of the
modeling digital citizenship and responsibility in their Internet, in your opinion, what are the
professional practices. Although previous studies have ethical/social/human issues we should consider when we
provided a variety of practical ways on the implementation of implement digital technologies, especially the Internet-
good digital citizenship in schools, such as assessments, plans, based technologies into today's education?
strategies, activities, etc. in general, few studies have attempted
EXPLORING TEACHERS’ BELIEFS ABOUT DIGITAL CITIZENSHIP AND RESPONSIBILITY 51
As indicated in Table 1, a total of 299 responses from • [Student A-Fall responded to Student B-Fall’s
participants were collected and analyzed by the instructor of response]
the course, who is the first author of this study as well. I think that if a teacher chooses to use the internet
during any type of project it is essential that the
TABLE 1 students learn it is not appropriate to copy someone
NUMBERS OF PARTICIPANTS AND RESPONSES
Participants Responses
else’s work. On top of that I think it is the teacher’s
Winter 09 26 92 job to double check each student’s work and make
Spring 09 20 69 sure they haven’t done so. I remember from student
Summer 09 21 91 teaching placement when I gave an essay assignment
Fall 09 20 47 and had a student plagiarize I took them aside and
Total 87 299
asked why they did so. They told me they had never
gotten caught before so they thought they could get
C. Findings
away with it. They understood they did something
Among 299 responses, while most of them addressed issues
wrong but had never gotten caught before. The
ranged over two or more indicators of NETS·T 2008 Standard
reinforcement of being caught would help prevent it
IV, some of them focused on one indicator of NETS·T 2008
in the future.
Standard IV. Table 2 listed numbers and percentages of
responses that pertained to each indicator. • [Student C-Fall responded to Student B-Fall’s
response]
TABLE 2 Whenever students are assigned to write an essay in
DISTRIBUTION OF PARTICIPANTS’ RESPONSES which they need to do research I feel the teacher
Frequency Percent should always give a mini lecture on the importance
Indicator A 210 70.2
Indicator B 44 14.7 of citing work to avoid plagiarism as well as making
Indicator C 125 41.8 students aware of the consequences they will face if
Indicator D 107 35.8 they are caught in the act of plagiarism. Students
should also be reminded of how to determine if a
First of all, the vast majority of responses (70.2%) were source is reliable and creditable as well as review on
falling into the Indicator A – “advocate, model, and teach safe, how to site works (or a handout) if necessary. My
legal, and ethical use of digital information and technology, content area is business but I always discussed
including respect for copyright, intellectual property, and the citations, references, and plagiarism with my students
appropriate documentation of sources.” The following online as it is important in every subject.
discussions in fall 2009, which involved three participants Secondly, although there were fewer responses (14.7%) on
exchanging their ideas on the plagiarism, were typical the Indicator B - “address the diverse needs of all learners by
examples: using learner-centered strategies providing equitable access to
• [The initial discussion on the plagiarism posted by appropriate digital tools and resources,” participants
Student A-Fall] thoroughly discussed this important area regarding the
Researching for papers the internet makes the implementation for their teaching practices. As were evident in
availability to plagiarism a lot easier for students. the following participants’ discussion in spring 2009:
Students even have the option to purchase papers of • [The initial discussion on the digital access posted by
the internet. The authenticity of individual work may Student A-Spring]
be more difficult to determine with the access of the The major social issue that I have been faced with is
internet. the fact that many of our students do not have access
• [Student B-Fall responded to Student A-Fall’s initial at home and it is important to be fair when giving
post] assignments which require this. In fact, I have gone
I agree that it is much more convenient and easy for so far as to meet students at the local library to help
students to plagiarize while using the World Wide them with a project that I assigned, so that they would
Web (sometimes without even knowing it!). Due to have the time outside of class, but my support, since
their early start to using the internet and their lack of they may not be as familiar with the technology as
knowledge about plagiarism this could definitely be a their peers.
problem that might occur when using the internet in • [Student B-Spring responded to Student A-Spring’s
the classroom. I think as educators we need to address initial post]
this problem, educate our students about plagiarism I applaud you for going to the local libraries to help
and teach them how to use the information correctly. your students. Why don’t they use the computers in
If a teacher chooses to use the internet for a research school? There are a number of students who don’t
project I think it is essential to address this issue in the have a computer at home do you have alternative
very beginning and spend time discussing how to version of the assignment for them?
complete a works cited to use the information legally. I do not think of lack of computers at home as a social
issues. Good thinking. I hope this is the reason why
52 YANG AND CHEN
some schools are developing laptop programs for their their new form of bathroom graffiti and harassment.
students. Teachers can't witness this cyber-bullying and parents
• [Student A-Spring responded to Student B-Spring’ don't even know it's going on most of the time. I
response] think this gives the bullies an even bigger sense of
Usually it's just one or two students who don't have power. They are unseen, and sometimes even
the access and it's for those kids who I will go above nameless, except for their screen name, which maybe
and beyond - although I could provide an alternate is known or unknown. When students are given
assessment, it is important to me that they have the computer usage time, it needs to be carefully
authentic experience like their peers. monitored by the network administrator and the
• [Student C-Spring responded to Student B-Spring’s supervising teacher. A students' history can be
response] tracked, but it is difficult for a teacher to monitor 20
My school had the laptop program (it was a students' activity when all it takes is a click of the
predictable disaster, and no longer is offered), and mouse to close a window you're not supposed to be in.
only the students who were leasing one of the laptops What are some effective ways to monitor computer
($$) got to take them home. The students who could usage by a classroom of students? How can educators
not afford to lease them only had access to them try to eliminate cyber-bulling in their schools?
during the day. They had to return them before going • [Student B-Winter responded to Student A-Winter’s
home. So, the disadvantaged students who probably initial post]
did not have a computer at home to begin with still As I think about your question about how educators
couldn't have access at home because they couldn't can try to eliminate cyber-bulling in our schools it just
afford to lease the laptops and take them home. From brings up such a difficult task. It makes me think of
the financial standpoint, it makes sense. The district just regular bullying and how much of it occurs
can't afford to have kids "lose" $800 laptops for which outside of school. How do we combat that? I've
there will be no financial recovery. talked with parents before whose children are on the
• [Student D-Spring responded to Student A-Spring’ social networks and they said from their standpoint,
initial post] it's just lots of monitoring. If anything, it seems like
As a teacher I am also facing with the same social cyber-bullying is easier to monitor because they leave
issue in my own classroom. It is great that students a trail of everything they are doing.
can access their homework from the internet or • [Student C-Winter responded to Student A-Winter’
interact with each other in a math game, but only for initial post]
the students that have a computer at home. Many of Great post and great questions! I found this article
my students do have a computer, but a few don't. The Tools Help Families Combat Cyber Bullying with
ones that do have a computer don't always have Alerts, Tips about a tool called "CyberBully Alert"
internet access at home. I choose not to really assign and thought the information was interesting to share -
any homework that requires using the internet for this “One such tool, recently unveiled by Vanden, allows
reason. We use it within our classroom. children to instantly notify selected adults when they
• [Student E-Spring responded to Student D-Spring’ are bullied or harassed online. CyberBully Alert
response] (http://www.cyberbullyalert.com/) also documents the
I agree, and only require use of computers when I am threatening message by saving a screen shot of the
supplying them in my classroom via laptop carts. It child's computer when the child triggers an alert. The
surprises me that so many people fail to consider Web-based solution is compatible with most
students lacking access to a computer when creating computers and Web browsers. It requires a single
assignments. download and registration. Parents then select which
Thirdly, approximately 42% of the responses addressed e-mail addresses and phone numbers will receive the
various aspects on the Indicator C – “promote and model alerts. Alerts are sent through e-mail and text message
digital etiquette and responsible social interactions related to when a child clicks on the CyberBully Alert icon…
the use of technology and information.” The following online The creators of CyberBully Alert also offer a School
discussions in winter 2009, which involved three participants Incentive Program” (cited from Information Week,
exchanging their ideas on the cyber-bulling, were typical October 1, 2008).
examples regarding one of new-fangled and concrete issues Finally, there were approximately 36% of the responses
related to Indicator C: falling into the Indicator D – “develop and model cultural
• [The initial discussion on the cyber-bully posted by understanding and global awareness by engaging with
Student A-Winter] colleagues and students of other cultures using digital-age
Bullying has always been an issue in schools among communication and collaboration tools.” Participants reported
students. Now they are able to do it behind computer their deep awareness of this indicator. As were evident in the
monitors and cell phones. Some of this is being done following participants’ discussion in summer 2009:
right from computers and cell phones at school. It's • [The initial discussion posted by Student A-Summer]
EXPLORING TEACHERS’ BELIEFS ABOUT DIGITAL CITIZENSHIP AND RESPONSIBILITY 53
My concern is that when students can be a part of the • [Student C-Summer responded to Student A-
“social clique” created with technology, how will that Summer’s response]
affect their ability to be social in the real world? This is an interesting thought because there are many
When someone is either at the keyboard or with communication resources that students can use. There
headsets on, will their real world collapse into email, is a fine line between replacing technology-based
YouTube and iPod? Will families still have the same communication with face to face communication.
connections, when each as an individual has the Personally, I believe the technology-based
ability to create their Second Life avatar, design their communication that we have today is helping our
own room and hang out with people who are only like society stay in touch. Although it is not face to face
them? Are we building communities for our children communication, it does not mean that people are
that are physical and in real time? I realize these are replacing technology with going out and meeting
questions, but when our students are being prepared people. This technological communication surge is
for a high technology workplace, are their social and simply helping stay in touch with people on a 24-7
moral obligations to each other as community citizens basis.
being addressed? I also want to note that as our youth
IV. CONCLUSION
are exposed the ethical and social differences of other
countries, other religions and the range of human The findings of this study may produce useful information
behavior throughout the world, will our culture be that will enable administrators, program planners, and teachers
affected? That is where educators have the moral to better understand teachers’ beliefs about digital citizenship
obligation of understanding how we are exposed to and related responsibilities. As Ribble and Bailey pointed,
differences and how they are discussed can have “The endless list of misuse and abuse includes hacking into
impact our personal lives and future choices. school servers, using e-mail to intimidate or threaten students,
• [Student B-Summer responded to Student A- illegally downloading music, plagiarizing information from the
Summer’s initial post] Internet, using cellular phones during class time, accessing
The issues you have raised should be a concern for pornographic Web sites, and playing video games during class.
everyone. More and more people are not Therefore, if you are using technology in your district, you
communicating face to face. I think the implications must begin to deal with digital citizenship in a significant way”
that this can have on society can be detrimental to [10]. To capture the improvements in digital citizenship and
how we all communicate with each other. I sometimes responsibility, teachers must be prepared – through sharing and
find myself texting to a friend or family member when expressing their beliefs and opinions – to learn what digital
I could easily call them and verbally say what I citizenship is and what it can do. Indeed, there have been some
needed to. Progressively, since the continuing line of of valuable resources and activities related to digital citizenship
new advancements in technology, the idea of existing for teachers and educators working on curriculum. For
community has declined in society. It’s scary to think example, for teachers to teach Internet safety, some of lessons
that more and more people have replaced the ideas of and activities including electronic simulations are available
community with thoughts concerning only online [18]:
themselves. As educators I think it is important that • Cybersmart (http://www.cybersmart.gov.au/) is
we continually stress the importance of community developed by the Australian Communications and
and help students to understand the value of working Media Authority, and “provides activities, resources
together. By exposing students to the benefits of and practical advice to help young kids, kids, teens
community it will help them understand that in order and parents safely enjoy the online world” [19].
to thrive, community cooperation is essential. Also, I • i-SAFE (http://www.isafe.org/) is endorsed by the
think that our culture will undoubtedly be affected. U.S. Congress, and “incorporates classroom
Cultural diffusion does not only take place through curriculum with dynamic community outreach to
means on face to face interaction. With the internet empower students, teachers, parents, law enforcement,
students can view videos of virtually any society that and concerned adults to make the Internet a safer
they wish. Students can find information on remote place” [20].
villages in the Amazon rainforest, something few • NetSmartz Workshop (http://www.netsmartz.org) “is
students of past generations could do. I do believe that an interactive, educational safety resource from the
this exposure to other cultures will be beneficial to National Center for Missing & Exploited Children®
students. Educators can utilize this resource to help (NCMEC) and Boys & Girls Clubs of America
teach students to understand other cultures and (BGCA) for children aged 5 to 17, parents, guardians,
societies, working to break stereotypes and prejudices. educators, and law enforcement that uses age-
I hope that affects on our culture will be beneficial; appropriate, 3-D activities to teach children how to
however educators need to understand their stay safer on the Internet” [21].
importance to this cultural diffusion. • ID the Creep (http://www.idthecreep.com) is an
online game sponsored by the National Center for
54 YANG AND CHEN
Missing & Exploited Children and The Ad Council [8] M. S. Ribble and G. D. Bailey, “Digital citizenship: Focus questions for
implementation,” Learning & Leading with Technology, vol. 32 issue 2,
for young students to practice identifying pedophiles pp. 13-15, 2004.
online. [9] M. S. Ribble and G. D. Bailey, “Developing ethical direction,” Learning
If the strong beliefs can be built, then better teaching & Leading with Technology, vol. 32 issue 7, pp. 36-37, 2005.
behavior and learning activities can be developed and [10] M. S. Ribble and G. D. Bailey, “Monitoring technology misuse & abuse:
A five-step plan for creating a digital citizenship program in your
implemented. For this reason, beliefs about digital citizenship school,” T. H. E. Journal, August 2004.
and responsibility among teachers may be an important < http://www.digitalcitizenship.net/uploads/T.H.E.JournalArticle.pdf>
consideration in the process of incorporating NETS·T into real- [11] M. S. Ribble and G. D. Bailey, “Is digital citizenship a problem in your
world teaching, with options for more concrete actions for school?” Technology Turning Points, December 2005.
<http://www.digitalcitizenship.net/uploads/TechnologyTurningPointsTea
teachers. cherTalk.pdf>
[12] M. S. Ribble and G. D. Bailey, Digital citizenship in schools, ISTE,
REFERENCES 2007.
[1] S. Wang and H. Yang, “Editorial,” International Journal of Technology [13] International Society for Technology in Education, National education
in Teaching and Learning, vol. 3 issue 2, 2007. technology standards for teachers, 2008.
[2] A. Lenhart, M. Madden, A. R. Macgill, and A. Smith, “Teens and social <http://www.iste.org/Content/NavigationMenu/NETS/ForTeachers/2008
media: The use of social media gains a greater foothold in teen life as Standards/NETS_for_Teachers_2008.htm>
they embrace the conversational nature of interactive online media,” Pew [14] L. A. Bryan, “Nestedness of beliefs: Examining a prospective elementary
Internet & American Life Project, 2007. teacher’s belief system about science teaching and learning,” Journal of
<http://www.pewinternet.org/~/media//Files/Reports/2007/PIP_Teens_So Research in Science Teaching, vol. 40 issue 9, pp. 835-868, 2003.
cial_Media_Final.pdf.pdf > [15] C. M. Clark and P. L. Peterson, “Teacher’s thought process,” in M. C.
Wittrock (ed.). Handbook of research on teaching, New York: McMillan
[3] M. S. Ribble, “Point of view on technology drivers’ licenses,” District
Publication, 1986.
Administrator Magazine, 2004.
[16] L. Cuban, “Computers meet classroom: Classroom wins,” Teachers
< http://www.digitalcitizenship.net/uploads/TechDL.pdf>
College Record, vol. 95 issue 2, pp. 185–210, 1993.
[4] M. S. Ribble, G. D. Bailey, and T. W. Ross, “Digital citizenship:
[17] F. M. Pajares, F. M., “Teachers’ beliefs and educational research:
Addressing appropriate technology behavior,” Learning & Leading with
Cleaning up a messy construct,” Review of Educational Research, vol. 62
Technology, vol. 32 issue 1, pp. 7-11, 2004.
issue 3, pp. 307-332, 1992.
[5] M. Villano, “Text unto others…as you would have them text unto you,”
[18] A. McAnear, “Digital citizenship,” Learning & Leading with
T. H. E. Journal, September 2008.
Technology, 2007.
<http://thejournal.com/articles/2008/09/01/text-unto-others-as-you-
<http://www.iste.org/Content/NavigationMenu/Publications/LL/LLIssues
would-have-them-text-unto-you.aspx>
/Volume_34_2006_2007_/March_6_/34604m.pdf>
[6] What makes you such a good digizen? < http://www.digizen.org/>
[19] < http://www.cybersmart.gov.au/>
[7] J. Sturgeon, “Raising good digital citizens: Teachers are developing new
[20] < http://www.isafe.org/>
ways for students to follow good internet and media use practices,” Cable
[21] < http://www.netsmartz.org/overview/aboutus.htm>
in Classroom, November 2008.
Faculty Driven Assessment of Critical Thinking:
National Dissemination of the CAT Instrument
Barry Stein, Ada Haynes, Michael Redding, Kevin Harris, Mark Tylka, and Elizabeth Lisic
Tennessee Technological University
Abstract- This paper reports the latest findings of a project to Tennessee Technological University (TTU) has been en-
nationally disseminate the CAT© instrument, a unique interdis- gaged in an extended effort during the last 9 years to develop
ciplinary assessment tool for evaluating students’ critical think- and refine an instrument to assess critical thinking that over-
ing skills. Tennessee Technological University originally part- comes many of the weaknesses of other existing tools. Pree-
nered with six other institutions across the U.S. to evaluate and
minent theoreticians and educators in the area of learning
refine the CAT instrument beginning in 2004. Through these
efforts a test with high face validity, high construct validity, high sciences and assessment participated in the project. Unlike
reliability, and that is culturally fair was developed In the cur- many other available assessment tools, the Critical thinking
rent national dissemination project, we are collaborating with Assessment Test (CAT) instrument uses short answer essay
over 40 institutions across the country to disseminate this unique responses to assess critical thinking. Although essay ques-
instrument and support the creation of assessment based efforts tions can be harder to score than multiple choice questions,
to improve student learning. These dissemination efforts involved they provide a better understanding of students’ thought
training representatives from other institutions to lead scoring processes and ability to think critically and creatively when
workshops on their own campuses as part of a broader effort to confronted with real world problems [6]. In addition, the CAT
improve student learning. A variety of findings indicate these
instrument is unique in that it utilizes a campus’s own faculty
dissemination efforts are successful. National user norms that are
being collected from this project indicate that largest gains in to evaluate student responses. Ewell [7] has noted that one
undergraduate critical thinking appear to occur in the junior and problem with many existing assessments is that “we let as-
senior years. 1 sessment get excessively distanced from the day-to-day busi-
ness of teaching and learning.” The CAT instrument allows
INTRODUCTION faculty to directly observe students’ critical thinking and un-
derstand their students’ weaknesses. This activity helps moti-
In our increasingly technological and information driven vate faculty to consider changes in pedagogy that might im-
society, the ability to think critically is a cornerstone to both prove students’ critical thinking skills [8]. This becomes in-
workplace development and effective educational programs. creasingly important as accrediting agencies such as ABET
The importance of critical thinking skills in higher education increase their focus on efforts to improve students’ critical
has become increasingly apparent. Accrediting agencies such thinking [1].
as Accreditation Board of Engineering and Technology TTU’s approach to developing the CAT instrument has
(ABET) recognize the need for higher order thinking skills been to involve faculty in the identification of a core set of
and real world problem solving in their accreditation standards skills they believe to be an important part of critical thinking
[1]. Faculty also recognize the importance of critical thinking across many disciplines (see Table 1). TTU worked with a
skills. According to Derek Bok [2], over ninety percent of the diverse group of institutions nationwide (University of Texas,
faculty in the U.S. feel that critical thinking is the most impor- University of Colorado, University of Washington, University of
tant goal of an undergraduate education. Despite the recog- TABLE 1
nized importance of critical thinking skills, it is difficult to SKILL AREAS ASSESSED BY THE CAT INSTRUMENT
find effective tools for assessing these skills. Many of the
assessment tools are plagued by problems related to validity, Evaluating Information
reliability, and cultural fairness [3]. According to Bransford, Separate factual information from inferences.
Interpret numerical relationships in graphs.
Brown, and Cocking [4] “a challenge for the learning sciences Understand the limitations of correlational data.
is to provide a theoretical framework that links assessment Evaluate evidence and identify inappropriate conclusions.
practices to learning theory.” Pellegrino, Chudowsky and Creative Thinking
Glaser also note that assessment instruments need to be devel- Identify alternative interpretations for data or observations.
oped based upon principles of learning and cognition [5]. Identify new information that might support or contradict a hypothsis.
Explain how new information can change a problem.
Learning and Problem Solving
1
Separate relevant from irrelevant information.
Partial support for this work was provided by the National Science Founda- Integrate information to solve problems.
tion’s CCLI Program under grants 0404911 and 0717654. Any opinions, Learn and apply new information.
findings, and conclusions or recommendations expressed in this material are
Use mathematical skills to solve real-world problems.
those of the authors and do not necessarily reflect the views of the National
Communication
Science Foundation.
Communicate ideas effectively.
55
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_10, © Springer Science+Business Media B.V. 2010
56 STEIN ET AL.
Hawaii, University of Southern Maine, and Howard Universi- nister the CAT instrument to their students and to conduct
ty) to further develop, test, and refine this instrument. The scoring workshops with their own faculty to score their stu-
results of that project revealed the CAT instrument had high dents’ responses on the CAT instrument. Since institutional
face validity when evaluated by a broad spectrum of faculty interest in the CAT instrument greatly exceeded that level of
across the U.S. in STEM and non-STEM disciplines, had good participation supported by NSF, we created an alternative me-
criterion validity when compared to other instruments that thod of institutional participation that was self-supporting.
measure critical thinking and intellectual performance, had We are currently collaborating with over 40 institutions across
good scoring reliability, good test/re-test reliability, and good the country. Grant funded participants were selected to insure
construct validity using expert evaluation in the area of learn- diversity in mission, institutional characteristics, ethnici-
ing sciences [8]. ty/racial demographics, and geographic location. The collabo-
Student response to the CAT instrument has also been very rating institutions range from community colleges to large
positive. Many students indicate appreciation for the real research institutions and include private, as well as, publicly
world problem solving tasks that make up the test [9]. Using funded institutions.
questions that are interesting and engaging helps motivate
students to perform their best. Materials
Performance on the CAT instrument is significantly corre- A variety of training materials were developed to support
lated with other measures of student performance including the training of new scoring workshop leaders and to help those
cumulative college grade-point average (r = 0.295), entering leaders train their own faculty to effectively score the CAT
SAT or ACT scores (r = 0.528, and r = 0.560, respectively), instrument on their respective campuses. These materials in-
and performance on several other tests that are thought to cluded a narrated overview of the development process for the
measure certain skills associated with critical thinking such as CAT instrument that streams from our website
the California Critical Thinking Skills Test (r = 0.645) and the (www.CriticalThinkingTest.org), a detailed training manual
CAAP Critical Thinking Module (r = 0.691). These correla- that covers issues from sampling and obtaining institutional
tions provide support for the criterion validity of the CAT review board approval to administering the test, recruiting
instrument. However, the magnitude of the correlations also faculty scorers, conducting the scoring workshop, and inter-
indicate that the CAT instrument measures something differ- preting institutional results. A detailed scoring guide and
ent than the other assessment tools [9]. Performance on the multi-media computer based training modules were also de-
CAT instrument has also been found to significantly correlate veloped to explain question scoring.
with several key questions on the National Survey of Student
Engagement (NSSE). One NSSE question that is significantly Train-the-Trainer Workshops
negatively correlated with students’ performance on the CAT Participating institutions sent two to three representatives to
instrument is the extent to which students think their courses a two-day training workshop where they received extensive
emphasized rote retention [10]. In contrast, CAT performance training in scoring the CAT instrument and preparation to lead
is significantly positively correlated with NSSE questions that a scoring workshop on their own campus. The workshop also
reflect the extent to which students’ think their college expe- covered other topics related to understanding and using differ-
rience emphasized critical thinking and real-world problem ent assessment designs, sampling, and using the scoring ses-
solving. sion as an opportunity for faculty development. One of the
Extensive work has been done to refine the scoring process advantages of the CAT instrument is that it gives faculty a
for the CAT instrument to improve scoring accuracy and re- good understanding of student weaknesses. Coupling this
liability. For example, the scoring process uses multiple scor- experience with a discussion of ways to improve student
ers for each question and uses a detailed scoring guide that has learning creates an effective faculty development experience
been refined with extensive faculty input to ensure scoring that can lead to real improvements in instruction and student
consistency. Although scoring accuracy is typically a problem learning.
for essay type tests, the scoring process for the CAT instru-
ment overcomes many of the problems typically associated RESULTS
with scoring essay tests.
The major focus of the new work reported in this paper in- The effectiveness of the train-the-trainer workshops were
volves efforts to disseminate the instrument to a broad range evaluated in a number of ways including; participant surveys,
of institutions across the U.S. These efforts involved training direct observations by our external evaluator, follow-up onsite
representatives from other institutions to lead scoring work- observations of campus scoring sessions, and scoring accuracy
shops on their own campuses as part of a broader effort to checks of tests scored by each institution.
improve student learning.
Training Workshop Surveys
METHOD At the conclusion of each of the training workshops partici-
pants were asked to complete an anonymous survey to eva-
Funding from the National Science Foundation supported luate various components of the workshop training. The sur-
collaboration with 20 institutions across the country to admi- vey used a 6 point Likert type scale: 6 – Strongly Agree, 5 –
FACULTY DRIVEN ASSESSMENT OF CRITICAL THINKING 57
Fig. 1 Fig. 2
Average Training Workshop Evaluation Responses (with 95% CI) Results of Scoring Accuracy Checks Across Institutions as Percent Error
Agree, 4 – Slightly Agree, 3 – Slightly Disagree, 2 – Disagree, about the success of this national dissemination model comes
1 – Strongly Disagree. Eighty-five participants were included from checking the accuracy of scoring performed by each par-
in this analysis. ticipating institution. Experienced scorers at our institution
The mean response on each survey item is shown in Figure rescore randomly selected tests from each scoring session at
1 (error bars indicate 95% confidence interval). The results other institutions. The sample includes about 15% to 20% of
indicate that the workshop evaluations have been quite posi- the tests scored by each institution. Error rates below 5% are
tive across the topics surveyed. considered acceptable. Thus far, overall accuracy has been
very good. Figure 2 shows the overall test scoring error found
Qualitative Evaluation of Training Workshops across 21 institutional scoring sessions that have undergone
Observations by our external evaluators and project staff accuracy checks. About 1 in 7 institutions has been found to
are used formatively to continually improve the training work- deviate more than 5% from our experienced scorers. All insti-
shops. For example, we noted early in the project that partici- tutions receive a scoring accuracy report together with a ques-
pants in the training workshops may find it difficult to process tion by question analysis of accuracy, and suggestions for im-
information from too many different sources (the workshop proving scoring accuracy in subsequent sessions. Recommen-
leader, the multimedia training modules, and the scoring dations are given to institutions that deviate more than 5% on
guide). This led to changes in how the multi-media training how to appropriately adjust their scores for comparison to
module was used in the workshop. It is now used during the other institutions.
second day as a resource for explaining scoring in situations
where scorers may need further explanation. Many other National Norms
changes were also made in response to qualitative evaluations Another goal of the current project is to begin assembling
(e.g., extending the discussion of sampling methods and ex- national norms for various populations of students including
ploring methods for obtaining representative samples at dif- community college students and students at four-year institu-
ferent institutions, and encouraging participants to develop tions. Although many institutions will track their progress by
discipline specific analog learning activities that correspond to
questions on the CAT instrument). 20.0
19.0
Onsite Evaluation of Scoring Workshops
Observations by our external evaluators and project staff 18.0
who have attended scoring workshops at other institutions
Mean
comparison to their own test scores over a period of time, oth- and can begin to explore modifications in teaching methods
ers have requested information that would allow comparison that might address these weaknesses. Many of the participants
to larger populations. Figure 3 shows the mean score on the in our regional training workshops feel that the CAT instru-
CAT instrument for approximately 3000 students of different ment is particularly useful because it helps faculty understand
class standings at four-year institutions across the country. how they can develop discipline specific assessments that en-
The largest gains in critical thinking appear to occur in the courage students to think critically instead of just encouraging
junior and senior years of undergraduate education. The aver- rote retention. This becomes increasingly important as accre-
age senior score was about 51% of the maximum possible diting agencies such as the Accreditation Board of Engineer-
score on the CAT instrument (maximum score = 38 points). ing and Technology (ABET) increase their focus on efforts to
These scores indicate there is considerable room for improve- improve student learning.
ment. The fact that no student or has obtained a perfect score
or a score of zero at any institution suggests the absence of a REFERENCES
ceiling or floor effect.
[1] Accreditation Board of Engineering and Technology, Criteria for Accre-
CONCLUSION diting Engineering Programs: Effective for Evaluations During the 2006-
2007 Accreditation Cycle. Baltimore, MD: Accreditation Board of Engineer-
ing and Technology, 2005
There have been several significant outcomes of the current [2] D. Bok, Our Underachieving Colleges: A candid look at how much stu-
national dissemination project. We have found that our dis- dents learn and why they should be learning more. Princeton: Princeton
semination model for training representatives at other institu- University Press, 2006.
[3] U.S. Department of Education, National Center for Education Statistics.
tions to use the CAT instrument at regional train-the-trainer The NPEC Sourcebook on Assessment, Volume 1: Definitions and Assess-
workshops is successful. This finding is supported by infor- ment Methods for Critical Thinking, Problem Solving, and Writing, NCES
mation from participant surveys, onsite observations, and 2000—172, prepared by T. Dary Erwin for the Council of the National Post-
scoring accuracy checks. We do, however, continue to modify secondary Education Cooperative Student Outcomes Pilot Working Group:
Cognitive and Intellectual Development. Washington, DC: U.S. Government
our training methods and supporting materials to improve ef- Printing Office, 2000.
fective dissemination. [4] J. Bransford, A. Brown, and R. Cocking, (Eds.) How People Learn:
We have found that many institutions are interested in find- Brain, Mind, Experience, and School. Washington, D.C.: National Academy
ing ways to engage faculty in quality improvement efforts. Press. 2000.
[5] J. Pellegrino, N. Chudowski, and R. Glaser (Eds.) Knowing What Students
Faculty involvement in the scoring of student essay exams Know. Washington, D.C., 2001.
greatly facilitates this process. We are also expanding our ef- [6] B. Stein, A. Haynes, and M. Redding. “Project CAT: Assessing Critical
forts to provide links to institutional and grant funded projects Thinking Skills,” in Proceedings of the National STEM Assessment Confe-
that have positively impacted student performance on the CAT rence, D. Deeds and B. Callen , Eds. NSF, 2007.
[7] P. Ewell. “The examined life: Assessment and the ends of general educa-
instrument. This type of information is just beginning to tion,” Association of American Colleges & Universities General Education
emerge from a broad range of institutional initiatives. We are and Assessment Conference, 2004.
also collaborating with several other NSF funded projects that [8] B. Stein, A. Haynes, M. Redding, T. Ennis, and M. Cecil. “Assessing
are using the CAT instrument to help identify potential student Critical Thinking in STEM and Beyond,” in Innovations in E-learning, In-
struction Technology, Assessment, and Engineering Education, M. Iskander,
gains in critical thinking and real-world problem solving that Ed. Springer, 2007, pp.79-82.
may result from innovative educational pedagogies. [9] B. Stein, A. Haynes, and K. Harris. “Assessing and Improving Critical
The CAT instrument is one of the few interdisciplinary as- Thinking Using the CAT Instrument,” ABET Best Assessment Processes
sessment instruments available that also provides an opportu- Symposium XI. Indianapolis, IN: Accreditation Board of Engineering and
Technology, 2009.
nity for faculty development. By participating in the scoring [10] M. Cecil, “The Relationship Between Student Responses on the Nation-
process, faculty become aware of their students’ weaknesses al Survey of Student Engagement (NSSE) and Performance of the Critical-
Thinking Assessment Test (CAT),” Dissertation, Tennessee Technological
University, 2006.
Enhancing the Instruction of Introductory Electric
Circuit Courses using MATLAB
Thomas Yang, Member, IEEE, Ilteris Demirkiran, Member, IEEE, Jianhua Liu, Member, IEEE,
Albert Helfrick, Senior Member, IEEE, David Pedersen, and Christopher Grant
59
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_11, © Springer Science+Business Media B.V. 2010
60 YANG ET AL.
engineering (AE), Mechanical Engineering (ME), Civil Sample instruction materials are included in Section III of
Engineering (CivE) and Engineering Physics (EP). this paper.
We used MATLAB to illustrate many important concepts in After students were taught how to use MATLAB to plot
circuit analysis, including: desired graphs, they were assigned simple problems for
practice purposes. Sample MATLAB codes were posted on
A. Solving linear systems of equations resulted from Nodal
Blackboard as examples. In MATLAB assignments, we
Analysis and Mesh Analysis
emphasize that the results obtained from computer software
As fundamental circuit analysis methods, Nodal Analysis should confirm, validate and match the results from analysis
and Mesh Analysis utilize Kirchhoff’s current and voltages by hand. For example, the solution of a linear system of
laws to solve linear circuit problems. Often, it is tedious to equations from Nodal Analysis obtained in MATLAB should
solve the resulting linear systems of equations by hand. In match the solution obtained by hand. Also, in determining the
MATLAB, the solution can be obtained via simple commands. types of passive filters, the answer can be found by comparing
In our project, we introduce two straightforward ways to the magnitudes of the transfer function at DC (Direct Current)
perform this task, i.e., solution based on matrix inverse and and at very high frequency. This answer should be confirmed
solution based on Gaussian elimination. by the magnitude response plot generated in MATLAB.
B. Voltage and current across capacitors and inductors In final review of course materials, MATLAB was again
used to assess students’ understanding of key concepts. Here
The current and voltage waveforms are plotted and
are some examples:
compared using MATLAB. The relationship between the
waveforms, which involves differentiation and integration, is
1. To evaluate students’ understanding of the relationship
clearly shown when they are plotted in the same window.
between the voltage and current across capacitors/inductors, a
Therefore, the related formulas become easier to understand
multiple choice question is constructed with MATLAB. For
and memorize.
example, a voltage waveform across a capacitor is plotted
C. AC Circuit Analysis together with three possible current waveforms. Students are
The AC current and voltage waveforms are plotted in the asked to identify which of the three is the current
same MATLAB window. The phase difference of the two corresponding to the given voltage.
waveforms, which is key to understand and analyze AC
circuits, is easily seen. 2. To evaluate if students are able to distinguish three types
of second-order transient responses, i.e., overdamped,
D. First and Second order transient analysis for circuit underdamped and critically damped responses, multiple plots
behavior after switches change position
are generated through MATLAB. Students are asked to
The model to describe transient process is differential determine the type of the transient response for each plot.
equations. Merely solving the differential equations does not
provide much insight into the dynamic evolution of the 3. To assess students’ ability to associate frequency
physical process. Using MATLAB, the transient responses are selectivity of bandpass filters with quality factor and
accurately plotted which give clear illustration of the physical bandwidth, magnitude responses of multiple bandpass filters
meaning of the solution. Moreover, the forced and natural with same center frequency are plotted using MATLAB.
responses are also clearly shown together with the complete Students are asked to order the quality factors and bandwidths
response. of those bandpass filters.
E. Frequency characteristics of passive filters and the
phenomenon of “Resonance” Sample review questions are included in Section 3 of this
paper.
Frequency response of passive filters is critically important
to understand the principles behind modern electronic
equipments. In our project, the magnitude response of the
filter transfer function is plotted using MATLAB. Important
concepts, such as resonance frequency, cutoff frequency,
bandwidth, and quality factor, are all clearly illustrated in the
plots. This makes the understanding of these concepts almost
effortless.
ENHANCING THE INSTRUCTION OF INTRODUCTORY ELECTRIC CIRCUIT COURSES 61
forced response
0.05
1
Current, A
0
0 5 10 15 0.5
time, (ms)
0.4
0
0.2
vL(t), (V)
0 -0.5
-0.2
-1
-0.4 0 0.005 0.01 0.015 0.02 0.025 0.03
0 5 10 15
time, (ms)
Time, s
B. Complete, natural and forced responses of first order D. Using magnitude responses of second-order bandpass
RC transient circuit filters to illustrate the concepts of resonance, quality factor
Complete, natural and forced response of RC circuit
and bandwidth
15
Q=4
Amplitude ratio
0.6 Q=8
0
0.5
0.4
-5
0.3
0.2
-10
0 0.5 1 1.5 2 0.1
time, (s)
0
-2 -1 0 1 2
10 10 10 10 10
Normalized frequency, w/wn
62 YANG ET AL.
E. Sample Review Question: given the capacitor voltage, challenging subject under our new teaching practice.
choose the corresponding capacitor current Assessing their subjective opinions is very helpful also
original capacitor voltage choice A because student motivation is of paramount importance in
6 0.8
STEM (Science, Technology, Engineering, and Math)
i(t), (amperes)
disciplines. Therefore, we designed an end-of-course survey to
v(t), (volts)
0.6
4
0.4 assess our teaching practice from the students’ perspective.
2
0.2 Below is the list of survey questions we used:
0 0
0 2 4
time, (seconds)
6
-6
0 2 4
time, (seconds)
6
-6
1. I can use MATLAB to perform matrix and vector
choice B
x 10
choice C
x 10
operations.
0.8 6
i(t), (amperes)
0.6 i(t), (amperes) 2. I can use MATLAB to solve linear systems of equations
4
0.4 resulted from Nodal Analysis and Mesh Analysis.
2
0.2
3. I can use MATLAB to plot data and mathematical
0 0
0 2 4 6 0 2 4 6 functions such as voltage and current across resistor, capacitor
time, (seconds) time, (seconds)
or inductor.
-6 -6
x 10 x 10
1
6. I have a better understanding of relationships between
0
0 1 2 3 4 5 voltages and currents across capacitors/inductors through
Time, s visualization by MATLAB.
Response of circuit B
1
Current, A
2
Response of circuit C
8. MATLAB plots of second-order transient responses help
Current, A
the overall evaluation of the teaching practice. consistently throughout the course of their college study. If
Survey results indicate that the majority of students believe more engineering courses, especially at the sophomore and
the graphic illustration in MATLAB substantially improves junior level, incorporate MATLAB in their instruction, only a
their understanding of relevant course materials. The average modest amount of effort is need to refresh students’ MATLAB
score for question 12 was 4.5/5. Specifically, the instruction of skills in each course.
second-order transient analysis (question 8) and passive Second, a balance needs to be achieved between the amount
filters’ frequency responses (questions 9-11) benefit most of course materials covered and the amount of time the
from visualization in MATLAB. Also, the plotting of instructor and students devote to MATLAB instruction and
voltage/current across capacitors and inductors (question 6) assignments. In promoting our teaching practice, we need to
helps students quickly grasp the relationship between the be cautioned against letting the MATLAB take too much of
voltage and current waveforms. instruction time and student efforts in assignments. The course
Also, we found students did appreciate the usefulness of materials need to be the top priority, and the MATLAB
MATLAB as a software tool to facilitate efficient solution of element should not increase students’ burden significantly.
electrical engineering problems. However, concerning After all, student motivation should never be compromised,
students’ ability to perform analysis using MATLAB especially for STEM disciplines.
(questions 1-5), there is much variation among students’
responses, with average scores varying between 3.5 to 4.2. REFERENCES
There are many students who are not proficient in using basic [1] MathWorks, “Getting Started with MATLAB”, Natick, MA, 1996.
MATLAB commands. This further confirms the necessity of [2] A. Haubold, “Matlab for First-year College Engineers”, The 37th
ASEE/IEEE Frontiers in Education Conference, pp. F1H-7 – F1H-12,
enhancing MATLAB instruction in engineering curriculum 2007.
consistently throughout undergraduate years. [3] MathWorks, “Learning Matlab”, Version 6, Natick, MA, 2001.
[4] A. Azemi, C. Stoock, “Matlab for First-year College Engineers”, The
26th ASEE/IEEE Frontiers in Education Conference, Vol. 2, pp. 599 –
602, 1996.
V. CONCLUSION/DISCUSSION [5] M. Durán, S. Gallardo, S. Toral, R. Martínez-Torres, F. Barrero, “A
In this paper, we described our recent efforts in integrating learning methodology using Matlab/Simulink for undergraduate
electrical engineering courses attending to learner satisfaction
MATLAB into the instruction of the course “Electrical outcomes”, International Journal of Technology and Design Education,
Engineering I”. It is concluded that the visualization of circuit Vol. 17, No. 1, pp. 55-73, 2007.
analysis concepts in MATLAB facilitates instruction and [6] A. Azemi, E. Yaz, “Utilizing Simulink and MATLAB in a graduate
nonlinear systems analysis course”, The 26th Frontier in Education
promotes student learning. Meanwhile, the new teaching Conference, Vol. 2, pp. 595 – 599, 1996.
practice demonstrated to the students the advantages and [7] E. Yaz and Azemi, “Utilizing MATLAB in Two Graduate Electrical
limitations of computer software in engineering problem Engineering Courses”, The 25th Frontiers in Education Conference,
Vol. 1, pp. 2c6.1-2c6.4, 1995.
solving, and the relationship between the results obtained by [8] S. Ayasun, C. Nwakpa, “Transformer Tests Using MATLAB/Simulink
software tools and by theoretical analysis. and Their Integration into Undergraduate Electric Machinery Courses”,
There are some additional observations based on the 2006 Wiley Periodicals Inc., pp. 142-150, 2006.
[9] V. Pires, J. Silva, “Teaching nonlinear modeling, simulation, and control
findings of this project. Here we discuss two of them: of electronic power converters using Matlab/Simulink”, IEEE
First, the basic MATLAB knowledge students learned in the Transactions on Education, Vol. 45, pp. 253-261, 2002.
freshman year still need to be substantially reviewed in our [10] R. Patel, T. Bhatti, D. Kothari, “MATLAB/Simulink-based transient
stability analysis of a multi-machine power system”, International
project, because students have not been using MATLAB Journal Electrical Engineering Education, Vol. 39, pp. 320-336, 2003.
Istopolis: Design Issues and Evaluation’ Results of
an Integrated Web-Based Educational Application
C. T. Panagiotakopoulos T. S. Papatheodorou G. D. Styliaras
Departmnet of Primary Education, Departmnet of Computer Departmnet of Cultural Heritage
University of Patras, Greece. Engineering and Informatics, Management and New Technologies,
[email protected] University of Patras, Greece. University of Ioannina, Greece.
[email protected] [email protected]
Abstract – In this study the design issues of a web-based These systems offer exceptional and continuously enhanced
application, named Istopolis, are described. Istopolis is an capabilities to educators, who eventually have the potential to
integrated system aiming to support History and Culture communicate with their students and assign them lessons by
courses of primary and secondary education in Greece. This using educational scenarios as well as exercises and projects.
hypermedia application system can be installed on a server
allowing access both to students and teachers. Authorized
Exercises and projects can be reviewed online, providing
teachers are able to modify or add modules with text, virtual students with feedback in terms of educational comments and
material and video, and also modify or add scenarios towards an grades [2] [10]. Thus, the actual virtual class can be realized.
optimum exploitation of the initially specified educational In the present work, the structuring (design and
modules. Furthermore, results of the summative evaluation development) of the Istopolis application is presented. This
obtained by a sample of active teachers are presented, showing application constitutes an integrated web-based hypermedia
that problems such as “lost in cyberspace” have been evaded educational system aiming to support History and Culture
and that, in general, educators are rather positive against the courses of the Greek primary and secondary education. The
developed application in terms of usability, quality of media and structuring and interconnection of the various sub-units of
aesthetics.
Istopolis are also described, which aim at serving all the users
(e.g. students, educators and administrator) interacting with
I. INTRODUCTION
the system. Finally, results of a predictive evaluation are
presented, in terms of usability, aesthetics, quality of included
Online systems aiming to support education are firstly met
media, and hierarchy of learning effect of differentiated
in the decade of ’60 with the appearance of the first online
information transmission media. The data used to perform
bibliographic retrieval system [1]. Systems developed from
this evaluation, were acquired from a sample of 36 educators,
1963 to 1995 could be classified as the first generation of
who used the developed system in real conditions.
educational online systems.
After the wide popularization of Internet (April 1995),
II. ISTOPOLIS APPLICATION: IDENTIFICATION
online educational systems met an increasing implementation,
and can be classified in two categories [2]: (a) oriented to
The name of the presented application comes from parts of
teacher-to-student communication [3] and (b) oriented to
the Greek words Isto-ria (history) and Poli-tismos (culture). It
student-to-student communication [4]. The generic attributes
is an integrated hypermedia web-based educational system
of systems from both categories present many similarities
constructed during 2007-2008 in the context of the Unit
with those of the first generation online educational systems.
“Chrysallides” (Development of integrated educational
Along with the emergence of the second generation of
activity products), of the Action “Pleiades” (Development of
systems an effort to support the learning process took place.
Educational Software and Integrated Educational Activity
This effort was mainly based on elements of the traditional
packages for Greek Primary and Secondary Schools &
class [2].
Distribution of Educational Software Products to Schools),
With the wide diffusion of Internet and WWW, online
with the Research Academic Computer Technology Institute
systems started to be employed as sources of information
(RACTI http://www.cti.gr) as the organization responsible for
retrieval [5] and means of design and development of learning
implementation monitoring. It was funded by the Greek
environments [6] [7].
Ministry of Education & Religious Affairs and the
The third generation of web-based learning systems or
Operational Program for the Information Society, Measure
online educational systems, as they are met in the literature, is
1.2, 3rd Community Support Framework.
based on novelties regarding the means of control and the
The system was integrated through a continuous formative
direction of information flow [8]. Contemporary educational
evaluation from the respective organization responsible for
online systems are nowadays exclusively based on Internet.
monitoring and controlling (i.e. RACTI) technical and
Along with these systems, the term “virtual teaching”
educational issues. It was finally positively evaluated by the
emerged [9].
Greek Pedagogical Institute from a technical as well as an
65
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_12, © Springer Science+Business Media B.V. 2010
66 PANAGIOTAKOPOULOS ET AL.
parameters about assigned projects. The system’s content is within a specific period of time, determined by the user.
generated on the fly by using PHP scripts that draw the Finally, through the “ap” choice modules and
appropriate content from the database. Exercises are complementary texts are classified according to their
implemented with Javascript and Java. The timeline is based association with a geographical area. If the module or the text
on Javascript and Dynamic HTML, whereas Flash is can be associated with a certain location, it then appears
employed for implementing the map interface. Flash positioned as a point on the map. Otherwise, it appears under
communicates directly with the database via Javascript in the map of the region it was associated with.
order to display the appropriate map and the points on the On the other hand, the user can access a specific content in
map. Flash is used also for displaying the virtual interface and a combinative way through the “Timeline” and the “Map”. In
the menu on the upper right of every screen. Finally, the other words the user can access a specific content through the
content management interface is based on PHP scripts. combined use of time and space.
As many researchers in the domain of hypermedia
IV. STUDENT’S INTERFACE applications have already reported, due to several design
issues, the phenomenon of “lost in cyberspace” has been
The application’s information flow, from the user’s observed. This phenomenon disorganizes and misleads the
perspective, depends on the kind of the user (administrator, user [12]. To avoid such problems, we did not connect with
teacher, and student). hyperlinks any words or phrases within the main and the
The general pattern of the information flow concerning the complementary text of each module. Instead, we used a term
student’s access to the software’s content is shown in Fig. 2. dictionary in every module that the user can access by
pressing a button in each module’s environment. All the small
dictionaries of the modules act as subsets of a bigger
dictionary that the user can find in the root menu and
approach it alphabetically. Moreover, all hyperlinks of the
module’s menu operate and expand inside the specific
module, apart from the “Resources”, the “References” and the
“Links”.
A. Content’s access
The student can access the software’s content in three
Fig. 3. An actual screenshot of the student’s interface.
different ways – choices. These choices are the “Thematic
Catalogue”, the “Timeline” and the “Map” (Fig. 2, Fig. 3).
B. Student’s Book
The shortest way to access the content is through the
An important “field” of the application from the student’s
“Thematic Catalogue” choice. Within the “Thematic
perspective, is the “Student’s Book”. In that “field” the
Catalogue” modules and complementary texts are classified
student can find:
through their positioning in a thematic hierarchy.
a) The scenarios that his teacher sent him for processing.
With the “imeline” choice, modules and complementary
b) The tasks that the student has fulfilled (exercises and
texts are classified through their positioning in chronological
assignments) and wants to be published in order to be
periods. This classification enables both unit and text search
68 PANAGIOTAKOPOULOS ET AL.
evaluated by his teacher. collection of the student’s projects, evaluation, and provision
Note, that the field “Student’s Book” is accessible by all of grades and constructive feedback to the students.
authorized users. This fact emerged under the consideration b) The construction of new scenarios.
that in this way, a better competition between the students Information flow of the teacher’s interface is depicted in
will be created and maintained since all of their projects and Fig. 5:
worked out exercises will be in public view.
Moreover, in that field every student can view the
evaluation of the projects submitted. Evaluation includes both
the project grade and the comments justifying the received
grade.
V. TEACHER’S INTERFACE
VI. INTERACTION
The teacher can be informed by the system for the
According to several researchers [14], the interaction
scenarios already assigned to a class in order to be effectuated between educational software and user is an essential element
along with those that remain unassigned. He is also able to
positively contributing to the learning process. Istopolis offers
extract a scenario in XML format, change or enhance it and
several capabilities concerning the interaction with the
import the modified scenario back to the server. The online
student, which include:
tools the teacher is offered are associated with:
1. Navigation in 3D space in order to discover the
a) The class administration, such as: assignment of a
software’s research and configuration tools, after the
scenario or a set of scenarios to a class, establishment of an
student has logged in. This way the students are
appropriate hierarchy for the scenarios to be accomplished
welcomed by an environment which follows a gaming
when more than one scenarios are assigned to a class,
fashion that is more close to the student.
ISTOPOLIS: DESIGN ISSUES AND EVALUATION’ RESULTS 69
2. Content search using simple research methods, such as standard deviation and the mode (the value that occurs with
the modules’ catalogue. the highest frequency) with the respective rates are shown in
3. Content search using complex search tools, coupling Table 1.
intense semantic charge with simple graphic interface
(such as the Thematic Catalogue, the Timeline or the TABLE 1
QUESTION, MEAN, STANDARD DEVIATION AND MODE
Map) and exploit the intrinsic hierarchical information Question (usability) Mean St.D. Mode
structuring. U1. At the very beginning working with Istopolis 5
3.97 0.97
4. Content search through complex combinations of search was: (36.1%)
criteria regarding "subject - location - time". U2. In general, you found the software’s 4
4.13 0.76
handling: (41.7%)
5. Access to same content through several paths. U3. Did you understood the way this software 4
6. Project writing through a small hypertext processor with 3.92 0.81
operates: (52.8%)
complete editing capabilities and project publishing. U4.Do you think that the students will be able to 3
3.58 0.69
7. Content collection, temporary storage and reuse in the use it: (52.8%)
projects. U5. Can you create a scenario with exercises 4
3.56 0.98
and projects through the software: (33.3%)
8. Fill out and submission of self-evaluation exercises. 4
The above are only the most important from the whole of U6. Can you assign a scenario to the students: 3.92 0,84
(47.2%)
the interactive functionalities offered to the students. U7. The software favors interaction between
4.17 0.74
4
Regarding the student-educator interaction, it is noted that student and content: (52.8%)
U8. Navigation in the software’s units was 5
the software is designed to be used inside the class, while the performed:
4.44 0.59
(50.0%)
educator is present supervising the students. U9. The use of the search tools (Thematic 5
4.21 0.72
Catalogue, Timeline, Map) is: (69.4%)
VI. APPLICATION’S EVALUATION U10. Functionalities of the “Student’s Book” for
4
the teacher and the student is in terms of 4.03 0.61
(63.9%)
usability:
The application, during its implementation, was structured U11. Did you feel “lost” at any time, while 1
through a continuous formative evaluation by the supervising 1.50 0.56
navigating in the software: (52.8%)
organization as well as the development team. The summative U12. Did you know at all times your exact
4.19 0.78
5
evaluation was performed with multiple groups of users and position in the software: (41.7%)
multiple evaluation subjects.
In the following, the results of a predictive summative Usability effects were analysed using a chi-square test.
evaluation received from a sample of 36 active teachers of Teachers found the software easy to handle (U3: x2=17.11;
primary and secondary education are presented. df=3; p< .001). This result is replicated when teachers
The evaluation procedure was performed with the sample comment whether their students will be as well able to handle
divided (for practical reasons) into two groups. The software it (U4: x2=9.5; df=2; p< .01). As a result, the high level of
was initially presented to each group by one researcher for usability makes educators feel confident to assign scenarios to
two teaching hours, while each participant was working at a their students (U6: x2=12.67; df=3; p< .01). From the
personal computer. After the presentation the participants teachers’ point of view, the navigation through modules poses
were asked to work with the developed software for three no usability constraints (U8: x2= 17.11; df=3; p< .001).
teaching hours. Finally, each participant was asked to provide In addition, the sample was asked to answer two more
feedback on specific questions filling an appropriate questions:
anonymous questionnaire. - In general, were there any points, where due to difficulties
The variables of the scale used for the evaluation [14] [15], in using the software you finally did not work with it?
aimed in collecting data regarding the applications’ usability, - In general, did you find any problems using the software
the aesthetics, the quality of the media included, and the or not?
establishment of an hierarchy for the learning effect of the Answers regarding the 1st question were 97,2% negative.
information transmission media from the teacher’s This means that only one subject claimed that there was a
perspective. point where working with the software was not possible due
Analysis of the self-evaluation of the participants’ to difficulties in its use. However, when this subject was
knowledge in basic issues of computer use (Microsoft asked with a follow-up question to spot this point, no answer
Windows/Office, Internet) showed an average of 3.9 in the was given. Answers regarding the 2nd question were 100%
five grade Likert scale (knowledge: none = 1, 2, 3, 4, 5 = negative.
excellent) with st.d. 0.7. The findings described above, show that the participants in
the evaluation believe that the software is easy and flexible in
A. Usability its use. Furthermore, the software’s design met with success
We asked the sample to provide answers to specific regarding the aim to avoid the “lost in cyberspace”
questions with expected answers within the five-grade Likert phenomenon.
scale (not at all = 1, 2, 3, 4, 5 = absolutely or difficult = 1, 2,
3, 4, 5 = easy). The questions, the answer’s average, the B. Aesthetics and quality of media
The sample was asked to evaluate the software’s aesthetics
70 PANAGIOTAKOPOULOS ET AL.
and the quality of information transmission media used by it. application, in its current form, comprises an integrated
The outcome is presented below (Table 2): platform aiming to support History and Culture courses of
primary and secondary education in Greece. It is a web-based
TABLE 2 application offering the potential to modify or enhance its
QUESTION, MEAN, STANDARD DEVIATION AND MODE
Question (aesthetics – quality of media) Mean St.D. Mode content through a CMS. Using the presented application, the
1Have you found image and video quality fair 5 educators are able to assign exercises and projects to their
4.08 0.97
enough? (41.7%) students through educational scenarios, evaluate them online
2. In which level did you find the software’s 5 and provide them with grades and constructive feedback.
4.02 0.91
images good regarding their quality? (38.9%)
3. Was the display’s color aesthetics 4
The results of the final predictive summative evaluation
3.92 0.73 with a sample of 36 teachers showed that the developed
appealing? (47.2%)
4. How much distinctive – readable was the 4 application benefits the educational process while being
3.97 0.81
text? (36.1%) handy, very appealing in terms of aesthetics and of high
5. Are the design and the presentation of the 4 quality with regards to the quality of information transmission
3.94 0.83
application’s screen appealing? (41.7%)
6. Is there cohesion regarding the presentation 4 media.
3.89 0.75
of the screen in every module? (52.8%)
A7. In general, how do you consider the
3.94 0.86
4 REFERENCES
software from an aesthetics perspective ? (36.9%)
[1] C. P. Bourne, and B. H. Trudi, “A History of Online Information
In general, application’s aesthetics were assessed as being Services, 1963–1976”, Cambridge, Mass.: MIT Press, 2003.
[2] R. Sheizaf, B. Miri, D-G. Yuval, and T. Eran, “QSIA – a Web-based
of high quality (A7: x2=9.78; df=3; p< .05).
environment for learning, assessing and knowledge sharing in
Establishment of a hierarchy regarding the quality of the communities”, Computers & Education 43, (3) 273–289, 2003.
information transmission media’ factors (text, pictures, [3] Y. J. Dori, M. Barak, & N. Adir, “A Web-based chemistry course as a
graphics-maps, videos) was analyzed with the assistance of means to foster freshmen learning”, Journal of Chemical Education,
80(9), 1084–1092, 2003.
the Friedman’s test and showed an important advantage of
[4] M. Guzdial, and J. Turns, “Effective discussion through a computer-
video (x2=25.43; df=3; p< .001) followed by pictures (Table mediated anchored forum”, The Journal of the Learning Sciences, 9(4),
3). 437–469, 2000.
[5] C. A. Merlic, B. C. Fam, M. M. Miller, “WebSpectra: Online NMR and
IR Spectra for Students”, Journal of Chemical Education, 78, 118–120,
C. Hierarchy of information transmission media’ learning
2001.
effect [6] Dori, Y. J.; Barak, M., “Virtual and physical molecular modeling:
The sample was asked to give their opinion for which Fostering model perception and spatial understanding”, Educational
information transmission media is considered as helpful for Technology & Society, 4, 61–74, 2001.
[7] Donovan, W. J.; Nakhleh, M. B., “Students’ Use of Web-Based
the student in order to better understand the concepts in the
Tutorial Materials and Their Understanding of Chemistry Concepts”,
developed application. Data analysis through the Friedman’s Journal of Chemical Education, 78, 975–980, 2001.
test showed again an important advantage of video (x2=19.23; [8] Nachmias, R., Mioduser, D., Oren, A., & Ram, J., “Web-supported
df=3; p< .001) followed by pictures (Table 3). emergent-collaboration in higher education courses”, Educational
Technology & Society, 3(3), 94–104, 2000.
[9] Sheremetov, L., & Arenas, A. G., “EVA: An interactive Web-based
TABLE 3
RANK OF MEDIA’ QUALITY AND EFFECT collaborative learning environment. Computers & Education, 39, 161–
Media Quality - Mean Rank Effect - Mean Rank 182, 2002.
Videos 1.92 1.72 [10] Panagiotakopoulos, C., Papaterpos, C., & Papatheodorou, Th.,
Pictures 2.06 2.53 “Presentation of a hypermedia system supporting History: Basic
characteristics and content’s evaluation”, Proceeding of 1st Educational
Graphics – maps 2.78 2.83
Conference for the Introduction and Use of ICT in Education (P.
Text 3.25 2.92
Politis, ed.), 759-765, 2009.
[11] Rafaeli, S., Barak, M., Dan-Gur, Y. & Toch E., “QSIA – a Web-based
The previous results show for one more time that an image environment for learning, assessing and knowledge sharing in
on its own is able to paint a thousand words. For this reason, communities”, Computers & Education, 43, 273–289, 2004.
[12] Otter, M. & Johnson, H., “Lost in hyperspace: metrics and mental
during the design of an educational material this should be models”, Interacting with Computers, 13(1), 1-40, 2000.
seriously taken under consideration [16]. [13] Panagiotakopoulos, C., Pierrakeas, C. & Pintelas, P., “Educational
Finally, all participated teachers, who answered a relative Software Design”, Patra: Hellenic Open University Publications, 2005.
question, stated that they would like to use the developed [14] Picciano, A.G., “Beyond student perceptions: Issues of interaction,
presence, and performance in an online course” Journal of
software in their classroom, as part of their everyday teaching Asynchronous learning networks, 6(1), 21-40, 2002.
practice. [15] Panagiotakopoulos, C., Pierrakeas, C. & Pintelas, P., “Educational
Software and its Evaluation”, Athens: Metaihmio Publications, 2003.
VII. CONCLUSIONS [16] Spiliotopoulou-Papantoniou, V., Karatrantou, A., Panagiotakopoulos,
C. & Koustourakis, G., “Visual representations of the internet in greek
school textbooks and students’ experiences”, Education and
In this study, issues regarding the design and development Information Technology, 14, 205-227, 2009.
of the application named Istopolis were analyzed. This
TEAC2H-RI: Educational Robotic Platform for
Improving Teaching-Learning Processes of
Technology in Developing Countries
Juan J. González1, PhD Jovani A. Jiménez2, PhD Demetrio A. Ovalle3
National University of Colombia
Carrera 80 numero 65-223 (M8A-309) Medellín-Colombia
1
[email protected], [email protected]; [email protected]
Abstract—The technological development achieved in each of the Pedagogy, which also claim, as the constructivism theory, that
epochs of the human history has changed the means of the knowledge is not transferred but created in the learner’s
production, communication, transportation and education. There mind, but additionally, that the method to achieve that is
are many results because of the incorporation of technology on through the construction of something tangible by the learner
education, including remote laboratories, online courses, and [3] [4].
educational robotics. In recent years, the latter has become very
Papert and his research group at the Massachusetts Institute
popular because of its significant contributions to the learning
process, strengthening of cognitive skills, including creativity, of Technology (MIT), based on the Construccionism
design and problem solving skills. However, this phenomenon has Pedagogy, developed a programming language for kids, called
been focused in the developed countries, and many of the Logo. Logo had a Graphic User Interface which allowed
applications of the developing countries rely on a technological programming a turtle robot. Afterward, LEGO and
transference from the developed countries. This model is not the Fischertecnnik with a similar concept developed kits to build
most suitable due to the fact that the tool should be adapted to the robots using many bricks, including tires, gears, lights,
environment and its user, and not the opposite. This article sensors. The Media Lab of the MIT and other institutions used
presents some of the main contributions in this issue and outlines these kits to explore the results of the interaction with them.
the TEAC2H-RI kit. This kit was developed based on the
Most of the researchers concluded that theses kits not only are
Colombian educational environment characteristics and the
experience in robotics of the research group GIDIA. It will be useful for the learning process but also for the strengthening of
used in Senior High School Students. the creativity, design, troubleshooting, team work and social
awareness, among other cognitive skills.
Index Terms—Society and Computing, Educational Robotics, These kits produced by Develop Countries have been
Robotic Swarm, Constructivism, Construccionism implemented in these regions, and in some cases in
Developing Countries. Although, the last one have social-
I. INTRODUCTION economic conditions very different from those of the formers,
As well as the societies evolve so do its means of then the performance of these kits without the correct reading
production, communication, transportation and education. In of the context could generate no satisfactory results. It is
the last one, one of the most important moments was in the because by this methodology the user and his environment has
XIX century when many researchers proposed the switch from to be adapted to the educational tool, but it is the tool that
the passive paradigm to an active paradigm where the should be adapted to the user and his environment. This
participation of the learner is more active. It comes as a result phenomenon is very common in the technological
of the works of pedagogues, educators and psychologist, implementation in the Latin American Countries and it is very
including, Johann Heinrich Pestalozzi, Friedrich Froebel and criticized by many different authors. [5]
Maria Montessori [1]. These previously mentioned In order to provide an autochthonous solution to this
approaches, along with the subsequent, are the root of the problematic the Artificial Intelligence in Education Research
development of the constructivism theory of the Swiss Group of the National University of Colombia addresses these
Psychology Jean Piaget. In this theory the knowledge is not difficulties in the research project: Educational Robotics:
assume like a fluid which can be transferred from one person Intelligent Machine in Education.
to another, but like a process where the knowledge is created This paper is organized as follows: Section II briefly
in the learner’s mind [2]. Based on it, the South African reviews some approaches in Educational Robots. Section III
mathematician Seymour Papert proposed the Construccionism presents some cases of the implementation of educational
robots in developing countries. Section IV describes the
71
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_13, © Springer Science+Business Media B.V. 2010
72 GONZÁLEZ ET AL.
proposed methodology. Section V presents the kit TEAC2H- developed and Latin American countries. There is very clear
RI (Technology, Engineering And sCience eduCation through that the research crisis covered almost the whole region, and is
Human-Robot Interaction). Finally, section five presents the less severe in Chile, Brazil, Argentina and México. The
conclusions and future work. research capacity of any country is one of the most important
resources to affront the social, ecological, technological, and
II. EDUCATIONAL ROBOTICS economical challenges. Therefore, it can be understood why
Since the creation of the Logo Turtle in the 60’s by the the countries with a low number of researchers per million
Media Lab MIT until our days the advancement in Educational inhabitants, have a lower progress than those with a greater
Robotics has covered not only the primary school students but number of researchers.
also the high school, undergraduate and graduate students, and TABLE 1
even the industry workers. Some of the most important RESEARCHERS PER 1 000 000 INHABITANTS (TAKEN FROM [12])
educational kits are: LEGO Minstorms, Logo, Fischertechnik Researchers per 1 000 000 inhabitants (FTE)
[6], Pioneer, K-Team, IroRobot y Handy Board [7], [8], [9]. Argentina 980
The three formers are used in education and toy industry. On
Bolivia 120
the other hand, the last five could be implemented in the
Brazil 629
school, university and in the industry. These kits are very
useful to explain and verify algorithms, to learn some concepts Canada 4157
of programming languages and to encourage the strengthening Chile 833
of different cognitive skills [10]. Colombia 151
From the different options available LEGO Minstorms and Ecuador 69
Fischertechnik are the most widely used in the world because Guatemala 25
its relative low cost, in comparison with the others, and the Japan 5573
educational resources that provide. The philosophy of both is
Mexico 460
very similar because is based on interlocking bricks. One of
Panama 62
these kits is RoboExplorer. It is manufactured by
Fischertecnik, and it is consists of ultrasonic range sensors, a UK 2881
color sensor, a track sensor, a NTC resistor and a USA 4663
photoresistor. The locomotion is based on two servomotors
and a crawler drive. Also, with this kit is provided the ROBO Aware of this fact, some research centres and universities
interface, and the programming software ROBOPro [11] have implemented different educational robotics courses, to
As can be seen the characteristics of these kits made of find a solution to these issues. Some of them are the Carnegie
them a very useful tool in the teaching of important subjects as Mellon University (CMU) and the MIT. Next is presented
mathematics, physics, electronics, and algorithms. These kits some of their results
at the university, in careers related with Computer Science, In Qatar and Ghana the Carnegie Mellon University
contribute with a deeper understanding of the algorithms associated with some higher educational institutions of these
concepts. On the other hand, in industry there are many regions, developed a course on autonomous robots with the
applications, including the process simulation, training in the aim of strengthening the knowledge on algorithms and the
control machine. However, its implementation must be creativity in the undergraduate Computer Science students.
adjusted to the particular context of the region. The following The educational robotic tool was the MIT Handy Board, the
section presents some approaches in this matter. software was the Interactive C and the hardware the bricks of
LEGO. At the end of the courses was concluded that robotics
III. EDUCATIONAL ROBOTICS IN DEVELOPING COUNTRIES as an educational tool in Computer Science meet the proposed
objectives and also increase the interest of the learners by
The great deal of problems in the majority of the developing technological topics and research. Another important
countries, especially in the Latin American Countries, requires conclusion was that the interaction, for the goal
that the inhabitants face them not only with a solid knowledge accomplishment, among the team members strongly depends
in their speciality area but also with high research skills. on the cultural environment. This fact has to be observed at the
Furthermore, it is necessary to increase the attraction of school design stage of a course in this thematic [13].
students for careers related with technologic production, such In Costa Rica the Massachussets Institute of Technology
as engineering. Statistically (see Table 1), it can be seen that (MIT) worked with a school located in one of the rural zone of
there is a research crisis in Colombia and in almost all the this country. The educational robotic tool was the LEGO
Latin American Countries, e.g. in Colombia in 2006 there was Mindstorms and the software was the Yellow Brick Logo.
only 151 researchers per million inhabitants, whilst for the From the results some of the conclusions were that educational
same year in USA and Japan there is 4663 y 5573 researchers robotics is a very useful tool for the student learning process
per million inhabitants, respectively [12]. In the table I, can be support and also for the strengthening of the creativity and the
seen the number of researchers per million inhabitants in troubleshooting skills. [14].
TEAC2H-RI: EDUCATIONAL ROBOTIC PLATFORM 73
As was illustrated before there are many robotic educational demonstrated that the active role of the student in his/her
tools, however, because they were developed in countries with learning process can be done through the interaction with the
socio-economical conditions very different from the Latin kit TEAC2H-RI.
American countries, its effectiveness is lower in those In the second workshop the objective is to increase the
countries than a autochthonous propose made from a fair motivation of the learners for robotics, especially for the kit
reading of the context. Moreover, they are very expensive in TEAC2H-RI. The success of the course depends on this fact
comparison with the economical resources of the majority of then it should be handled with caution. First of all, we will
the educational institutions of these countries, restricting the present a movie or documental about robotics. Then, they will
widely spread of this tool. Due to this fact, it is necessary to be induced to observe and analyze the behavior of the robotic
make a deeper reading of the Latin American reality and bring swarm. After, teams of four students will be formed and each
an autochthonous solution according to the region necessities of them will work with a tutor and a robot. It is done in order
[5]. From the educational view, one important constraint of the to make of robotics meaningful material for them [16].
majority of these kits is the specific number of pieces that they In the third workshop the objective is that the learners
have, which influence the mind of the learner to develop some understand the theories that rule the behaviour of the robots
specific prototypes. Technically, one of the most important through the modification of the parameters of them. The tutor
constraints is that each brick has a group of equations that will stimulate the interest of the learners to interact with the
rules its behavior which can not be modified by the user [15], kit, and also he/she will provoke situations where the learners
additionally the learner can not integrate new bricks of his/her could conclude some important theories that rule robotics. For
own creation. These constraints restrict the learning process, example, the tutor will incite the learners to move the batteries
and the strengthening of creativity and design. position and observe how the change in the centre of mass
affects the robot’s stability. There are many other important
IV. PROPOSED METHODOLOGY concepts that could be understood through the interaction with
the kit, including overlapping of light beams, angular
The Artificial Intelligence in Education Research Group of momentum, friction, resistance [17]. Some of the skills that
the National University of Colombia, addressed the will be strengthened are analyses, problem solving, and the
difficulties mentioned above through the research project, team work.
Educational Robotics: Intelligent Machine in Education. This In the next workshop the objective is that the teams design,
project is ongoing. In this project, based on a objective view of and develop an addition, improvement, or a modification on
the Colombian reality and the experience in robotics of the the robotic platform. Also, the learners could choose to design
Research and Development Group in Artificial Intelligence of and develop a small automation system which must be
the same institution, was developed the robotic swarm associated with the topics of the course. For both activities, the
TEAC2H-RI (Technology, Engineering And sCience tutor will be a supporter to the learner during his design and
eduCation through Human-Robot Interaction) which will be construction stage. The skills to be strengthened in this
implemented in the Senior High School students of the workshop are the design, cooperation, creativity and research.
Metropolitan Area of Medellin. In the fifth workshop the objective is to strengthen the
In this zone will be chose ten schools and to each of them writing skills. In order to do that, each student will be asked to
will be delivered a kit TEAC2H-RI, composed of 4 robots. To write an article, which must contain abstract, introduction,
be eligible, each school must fulfil with some minimum theoretical support and their experience in the previous
requirements, i.e. a computer room. If the school not fulfils workshops. Other issues that have to be mentioned in the
this requirement, our university will provide a special room for article are: “what are the characteristics of the device that you
this purpose. Each course will have duration of 20 hours developed in the previous workshop?”, “Name an industrial
distributed in five weeks, with seven workshops. sector where it can be implemented”, “Why did you choose
The aim of the first workshop is to get the commitment of this design over others?” among others. Finally, the article
the teachers with this process, by means of the increasing of should have the conclusions and references section. This
their comfort with the topics of the course and dismissing the strategy is used because people like to write about their
negatives or technophobic paradigms. This will ensure that, interests [18].
even when the course has finished, the tool remains its The last objective is to promote the research discussion
pertinence to be used in the institution. spaces among the students. For this purpose, each group will
For this purpose, it will be presented in the workshop, the present, in front of their partners, the results obtained in the
topics to be studied and the expected results. Additionally, it fourth workshop. The other students in the audience will be
will be illustrated some pedagogies and theories of active encouraged to criticize constructively the projects, to let the
education, such as the Piaget’s Constructivism theory, the team to identify the improvements or limitations of the
Ausubel’s learning theory and the Contructionism of Papert. prototype. It is also with the aim to strengthen the team ability
Afterward, it will be exposed and argued the reasons to use to appreciate the diverse points of view, and as a consequence
active pedagogies of education as their teaching methods expand his community sense [17], [13].
supporting this in the social, economical and productive Finally, the Artificial Intelligence in Education Research
necessities of the Colombian context. Finally, it will be Group, based on the previous experience in educational topics,
74 GONZÁLEZ ET AL.
will develop the test to evaluate the effect of the course in the groups, making with this that the group and intergroup
learners, testing the performance of the students before and interaction skills be strengthened. This is the fundamental
after of their participation in the course. model of relations in enterprise and research projects. One
Despite de fact that the course finishes in this stage, the deeper explanation is giving next: every member of the group
support of our group will continue, because through the web interact with his/her partner to get the desired result in the
page, http://xue.unalmed.edu.co/~dbetanc0/index-3.html, we robotic agent, but also the interactions of each team with the
will present updated information about the kits, such as others must be coherent to get the desired behaviour of the
designs, algorithms or manuals that could improve the robotic swarm.
performance of the robot. Furthermore, the students could Following, it is explained each component of the robotic
interact with the Developer research group of the kit through swarm.
the web page or the e-mail, facilitating in this way the
feedback of the concepts and validity of the kit through time. A. Sensors
Next, it is mentioned some software and hardware
characteristics of the kit TEAC2H-RI. Traditionally, the educational robotic systems for the school
do not accept devices of other manufacturer. This fact makes
V. EDUCATIONAL ROBOTIC KIT: TEAC2H-RI difficult the access to spare parts and the design of new
2
prototypes. On the contrary, TEAC2H-RI is an open and
TEAC H-RI is a robotic swarm composed of four robotic modular architecture, i.e. the accessories that come with the kit
agents, i.e. three sons and one mother. It is a hybrid system can be improved or replaced. Even more, new accessories can
because it has distributed and centralized characteristics. It is be added, being therefore more adaptable to other
distributed because the son-robots, with similar behaviour as technologies. These concepts are very important for the
the mother-robot, could interact among them. It is centralized strengthening of the design, creativity and development skills
because the mother-robot could perform the role of group in the user. The kit has five different sensors for each robotic
leader. See Fig. 1 where a mother-robot is interacting with a agent. In these sensors the learner can experiment and modify
son-robot. the parameters of the equations that rule them. Also, additional
The robots not only interact among them, but also with the pieces are provided to the learner, in order that he/she may
user by means of sound signals (buzzer) and light (LEDs). It develop new sensors. In Fig. 2 it can be seen the infrared
facilitates to the user to conclude if the sensor connections or sensor. In 2a the emitter of the infrared sensor does not have
the robot behaviour is appropriated, and hence take the any resistance. In 2b a resistance of 10ȍ is put on the emitter.
corrective actions [19]. In 2c the Velcro, used to hold the sensor to the robot, can be
Each robot, when has been properly configured by the user, observed. In 2d the sensor is in the front side of the robot. Due
it could carry out navigation activities and obstacle avoidance. to the fact that the resistance of the infrared sensor emitter can
Even so, the mother-robot is the only one that could also send be changed, the maximum distance sensed can be modified. In
and receive the information from and to a remote PC, either by this manner and based on the textbook, the concept of
way of a radio link or a data cable. The modes of navigation electrical resistance, current and voltage can be understood.
are: track follower, by means of the QRB1114, obstacles
avoidance through infrared sensors or touch sensors.
a b
c d
This robotic swarm will be used in the workshops with the
schools. In each workshop the maximum number of Fig. 2 Infrared Sensor a) without resistance b) with a 10ȍ resistance c)
participants is 16, divided uniformly in four groups and a with the Velcro at the bottom side d) in the front side of the robot
robot assigned for each team. This work topology with the kit
TEAC2H-RI facilitates not only the cooperation among the
members of each team, but also the cooperation among the
TEAC2H-RI: EDUCATIONAL ROBOTIC PLATFORM 75
In the kit each robot has infrared sensors to avoid obstacles phobia for the technological topics [6]. The figure 3 was taken
or to follow tracks, photoresistance to recognize day and night, from the textbook.
touch sensor to avoid obstacles, sound sensor to detect claps,
and also it can be added other sensors including, temperature VI. CONCLUSION AND FUTURE WORK
sensor using NTC resistance and ultrasonic range sensor. The production cost of each robot is around $65US, which
is very attractive in comparison with other options in the
B. Mechanical Design market. This fact facilitates the expansion of the educational
robotics to a wider number of regions.
The chassis of both robots is very similar; its size is 23cm The implementation of educational robotics in developing
long, 14cm wide and 10cm high. The chassis has a first level countries must be done with a previous reading of the
where the motors card is and a second level where the Main economical and social context. It is important to select the best
Board is. In the outline of both levels, it is glued male Velcro. tool that fit the reality. In our project, this reading was done
It allows changing the position of each sensor and explores the and a new educational tool was developed: TEAC2H-RI.
operation in function of the position. The design of the chassis, The open architecture of TEAC2H-RI allows implementing
which could be assembled or disassembled, is ideal for new devices in the robot. This fact is very important because
experimentation and understanding of physical and the learner would like to improve the performance of his/her
mathematical concepts. For example, when it is modified the robot through the design and development of new sensors or
distance between the motors, the position of the sensors or the actuators. This characteristic, strengthen the creativity and
distribution of the mass in the platform , it could be design skills.
understood some concepts, including centre of mass, torque The proposed methodology is related to the steps to follow
and beam interference. An additional advantage of the in a research project. In this way, implementing this
mechanical design of TEAC2H-RI is that its components are methodology could be strengthened the research skills of the
very cheap and easy to acquire in Colombian market. Fig. 3, learner. It is extremely important in the developing countries,
shows the mechanical construction of the chassis of one like Colombia.
TEAC2H-RI robot. In this first approach, the social group chose to work on is
the Senior High School students from the Metropolitan Area
of Valle de Aburra – Antioquia – Colombia. However, future
works will cover not only the Metropolitan Area but also other
Colombian and Latin American places.
In upcoming projects will be developed new programming
languages with an easier interaction of the learners with the
educational tool TEAC2H-RI. Additionally, it will be brought
an introduction in the programming world.
Currently, there is a human-robot interaction but it is very
weak and it is necessary to consider the addition of new
hardware and software devices which must adapt the kit to the
user necessities [19].
Fig. 3. First construction stage of the chassis of one TEAC2H-RI robot. The kit TEAC2H-RI should not be taken as the panacea for
the educational problems in Colombia, neither in the
C. Programming Language developing countries, because it is necessary the collaborative
work between Government, School, University and Private
It will be used the programming language CodeWarrior of Sector to evolve to the suitable educational model [5, 3], in
the Metrowerks Company, which is supported in C. However, which the teacher will give a greater commitment and effort
we are considering developing an easier programming but also will receive better fruits.
language to the user, which not required previous knowledge
in algorithms and give an introduction to this subject. [20] [21] REFERENCES
D. Textbook
[1] Zuckerman O., “Historical Overview and Classification of Traditional
and Digital Learning Objects”. MIT Media Laboratory, 2006
The textbook contains the topics of physics, mathematics, [2] Piaget, J. Inhelder, B. 1966. La psychologie de L’enfant. Paris. P.U.F.
electronics, algorithms, and artificial intelligence that will be [3] Papert S. 1990. A Critique of Technocentrism in Thinking about the
addressed in the course. This is one of the most important School of the Future.1990 M.I.T. Media Lab Epistemology and Learning
Memo No. 2.
issues, which has to be handled carefully, because require a
[4] Papert S. 1980. The Gears of my Childhood. Mindstorms: Children,
language, concepts and adequate examples, to catch the Computers, and Powerful Ideas. Basic Books, New York.
attention of the learners. If it is not regarded with great [5] UN Millennium Project 2005. Innovation: Applying Knowledge in
caution, the reverse effect can be caused, i.e. rejection or development. Task Force on Science, Technology, and Innovation.
76 GONZÁLEZ ET AL.
[6] Chiou A. 2004. Teaching Technology Using Educational Robotics. En: [14] Urrea C. 2001. Designing Robotic Artifacts: Rural School-Community
Scholarly Inquiry in Flexible Science Teaching and Learning Collaboration for Learning. 2001 Media Lab MIT.
Symposium. (Sydney, Australia, Octubre 1, de 2004) [15] Zuckerman O., Resnick M. 2003. System Blocks: A Physical Interface
[7] Anderson M., Thaete L, Wiegand N. Player/Stage: A Unifying Paradigm for System Dynamics Learning. MIT Media Laboratory. 2003.
to Improve Robotics Education Delivery. RSS Workshop on Research on [16] Ausubel, D.P. 1963. The Psychology of Meaningful Verbal Learning: An
Robotics in Education 2007 (Georgia, USA, June 30 , 2007) Introduction to School Learning, Grune & Stratton Inc., New York, 1963
[8] Barrera P., Martin F., Robles G. Cañas J., Matellan V. 2005. Modelos y Bureau of Educational Research, University of Illinois. Grune &
herramientas de software libre en el campo de la robótica. En las Stratton, Inc.
memorias del congreso Tecnologías del Software Libre (La Coruña, [17] McKerrow P. Robotics research and education with LabVIEW. In
España, Julio 7-8, 2005) Proceedings of the Apple University Consortium Conference (New
[9] Martin F. 2001. Robotic Explorations. A hands-On Introduction to South Wales, Australia, Abril 25-28, 2000)
Engineering. New Jersey Prentice Hall. 2001. [18] Acuña A. 2006. Projects for Educational Robotics: Engines for the
[10] Blank D. 2006 Robots Make Computer Science Personal. In Journal of Innovation. In the proceedings of the Fourth International Conference on
Communications of the ACM. Vol. 49. Ed. ACM Press, New York, NY, Multimedia and Information and Communication Technologies in
25-27. Education. Current Developments in Technology-Assisted Education
[11] Fischertechnik. http://www.fischertechnik.de/en/index.aspx. Fecha de (Sevilla, España, Noviembre 22-25, 2006)
acceso: 09 de October de 2009 [19] T. Fong, I. Nourbakhsh, and K. Dautenhahn, “A survey of socially
[12] UNESCO Institute for Statistics, 2006 interactive robots,” Robotics and Autonomous Systems vol. 42, no. 3–4,
[13] Dias M., Browning B., Mills-Tettey A., Amanquah N. (2007) Robotics pp. 143–166, 2003
Education in Emerging Technology Regions. En AAAI Spring [20] Piperidis, S. Doitsidis, L. Anastasopoulos, C. Tsourveloudis,
Symposium: Robots and Robot Venues: Resources for AI Education. N.C. 2007. A Low Cost Modular Robot Vehicle Design for Research
(Menlo Park, USA, 26-28 Marzo 2007) and Education In proceeding of the 15th
[21] Wedeward K., Bruder S., Sarkodie-Gyan T. 2001. Exploration in
Robotics: Outreach to Secondary Education. In Proceedings of the AAAI
2001 Spring Symposium Series. (California, USA, Marzo 26–28, 2001)
Selecting Patterns for the Evaluation of Systems
Dmitry Zub, Erki Eessaar
Tallinn University of Technology, Estonia
[email protected], [email protected]
77
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_14, © Springer Science+Business Media B.V. 2010
78 ZUB AND EESSAAR
situation) in which a problem can occur. Secondly, the prob- So if there are more such possible alternatives (systems),
lem itself is described. And finally, the solution to the prob- then it is necessary to further limit the set of systems.
lem is described together with discussion of how to implement One should select the systems that have high popularity or
this solution. Now, we describe our method and how it was quality (in terms of the goal of the AHP model). For the
used for the evaluation of usability of e-learning systems. selection of alternatives one could use the results of sur-
veys, opinions of experts, and personal experiences. If
C. A Pattern-based Evaluation Method of Systems. one selects systems that have low quality according to ex-
We use Analytic Hierarchy Process (AHP) [13] for the perts, then the good result of evaluation means only that s
evaluation of systems. The general idea of our proposed me- has a good quality compared to the other selected sys-
thod is to compare a system s, which we want to evaluate in tems. Our approach is not well usable if one wants to
evaluate a groundbreakingly new system that has few or
terms of an overall goal g, with similar systems by using a set
no similar systems.
of patterns that correspond to g.
3. Selection of patterns (criteria) for the evaluation. We
We have to create a hierarchical model to solve a complex
discuss the selection of criteria in the next section.
evaluation problem, according to AHP. On the highest level of 4. Analysis of the patterns, which are selected for the evalu-
this model is the goal, which is the overall objective that we ation, by comparing them pairwise in relation to the goal
want to achieve. For instance, the goal may be to find relative g of the evaluation. The result of the step is an AHP
usability of different e-learning systems. model that shows us the relative importance of each pat-
At the next level of the model are objectives. They are the tern in terms of g.
main aspects that we should take into account while working 5. Writing of tests that are needed to evaluate the systems in
towards achieving our goal. These objectives may have some terms of the selected patterns.
sub-objectives that are located at the next level of the model. 6. Performing the tests based on the selected patterns.
We propose to use patterns as criteria to evaluate how well 7. Analysis of the results of the tests. Pairwise comparison
systems satisfy objectives. of the systems in terms of each pattern based on the re-
The selected patterns must correspond to the overall goal of sults of tests.
the model. In case of evaluating an e-learning system, we 8. Calculation of final weights for each system based on the
might use web-usability and e-learning patterns as criteria. relative importance of patterns. The resulting model
And finally, at the lowest level of the model, there are al- shows us how well each selected system (including s) cor-
ternatives. They are objects, between which the choice will be responds to the goal of evaluation.
made. We propose to use systems as alternatives. One of the 9. Identification and solving of the main problems that were
systems must be a system that we want to evaluate. discovered during the testing. It means partial redesign
So in general, we are evaluating which of the alternatives and reimplementation of system s.
corresponds better to our objectives. AHP method uses pair- 10. Retesting of s to find out whether it has improved.
wise comparisons where elements of lower level are compared 11. Analysis of the patterns used in the evaluation.
with each other in relation to some element of the model that III. SELECTION OF PATTERNS
is located at the next higher level. Therefore, objectives are
In our previous study [1], the selection of patterns was
compared with each other in terms of the goal. Possible sub-
based on the experience of an evaluator. In this paper, we
objectives are compared pairwise in terms of the objective.
explicitly explain the method for selecting patterns. Firstly,
Finally, alternatives are compared with each other against ob-
we introduce the method based on an example.
jectives (or sub-objectives). For example, in [1], we used an
Initially, we have to select collection of patterns that we
in-house developed e-learning system and in addition three
will use for the analysis. We select Web Usability (WU) pat-
other e-learning systems as alternatives.
tern language [15] because all the systems under discussion in
The steps of our method are the following.
the study [1] are web-based systems and we used patterns
1. Selection of the overall goal g of the evaluation. Based
from this language in [1]. Therefore, we can compare the se-
on that we have to select criteria (patterns) and alterna-
lection based on the experience of an evaluator and the selec-
tives (systems).
2. Selection of systems (alternatives) for the evaluation. tion that is proposed in this paper.
One of the alternatives must be the system s that we want One should take into account the following aspects during
to evaluate. Other alternatives must be systems that are the selection of patterns.
similar to s. The first criterion for selecting a system, • The number of patterns that will be used for the evalua-
with which we want to compare s, is its availability. The tion. If we use AHP method, then probably nine is the
evaluators must have access to the selected systems so maximum number of patterns one could use (see sec-
that they can perform tests based on these systems (see tion II) as criteria. We can use more patterns if the crite-
steps 5–6). Forman and Selly [14] note that psychologi- ria have sub-criteria but it complicates the evaluation. On
cal as well as mathematical evidences suggest “when de- the other hand, WU language contains 79 patterns [15].
riving priorities from paired comparisons, n, the number • Functionality of the system. If a pattern is about a func-
of factors being considered must be less or equal to nine”. tion that is not necessary or important for the system,
SELECTING PATTERNS FOR THE EVALUATION OF SYSTEMS 79
which we want to evaluate, then we do not have to con- opinion some patterns could belong to more than one group.
sider the pattern. The next step in the process of selecting patterns is to ex-
• A pattern p could be unsuitable for the evaluation because clude patterns that one cannot use for the evaluation due to
it is not important in terms of the overall goal of the anal- their unsuitability for evaluating already existing systems.
ysis or the evaluators do not have enough information
about systems to evaluate them in terms of p. C. Excluding Patterns from the Evaluation
• Some patterns could be related to each other. For in- As we mentioned in the previous section, WU pattern lan-
stance, they could be similar to each other. In this case, guage has some patterns that deal with the initial planning of
we should decide what patterns to choose from the set of systems (Group “Getting started”). Most of the patterns in
related patterns. this group are not suitable for our evaluation. In other groups,
The first step of pattern selection is the classification of pat- there are also some patterns that deal with rather technical
terns. Firstly, we should examine what information the se- issues that are hard (or impossible) to check in the evaluation
lected pattern language already contains. Hence, we now de- process. For example, pattern “Cache transactions”.
scribe the format and grouping of patterns in WU language. The only pattern, which could be useful for us from the set
“Getting started”, is “Two years old browser”. The reason is
A. Groups of Patterns that it describes not only how to test but also what to test.
The patterns of WU language are described in a classical (Check whether the system works in computers with old
format close to the one that is used in [9] by Alexander. hardware and with different web-browsers.) Hence, we pro-
Along with traditional pattern information (like name, context pose to use this pattern in the evaluation.
etc) we could use information about related patterns and star
ranking of patterns. D. Finding the Best Candidates from the Groups
Patterns of WU language are also grouped in categories ac- The goal of this step is to determine for each remaining
cording to a possible order of their use. The groups range group one or more patterns that are the most important in the
from “Getting started with the site” to “Dealing with context of the evaluation.
workflow and security”. Patterns About the Effectiveness of use of the System
Unfortunately, the existing grouping of patterns of WU lan- In this group there are several patterns about the speed of
guage is not very helpful. The reason is that patterns that could using the system like “Download times” and “Content before
be of interest to us are located in different groups. Therefore, graphics”. We do not include these patterns in the evaluation.
in the next section we divide the patterns into other groups We assume that nowadays there are not many problems with
based on the problem, the solution of which they specify. internet speed and there is no need to download large files in
e-learning systems. As you can see, selection of patterns re-
B. A Classification of Web Usability Patterns quires assumptions about systems that we evaluate.
After examining the patterns and problems that they de- Search for information is very important task for all sys-
scribe, we decided to group them based on the types of prob- tems, not only educational ones. Hence, the majority of the
lems that the patterns help us to solve. This way it would be patterns that describe how to organize navigation are impor-
easier for us to select patterns for the evaluation. The groups tant. Therefore, we discuss them next in more detail.
(and the problem areas) are the following. (See table one for Navigational Patterns
number of patterns in each group) Firstly, we shortly describe patterns (from [15]) that in our
• Getting started – development process of systems (from opinion cover navigational aspects.
setting of business objectives to testing). “Site map” – Provide a single page that gives visual and
• Effectiveness of use – problems with navigation, search, textual representation or overview of the entire site.
speed of using the system. “Search box” – Allow visitors to search from a site.
• Functional patterns – patterns that give recommendations “Sense of location” – In case of each page, provide visitors
about the functionality of systems and organization of clear clues about where they are on the page.
functions within a system. “Canonical location” – Make sure that the page name is
• Help patterns – the application of these patterns makes it displayed prominently
easier for users to learn to use the system. “Breadcrumbs” – Do not rely solely on breadcrumbs for
• Content patterns describe what information should be navigation unless you are short of space.
published at the site and how the text should be written. “Site logo at top left” – Clicking on the logo is a way to go
• Design patterns in our case describe how to implement to the homepage.
functions of the system, and how to organize the build- “Navigation bar” – Provide a bar on the home page and
ing-blocks of systems. possibly on other pages that allow users to go to the main sec-
• Technical patterns concern such things like the use of tions of the site.
HTML, security, and others. “Go back to safe place” – Allow users to mark the place
These groups were quite roughly defined. Therefore, in our they want to return if they are lost.
80 ZUB AND EESSAAR
“Back button” – Do not disable back button of browsers. enables a user to get necessary information about the task
Design transactions the way that they can be undone. without browsing help documentation.
If we look at the patterns described above, then we see that Design Patterns
there are some patterns that could be grouped together. For There are a lot of design patterns according to our classifi-
example, “Search box” can be an addition to “Site-map”. cation. Large part of them could be useful in our opinion be-
“Sense of location” is more general variant of “Canonical lo- cause they describe how to improve different aspects of usa-
cation”. “Canonical location” provides a recommendation to bility. Let us describe those patterns in more detail.
use cross-links across site hierarchy. One can use the solu- “Follow standards” and “Keep it simple” are rather abstract.
tions from “Breadcrumbs”, “Navigation bar”, and “Site logo” However, if these two patterns are followed, then it would be
patterns to create these links. “Back button” (or “Site logo”) easier to learn and to use a site.
can be used as a variant of “Go back to safe place”. There- “What you see is what you can use” – If a user sees on the
fore, we think that from these patterns the most important are site only things that he/she can use in the learning process and
“Site-map”, “Sense of location”, and “Go back to safe place”. he/she is not distracted by something else, then it is also good
Content (and Text) Patterns for learning.
This group of patterns explains what information should be “Mandatory fields” – Avoid the use of mandatory fields,
on a web-site and also how it should be represented. In our and while using them provide information about how to fill
opinion, most of content patterns (like “Links to many sites” them in forms.
and “Human touch”) are not very important for e-learning “Oblique Landmarks” – Indicate somehow whether some
systems. From this group we add to the evaluation “Natural features of a system are usable indirectly or when a user is in a
metaphor”. If this pattern is used, then probably it is easier to certain state.
learn to use the system and to understand it. In addition, “Button gravity” – Users usually expect “Submit” buttons
“Home page” pattern could be used in a sense that it would be at the bottom of the page. Therefore, you should locate them
nice if links to all main features of the system would be avail- there (or make them more visible).
able at home page. “Pipeline interaction” – Give users chance to revisit the ear-
There is a similar situation with the text patterns. Most of lier steps of the workflow and correct provided data if needed.
them provide recommendations about how text should be Selecting the Final Subset of Patterns
represented (“Design page for scanning”, “Short text”, and Based on the previous discussion, we have the following
“Separate print pages”). That is surely important for learning patterns left as candidates.
materials, but not for the system itself. From that category of
patterns, we include “Acceptable wording” for the evaluation. 1. Two year old browser 11. Sense of progress
If the system uses well-known terms, then it is easier for users 2. Site-map 12. Context sensitive help
to understand the new system and subsequently learn to use it. 3. Sense of location 13. Keep it simple
Technical Patterns (HTML, Security) 4. Go back to safe place 14. Follow standards
We found no suitable patterns in this group as all of them 5. Natural metaphors 15. What you see is what you
rather technical. 6. Home page can use
Functional Patterns 7. Acceptable wording 16. Mandatory fields
The most important functional pattern in the context of usa- 8. No unpleasant sur- 17. Oblique landmarks
bility evaluation is probably “No unpleasant surprises” – sys- prises 18. Button gravity
tem behavior should be predictable and reliable. Also “Ex- 9. Exploit Closure 19. Pipeline interaction
10. Feedback
ploit closure” can be suitable – this pattern means that before
completing some main tasks it should be guaranteed that a
We found too many patterns (19) that can be used in the
user has completed all the necessary preliminary tasks.
evaluation. How can we reduce the amount of patterns to nine
Help Patterns
[14]? Firstly, the list contains some patterns that specify a
Help patterns are quite important from the point of view of
possible solution (or part of solution) to some other more gen-
usability and they improve the ease of learning and the effec-
eral (abstract) pattern. For example, if we use pattern “Button
tiveness of use of a system. We have selected three patterns
gravity”, then it means that we also use pattern “Follow stan-
from that area. If a system gives a user a “feedback", then
dards”. If users expect that “Submit” buttons are located at
he/she understands better what is going on while working with
the bottom of the page, then it means they are used to that
the system. “Sense of progress” helps user to understand bet-
because of previous experience and it became a standard of
ter the task he/she is completing. “Context sensitive help”
TABLE I
PATTERN GROUPS AND NUMBER OF PATTERNS
Groups of patterns GS Eff Funct Help Cont Des Tech In all
Number of patterns 11 21 8 5 10 20 13 88
Of them selected 1 3 2 3 3 7 0 19
SELECTING PATTERNS FOR THE EVALUATION OF SYSTEMS 81
some kind. So, we should remove the more specific pattern Welie [16] and Tidwell [17].
from the list of selected patterns and only use the more general There exist some examples of the use of patterns for the
pattern. According to this principle, we propose to eliminate evaluation of systems. Kramberg [18] examines how much
pattern 11 (as it is a sub-pattern of 3), patterns 5 and 7 (and IBM WebSphere software supports the use of control-flow
keep pattern 13), keep pattern 17 and eliminate pattern 15. In patterns from [19]. Kramberg [18] does not select a subset of
addition, we keep pattern 14 and eliminate pattern 18. Pattern patterns for the evaluation. There exists a proposal [8] to use
19 is a sub-pattern of 11, which in its turn sub-pattern of 3. patterns in the usability inspection. Unfortunately the method
This way, we have eliminated six patterns (5, 7, 11, 15, 18, of selecting patterns is not clearly defined. Moynihan et al.
19) and decided to keep four patterns (3, 13, 14, 17). [20] describe a prototype of an expert system for the selection
The second way to decrease the number of patterns is to of software design patterns. But the goal of this system is
look through all of them once again and to decide if some of “leading a designer to a suitable design pattern to the problem
the patterns are really important for the systems, which we at hand” [20] and so it cannot be used to evaluate a system.
want to evaluate. For example, while we think that “Oblique During our selection process, we tried to classify the pat-
Landmarks” pattern increases the usability of systems, it is terns. The idea of classifying patterns is not new. One of the
possible that in e-learning system there are no functions that reasons for classifying patterns is the big amount of pattern
are available only to some special types of users. “Home collections in the literature. Henninger and Correa [21] state
page” pattern states that we should draw the attention of po- for instance that there exist 14 collections of user interface
tential users to a web-site with a home page. That is not im- patterns containing 425 patterns in total. Hence, some papers
portant for learning. The requirement that all the important discuss how patterns and pattern languages can be organized.
functions should be visible at home page can in some way Van Welie and van der Veer [22] discuss how pattern lan-
satisfied with “Site-map” pattern. “Exploit closure” pattern guages in interaction design can be structured. There are at-
concentrates more on how system should be designed so that tempts, like [23], to create repositories of patterns. One of the
all the tasks would be completed in the right order. “Sense of problems of storing many patterns in one repository is that
location” should provide users necessary information so that it different authors may use somewhat different formats for de-
would be clear in what order tasks should be completed. Fi- scribing patterns. In order to solve this problem in Human
nally, “Context sensitive help” is quite important because if a Computer Interaction, a unified pattern format was developed
user needs to understand how a function works, then he/she named Pattern Language Markup Language (PLML) [24].
can get the necessary information without searching that in V. A GENERALIZED PROCEDURE
help files. A recommendation to give information about fill-
ing “Mandatory fields” in forms is similar to the “Context We have developed a method for evaluating systems by us-
sensitive help”, so we do not include “Mandatory fields” in ing patterns as the criteria. In [1, 2], we demonstrated how to
the final subset of patterns. use the method to evaluate usability of e-learning systems
At the end, we have selected patterns 1–4, 8, 10, 12–14 for based on usability patterns. In [1], we selected patterns for the
the evaluation. In the next section, we compare this selection evaluation based on the personal experience and knowledge of
with the one that was made in our earlier research [1]. an evaluator without trying to explicitly explain the process of
selection. In this study, we repeated the selection process and
E. Comparison of the two Selections. divided it into steps so that it would be possible to reason
In our previous research [1] two types of patterns were used about it. The process itself consists of four steps.
for the evaluation: web-usability patterns and e-learning pat- Step 1: Selection of one or more pattern languages (sets of
terns. Therefore, five patterns of each type were selected. For patterns) or individual patterns. The selected patterns form the
the web-usability, the following patterns were selected. “Two initial set S1 of patterns, from which we select the final set of
year old browser”, “Site-map”, “Sense of location”, “Feed- patterns that will be used for the evaluation.
back”, and “Follow standards”. All these patterns were se- Step 2: One has to look through all the patterns in S1 and
lected in the current research as well. It means that the choice divide them into groups. All the patterns in one group must
of patterns in [1] was good and repeatable. But it was based help us to solve similar problems.
solely on the experience of one expert and the process of mak- Step 3: Some patterns might be unsuitable for the evalua-
ing this choice was not explicitly explained. tion process (at least for the evaluation of existing systems).
One has to eliminate these patterns from S1.
IV. RELATED WORKS
Step 4: From each group (step 2) one has to select a subset
As it was already mentioned in Section II, a pattern de- of patterns. The selected patterns are the main criteria and
scribes “a solution to a problem” so that the solution could be form set S2. The cardinality of S2 must not be bigger than
used over and over again in different contexts. There exist nine. Selection must be based on the suitability of the patterns
collections of patterns (pattern languages) that describe prob- to the systems in the analysis. For example, while “Interna-
lems and solutions of a specific domain. Examples from the tionalization” is important for the sites of organizations that
field of usability are Wu Language [15] and collections of van operate in several countries, it is not so important for an
82 ZUB AND EESSAAR
e-learning system where learning is going on in one language. AHP, should not be bigger than nine. For instance, WU lan-
In our opinion a pattern p, which defines a general rule, is guage of usability patterns contains 79 patterns. Therefore, it
very suitable for the evaluation if there are one or more must be possible to select a set of patterns for the evaluation
“sub-patterns” that can describe specific (variants) of solution from a bigger set of patterns.
for a problem defined in p. We can use this principle also for In this paper, we proposed a systematic and structured
selecting a subset of patterns from each group (step 4). If a process how to select a set of patterns that one could use to
selected pattern p has a set of specific sub-patterns, then it is perform evaluations of systems. We illustrated the process by
also possible to use these patterns as sub-criteria of the main selecting a set of usability patterns that one can use to evaluate
criterion (p) in the AHP analysis. The amount of sub-criteria e-learning systems.
of a criterion should also be no more than nine, because we
REFERENCES
have to compare the sub-criteria pairwise.
The proposed process still needs decisions of experts who [1] D. Zub and E. Eessaar, “Pattern-Based Usability Evaluation of E-
Learning Systems,” International Conference on Engineering Education,
perform evaluation. However, the procedure should facilitate Instructional Technology, Assessment, and E-learning (EIAE 07), pp.
the work of evaluators. If decisions, which are made during 117–122, Springer, 2008.
the selection, are explicitly documented, then it also allows [2] D. Zub and E. Eessaar, “Pattern-Based Usability Analysis: Comparison
with a Case-Study,” International Conference on Engineering Education,
interested parties to better understand the evaluation results. Instructional Technology, Assessment, and E-learning (EIAE 08),
To get the final set of patterns for the evaluation, we can re- Springer, 2009, in press.
peat step 4 once more to compare patterns from different [3] J. Nielsen, “The Usability Engineering Life Cycle,” Computer, vol. 25,
pp. 12–22, 1992.
groups. The pattern-based evaluation method is not well usa- [4] E. Folmer and J. Bosch, “Architecting for Usability: A Survey,” The
ble if one wants to evaluate an aspect a of systems, and a has Journal of Systems and Software, vol. 70, pp. 61–78, 2004.
few corresponding patterns that specify the best practices. [5] J. Kontio, L. Lehtola, and J. Bragge, “Using the Focus Group Method in
Software Engineering: Obtaining Practitioner and User Experiences,”
To facilitate the process of pattern selection, the following International Symposium on Empirical Software Engineering,
additional information about each pattern should be recorded pp. 271–280, 2004.
in the repository of patterns: (a) a classification of patterns [6] J. Nielsen, Usability Engineering. Academic Press, 1993.
[7] J. Nielsen, “Heuristic Evaluation,” In: Nielsen J., Mack R. L. (eds.)
based on the problems the patterns deal with; (b) is this pattern Usability Inspection Methods. New York: John Wiley & Sons, 1994, pp.
suitable for the evaluation process?; (c) the history of using 25–62.
the pattern for the evaluation of systems; (d) does this pattern [8] M. Schmettow and S. Niebuhr, “A Pattern-based Usability Inspection
Method: First Empirical Performance Measures and Future Issues,” HCI
provide a general advice or some practical solution?; (e) what 2007, vol. 2 of People and Computers, pp. 99–102, 2007.
is the rating of the pattern?; (f) how is this pattern connected [9] C. Alexander, The Timeless Way of Building. New York: Oxford Uni-
to other patterns in the collection? versity Press, 1979.
[10] E. Gamma, R. Helm, R. Johnson, and J.O. Vlissides, Design Patterns –
VI. FUTURE WORK Elements of Reusable Object-Oriented Software, Addison-Wesley, 1994.
[11] D.C. Hay, Data Model Patterns: Conventions of Thought. New York:
The future work in this area includes the creation of a sys- Dorset House Publishing Company, 1996.
[12] J. Bergin “Fourteen Pedagogical Patterns,” Fifth European Conference
tem, which supports the evaluation process. Among other on Pattern Languages of Programs, 2000.
things it should support the management of patterns and all the [13] T.L. Saaty, “How to Make a Decision: The Analytic Hierarchy Process,”
necessary information for the selection (see Section V) of pat- European Journal of Operational Research, vol. 24, pp. 19–43, 1990.
[14] E. H. Forman and M.A. Selly, Decision by Objectives. World Scientific
terns. Another direction of research is to study the following Publishing Company, 2002.
aspects of pattern-based testing process. [15] Web Usability Patterns. Retrieved September 25, 2009, from
• What kind of tests to use in the evaluation? http://www.trireme.com/WU/home.htm
[16] M. van Welie, Interaction Design Patterns Library. Retrieved September
• How many patterns to evaluate in one test? Should a test 25, 2009, from http://www.welie.com/patterns/
cover one pattern or a set of them? This information [17] J. Tidwell, Designing Interfaces. Retrieved September 25, 2009, from
could be later stored in a repository so that the future use http://designinginterfaces.com/
of this test would be easier. [18] V. Kramberg, “Pattern-Based Evaluation of IBM WebSphere,” BPEL
Master’s Thesis, Universität Stuttgart, 2006.
• Who will be the testers of the system and how many [19] V.M.P. van der Aalst, A.H.M ter Hofstede, B. Kiepuszewski and A.P.
people should participate? Barros, “Workflow patterns,” Distributed and Parallel Databases,
• How to determine the relative importance of the problems vol. 14, pp 5–51, July 2003.
[20] G.P. Moynihan, A. Suki, D.J. Fonseca, “An Expert System for the Se-
discovered during the testing, and how to classify them? lection of Software Design Patterns,” Expert Systems, vol. 23,
pp. 39–52, 2006.
VII. CONCLUSIONS [21] S. Henninger and V. Correa, “Software Pattern Communities: Current
We have proposed an evaluation method of systems that is Practices and Challenges,” 14th Conference on Pattern Languages of
Programs, 2007.
based on Analytical Hierarchy Process (AHP) and patterns. [22] M. van Welie and G. C. van der Veer, “Pattern Languages in Interaction
Patterns are criteria, based on which is possible to evaluate Design: Structure and Organization,” Interact’03, pp. 527–534, 2003
how well systems correspond to different objectives. Psycho- [23] A. Gaffar, D. Sinnig, H. Javahery, and A. Seffah, “MOUDIL: A Com-
prehensive Framework for Disseminating and Sharing HCI Patterns,”
logical and mathematical evidences suggest that the amount of CHI 2003 workshop on HCI Patterns: Concepts and Tools, 2003.
criteria (that correspond to the objectives), that one can use in [24] S. Fincher, “Perspectives on HCI Patterns: concepts and tools (introduc-
ing PLML),” Interfaces, vol. 56, pp. 26–28, 2003.
Effectiveness of Collaborative Learning in Teaching
Information Systems
Ioana Tuugalei Chan Mow
National University of Samoa
Abstract: This paper is based on the second phase of an collaborative learning could be proven effective in teaching
investigation which explored student attitudes and experiences in computing subjects, this would have a direct beneficial
collaborative and individualised learning in computer based bearing on resource implications.
instruction. It attempted to answer the questions a) whether The current research was conducted in two phases. The
Samoan students prefer learning in a cooperative learning
environment as opposed to individualised learning and secondly
research was structured such that the first phase initiated a
b) whether collaborative learning is more effective than general comparative investigation of the attitudes and
individualised learning? Research results indicated that students performance of students in collaborative learning as compared
enjoy and show positive attitudes towards collaborative learning to individualised learning. This initial phase spanned three
and that they perform well in a collaborative environment weeks. The second phase provided a semester long and more
provided the essential conditions for effective collaborative in-depth study of the essential conditions necessary for
learning are satisfied or present. optimal collaborative learning.
II METHODOLOGY
I INTRODUCTION
The research design used in the study was quasi-experimental,
This paper describes the second phase of an investigation
where the independent variable (learning environment) was
which explored student attitudes and experiences in
manipulated and subsequent changes in the dependent
collaborative and individualised learning in computer based
variables (student attitudes to learning environment and
instruction. It attempted to answer the questions
student performance) were investigated. The quasi-
• Do Samoan students prefer learning in a cooperative
experimental design was chosen for both phases in the study,
learning environment as opposed to individualised
as its strict control of variables enabled a distinct focus on
learning?
causal relationships between learning environment and student
• Is collaborative learning more effective than attitudes to learning.
individualised learning?
The Computing Department at the National University of The population for the second phase was a ‘captive’ one
Samoa offers courses in Computer Science and Information consisting of students in a Computer Information systems
systems. One of the challenges in teaching Information class. The students in this phase had already completed an
Systems is the provision of students with skills in developing Introduction to Information Systems class in Information
real life systems in the classroom. In the workplace, these systems analysis and design. Phase 2 involved a much smaller
systems are developed using the team approach.Therefore, the group of students (9 students) and group selection was on
use of collaborative learning in the learning environment for ability so that each group of three was of mixed ability (high,
Information systems coordinates well with the team approach low and average). The composition of the group is an
to Information systems design in real world applications. important factor in CL and research in CL recommend
Another motive for undertaking this study was the interest in heterogeneous groups as some difference in view points is
the cultural effects of collaborative learning in the field of required to initiate interactions between group members [2].
computer education in Samoa. Samoan culture is communal Ability was based on previous grades in CSB 283, the
and based on communal practices[1]. Family members prerequisite class to this Information Systems class.
cooperate and share in all manner of responsibilities ranging
from domestic tasks to maintaining family lands and providing III PROCEDURES
for the welfare of the family.
A major drawback in Educational computing in Samoa has Phase II of the research was conducted on students in an
been the high cost of computer technology, manifesting itself Information systems and database design class at the local
in the shortage of computers and the high student to computer university. The class was divided into heterogeneous groups of
ratio. Enrolment numbers for computer courses have always threes according to ability (high, average, and low ability).
been restricted by the availability of computers. If Selection and composition of the groups were decided by the
lecturer, who was also the principal researcher.
83
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_15, © Springer Science+Business Media B.V. 2010
84 MOW
To provide the context rich environment of the real world, the Evaluation of group projects involved one group presentation
groups were assigned two real life projects. The first project for the Library system and written documentation such as data
involved the development of a computerised system to flow diagrams, entity relationship diagrams, data dictionary,
automate the checkin/checkout transactions of the university user manuals and a disk copy of the final database system.
library. The second project involved the development of a Assessment was done on a group basis with one mark given to
computerised invoice entry system for a local water company. the whole group. Assessment from the two projects accounted
The major activities in the two projects involved for 80% of the overall assessment. The remaining 20% came
• the groups carrying out investigative interviews on the from a 1-hour test which tested their knowledge and
clients to elicit such things as user requirements, problems comprehension of database theory.
in existing systems leading to a need for an improved
system and also the feasibility of the new system. B Student attitudes
• Systems analysis using a CASE software tool called
ASCENT Student preferences for a particular learning environment and
• Systems design using ASCENT and Word for Windows their feelings towards such an environment were measured by
for prototyping design of menus and user interfaces self-administered questionnaires, structured interviews and
• Implementation of the system using Microsoft Access. observations by the principal researcher. The questionnaire
• Testing of the database system and making the necessary focused on issues of, the enjoyment of learning, self-esteem,
revisions and improvements. motivation, learner control, and social interactions. For the
Since the two systems were planned and designed using purpose of the current study, the enjoyment of learning was
Systems Development Life cycle approach, which involved interpreted and defined as students’ valuation and appreciation
several phases, assessment of the projects was continuous. For of the learning experience.
example, the deliverables for the early planning and
investigative phases were assessed and feedback given on V DATA GATHERING TECHNIQUES
areas needing improvement before proceeding to the systems Student attitudes to the learning environment were measured
analysis and design phases and eventually to the by self- administered questionnaires and also by observations
implementation. by the researcher. In addition to the self-administered
Throughout the two projects, which spanned the duration of a questionnaires, groups in Phase II were also interviewed on
semester, students were expected to carry out planning, their views and experiences with collaborative learning.
analysis and problem solving by means of extensive group
discussions.
VI RESULTS
IV OPERATIONALISATION OF VARIABLES The aim of the current research was to establish from the
results of the study, the following propositions. Firstly, the
In this study, the independent variable was the learning study attempted to confirm that students in Samoa enjoyed
environment which had two values in its domain, collaborative collaborative learning and preferred it to individualised
learning and individualised learning. The dependent variables, learning. Secondly, the study attempted to establish that
were student performance and student attitudes to the learning students in Samoa performed better in a collaborative
environment. In the course of the study, the independent environment.
variable (learning environment) was manipulated and Hence, for purposes of the discussion, the main areas of focus
subsequent changes in the dependent variables (student were student performance and student attitudes to learning.
attitudes to learning environment and student performance)
were gauged using questionnaires, surveys and pre and post- A Student attitudes to Learning
tests. Student attitudes to collaborative learning were gauged from
responses to the questionnaire administered at the end of the
A Student performance session and also from interviews with the researcher. Results
from the various questionnaires and interviews, showed that
In Phase II, the effectiveness of the learning environment is the majority of the students viewed collaborative learning as
student performance measured or assessed as a group using an effective instructional strategy with positive effects on
criteria such as group presentations, grades for deliverables problem solving, self esteem, motivation and enjoyment.
such as the Initial Data Model, Initial Process Model, However, despite these outcomes, the results did not show that
Feasibility study, database tables, forms, menus and reports. students preferred collaborative learning over individualised
The marks for the collaborative learning sessions were learning as the analysis of Phase 1 results did not reveal any
compared to marks from the previous Information systems significant differences in attitudes between the two treatments.
class and the difference between the two sets of grades, Discussion with the students in the interviews revealed the
analysed statistically, for significance. following:
EFFECTIVENESS OF COLLABORATIVE LEARNING IN TEACHING INFORMATION SYSTEMS 85
Question 1 Did you enjoy working together as a team? as the group members discussed the various issues in the
Of the three groups in this second phase, the members of two project.
of the groups, enjoyed working together as a team. Two of the
members of Team 2 in the early stages of the exercise, initially Question 7 Did working in a group help stimulate thinking?
resented working in a collaborative mode, due to the inability As revealed in the interviews, all of the participants, agreed
to resolve conflicts between group members. However, their that group learning stimulated their thinking, in that the group
attitudes changed as the exercise progressed. discussions gave them new perspectives on the subject area
and brought up other valuable issues related to the topic under
Question 2 What problems, if any, did you have working discussion.
together as a group?
Members in Group 2 indicated that they had had problems in Question 8 Did working in a group give you new perspectives
scheduling times in which everyone would be available for on the subject area?
group discussions. There were also feelings of resentment All of the students agreed on the value of group learning on
where some members felt that members in the group were not broadening one’s perspectives.
contributing their share of the work. Most of the members in
the other two groups, felt that they had progressed well and
had not had problems working as a team. VII DISCUSSION
The aim of the current research was to establish from the
Question 3 Did everyone contribute equally or did some results of the study, the following propositions. Firstly, the
dominate the group? study attempted to confirm that students in Samoa enjoyed
The general feeling in Group1 and Group3 was that everyone collaborative learning and preferred it to individualised
contributed equally to the group. In Group 1, Student 1A and learning. Secondly, the study attempted to establish that
Student 1B were Accounting experts while Student 1C students in Samoa performed better in a collaborative
provided the expertise in computing applications, so there was environment. Hence, for purposes of the discussion, the main
some form of division of labour, as the members areas of focus were student attitudes to learning and student
complemented each other’s skills and shared their expert performance. Student attitudes to Collaborative learning were
knowledge. both positive and negative.
Question 4 Did working together help the student’s i) Positive attitudes to Collaborative learning
understanding of the project? The groups in which everyone participated and contributed to
All of the students agreed that collaborative learning improved team progress enjoyed the experience and found that
their understanding of the project. There were aspects of the collaborative learning led to increased social interactions
project one would not have thought of, had they not been told between the members of the group. As evident in all of the
of, by other members of the group. Discussion of some of the student interviews, students felt that solving problems and
issues in the project, led to awareness of further issues in the carrying out tasks in groups was a lot more effective than
project that they, as a group, had to resolve and address. doing it individually. From student responses in student
However, Student 2D and Student 2E in Group2, felt that interviews, it was clear that practically all of the students in
working as a group actually led to confusion of issues in their the collaborative group felt that it was an effective
group. When asked to elaborate, they explained that both instructional strategy for the following reasons :
members were convinced that their own views of the way in
which some aspect of the project should be approached were • Collaborative learning taught students teamwork
the correct one and refused to accept any other perspective. Students felt that collaborative learning taught them to work
together as a team and share both the work-load and ideas.
Question 5 Do you feel that collaborative training helps in This was consistent with the findings of researchers such as
pooling your knowledge and experience? [3] and [4] that demonstrated the positive effects of
When asked about the value of collaborative learning in collaborative learning in fostering team-work.
pooling their knowledge and experience, all of the members in
Group 1 and Group3 felt that they would not have been able to • Students enjoyed Collaborative learning
develop and implement the systems without the help from Practically all of the participants, claimed, they had enjoyed
fellow members and the lecturer. working in the collaborative mode. This was especially true in
Student 2D and 2E initially had doubts about the worth of groups where members had successfully worked together as a
group learning, but as the research progressed, they changed group and enjoyment was probably due to feelings of success.
their opinions in favour of collaborative learning. It could also be argued that enjoyment led to an increase in
intrinsic motivation. These findings were consistent with the
Question 6 Did you get helpful feedback in your group? literature and it could be argued that these feelings of
All of the students participating in the collaborative mode, enjoyment were due to increased motivation and increased
indicated that they received helpful feedback from their group social interactions. Collaborative learning, as [5] pointed out
86 MOW
was a social event where learning was fostered through peer ii) Negative attitudes to Collaborative learning
interactions and social discourse. It should be acknowledged, that student attitudes to
collaborative learning in the current study, were not all
• Collaborative learning led to increased self confidence positive. The research results also included issues in
The positive effects of collaborative learning on self collaborative learning, which the participants found
confidence and esteem have been established by findings of troublesome and aspects of the learning environment, which
studies by [7] and [9]. This was confirmed by responses of the they found disagreeable. However, in viewing all of the
majority of the participants in the collaborative treatment, who student feedback, these negative aspects were quite minimal
maintained, that collaborative learning had boosted their self- (less than 10 %) when compared to the overwhelming volume
confidence. Again this was evident mostly in groups where the of positive feedback.
collaborative mode had been successful leading to feelings of
increased self-confidence. Furthermore, the mere fact that • Lack of cooperation
most of the decisions were the product of group discussions Perhaps the most common complaint raised by the participants
and consensus, meant that students felt a lot more confident in was the lack of cooperation from some team members who
any decision making and tasks done by the group as it had the were not doing their “fair share” of the work. The lack of
approval and confirmation of other people. cooperation also resulted in the lack of group processing,
which, as recommended by key researchers ([3]; [11]) in the
• Collaborative learning promoted sharing of ideas field, was one of the essential factors for successful
Practically all of the students in the collaborative group, collaborative learning.
confirmed, that they felt that the benefits of collaborative
learning lay in the ability to share ideas through group • Slow pace of work
processing. Another complaint voiced by the students in the collaborative
mode was the slow pace of progress. It could be explained,
• Encouraged discussion of issues from various view points that apart from some students wasting time on “off task
enabling an in-depth probing of problem, task or concept. activities”, the longer time spent on tasks, was mainly due to
As pointed out by [5], knowledge is a social construct and the more time spent on discussions and increased interactions
learning process is facilitated by social interactions amongst amongst group members. However, to this end, the author
students as they collaborate with each other to actively would like to claim that this slow pace was actually beneficial,
construct their learning. In the collaborative learning groups, as the group members needed this time to carry out in-depth
the sharing of ideas and varying viewpoints through discussions, to critically reflect on the group’s progress and
discussions led to a deeper understanding and analysis of a plan of action and to resolve any conflicts and controversy
problem, task or concept as it exposed the student to concepts which may have arisen. These activities, if carried out
they had not thought about before. These supported the effectively, would have resulted in improved problem solving
findings of studies by [3], [4], [6], [7], [8] and many others, and promoted critical thinking within the group.
which showed that collaborative learning led to increased
cognitive analysis and enhanced problem solving in learning. • Resolving controversy
Another issue, brought to the fore by the author’s
• More able students provided ‘scaffolding’ for the weaker observations, was that of resolving conflict and controversy
ability ones amongst group members.
According to [12],
The concept, referred to by [10] as the “zone of proximal “Conflict exists whenever incompatible activities
development”, that the more able students provide mentoring occur...”
for the less able, was quite noticeable in the successful groups Classification of conflicts include “controversy, which occurs
in Phase II. In these groups, the more able students helped the when a person’s ideas or opinions are incompatible with
weaker students in understanding concepts. A direct result of others” ([12]), and there is a need to arrive at some consensus
this mentoring was the obvious improvement in self- or agreement. [12] have argued that controversy, when
confidence in the less able students and, more importantly the managed constructively, could lead to enhanced problem
significant improvement in achievement of the weaker ability solving, enhance the quality of reasoning and decision
students in Phase II. making and strengthen liking, respect and trust amongst
members.
In Phase II, the feelings of hostility between Student 2A and
• Increased motivation Student 2B, were due to their inability to resolve conflicts
The students also indicated that collaborative learning within the group. These students had opposing views on most
promoted increased feelings of motivation, and this was issues. However, neither was prepared to come to some
particularly true in cases where collaborative learning had agreement or work out a settlement. Controversy, when
been implemented successfully. These findings are in tune resolved constructively would prove beneficial to
with the findings of [3], and [8]. collaborative learning. However, controversy, when
EFFECTIVENESS OF COLLABORATIVE LEARNING IN TEACHING INFORMATION SYSTEMS 87
[12]Johnson, D. W., & Johnson, R. T (1995). Creative Controversy: [14]Springer, L., Stanne, M. E., and Donovan, S. (1998). "Effects of
Intellectual challenge in the classroom (3rd edition). Edina, MN: Interaction cooperative learning on undergraduates in science, mathematics, engineering,
Book Company. and technology: A meta-analysis." (Research Monograph No. 11). Madison:
University of Wisconsin-Madison, National Institute for Science Education,
[13] Kimber,.D. (1994) Collaborative Learning in Management education: Review of Educational Research.
Issues and Benefits, problems and solutions: A Literature Review.
Implementation of a Curriculum Management
System at a Faculty of Medicine – Lessons Learned
from a Pilot Project
Aarti Paul Daniel Yaworski Keith McConnell Bruce Martin Dean Sandham Sat Sharma
Faculty of Faculty of Faculty of Faculty of Faculty of Faculty of
Medicine Medicine Medicine Medicine Medicine Medicine
University of University of University of University of University of University of
Manitoba Manitoba Manitoba Manitoba Manitoba Manitoba
[email protected] daniel_yaworski@ [email protected] [email protected] Dean_Sandham@c [email protected].
ba.ca hotmail.com nitoba.ca toba.ca c.umanitoba.ca ca
Abstract— Integration of electronic curricula remains highly attitudes [3] [5]. Recent advances in information technology
variable among medical schools in North America. A major have laid a strong foundation for the use of electronic
challenge has been availability and implementation of an ideal technology to deliver, support and enhance teaching, learning,
curriculum management software system (CMS). The objective and evaluation [6]. Using well-designed computer-based
of this study was to test the feasibility of using a CMS system at
the University of Manitoba (U of M). At the U of M, we
courseware, electronic learning can facilitate live teaching
implemented a multi-user software system designed to provide with streaming lectures, whiteboards, downloadable slide sets,
faculty and students with an integrated curriculum, scheduling, and discussion forums [6] [9]. The benefits of e-learning
calendar functionality, and communication tools. The goals of the include faster learning at reduced cost, increased access to
pilot project were to explore the feasibility of an electronic learning, and accountability of all participants in the learning
curriculum and to evaluate strengths and weaknesses of the process [10] [11]. Changes in health care delivery, advances
software. One block each from the 1st and the 2nd year medical in medicine, increased demands on academic faculty, and
school class were chosen for the pilot project. Data were collected transformation of health care delivery sites from tertiary care
through student and faculty surveys, student and faculty login- institutions to community-based settings have challenged
ins, student page views, and through direct communication with
various users. Most students (100%) and faculty (78%) indicated
traditional teaching time and resources [12].
a strong need for an electronic CMS. A statistically significant
increase in student log-ins (p<0.01) and page views (p<0.05) II. BACKGROUND WORK
occurred from week 1 to week 6 and then reached a plateau, Despite gaining popularity over the past decade, the use of
whereas faculty log-ins remained constant. The 1st year and 2nd
e-learning remains highly variable among medical schools and
year medical classes were dissatisfied (34% and 57%), neutral
(24% and 25%), or satisfied (42% and 18%) with the CMS, its integration has been a genuine challenge, especially when
respectively. This pilot project demonstrated that an electronic teaching clinical disciplines [6] [12]. Creation and
CMS was feasible at the U of M. In addition, we learned how to deployment of learning management systems that support e-
effectively implement such a system, and how to encourage learning are arduous and expensive, and their integration into
faculty and students to adopt the system. The lessons learned at U medical education is equally challenging. E-learning material
of M may be instructive for other faculties attempting to deploy involves several components [13] [14]. Once the content is
electronic curriculum for undergraduate medical education. developed, it must be managed, delivered, and standardized.
Content management includes all the administrative functions
Keywords- eLearning; electronic curriculum; curriculum (e.g., storing, indexing, and cataloguing) needed to make e-
management system; learning management system. learning content available to the learners [15]. Examples
include portals, repositories, digital libraries, learning-
I. INTRODUCTION management systems, search engines, schedules, and
personalized e-portfolios. A curriculum management system
Medical education aspires to provide future doctors with the includes web-based software that facilitates the delivery and
knowledge and skills to practice medicine and to develop tracking of e-learning across an institution. It may also serve
lifelong learning skills [1], [2]. The traditional model of several additional functions [16] [18]. For example, CMS can
medical education is based on transferring scientific facts to simplify and automate administrative and supervisory tasks,
students via lectures, notes and books, culminating in track learner’s achievement of competencies, and provide a
knowledge synthesis, and acquisition of skills and professional
89
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_16, © Springer Science+Business Media B.V. 2010
90 PAUL ET AL.
readily accessible and life-long repository for learning uploaded their own material, while others e-mailed it to the
resources [19]. The American Association of Medical course director, course administrator, or to us for uploading.
Colleges has suggested various approaches to implement Personal communication with the faculty and course directors
informatics skills into the medical school curriculum [1]. appeared to be the most successful method of preparing the
In 2004, the Faculty of Medicine (FOM) at the U of M curriculum for the pilot project. An implementation committee
began exploring the feasibility of acquiring a multi-user, was created to oversee the pilot roll-out. Members of the
integrated electronic curriculum management system for its committee included directors from 1st and 2nd year medical
medical school. The goal of this curriculum system was to school classes, course directors, pre-clerkship administrators,
provide medical students and faculty the best available and student representatives from each class. Feedback and
technology for knowledge transfer. Such a system was counsel was obtained from the committee and other users prior
expected to offer a simple, user-friendly interface that could be to customization of system for the U of M pilot project.
utilized by individuals with varying levels of technical skill. In (Fig. 1)
this paper, we describe our experience and the challenges we We used multiple strategies to prepare medical school
encountered at U of M during selection, preparation, faculty, students, and administrative staff for a successful
customization, implementation and evaluation of a web based launch of the pilot project. These included personal
comprehensive electronic curriculum management system. communications, presentations, and discussion at various
undergraduate administrative/education committees. Critiques
and suggestions obtained from multi-user groups were
III. METHODS
collated, analyzed, and put into practice. Hands-on training
A. Selection of the Curriculum Management System sessions were scheduled and well attended by faculty,
We began by analyzing curriculum management systems students, course directors, and their administrative staff. A
currently in use at various Canadian and American medical videoconference was arranged to communicate, share and
schools. Several candidate software systems were carefully learn from the University of Calgary’s experience with the
evaluated for functionality, architecture, support, and cost implementation. They experienced numerous adoption
efficiency. Following a preliminary review, detailed and difficulties as their curriculum gradually evolved into a
exhaustive analysis of three systems was performed. These paperless system. A focus group included 15 to 20 students
included: 1) Curriculum Information System initially from each class, CMS representatives, and academic
developed by the Med-IT Department, University of Calgary, representatives to enhance communication between the
and its intellectual property rights are currently owned by U of implementation team and the class.
M.; 2) Web Eval, developed by One 45, a commercial medical After approximately six months of preparation and
informatics company; and 3) WebCT, commercial e- planning the CMS was launched in January of 2007. We
curriculum software. targeted one block from each pre-clerkship year for the pilot
The product analysis substantiated Curriculum Information project. Block III, taught during the 1st year consists of
System (CIS®) compared to Web Eval and Web CT is a Respiratory Medicine, Cardiovascular and Otolaryngology
comprehensive software package that links scheduling, courses. Block VI, launched in March of 2007 is taught during
learning objectives and resources. Successfully implemented the 2nd year and includes Gastroenterology, Hematology, and
previously at the University of Calgary, CIS® is an integrated Dermatology courses. There were 105 students in the 1st year
electronic medical curriculum and scheduling software that class and 98 students in the 2nd year class. The launch
can be delivered via the Internet and wireless hand-held consisted of fully integrated individualized schedules linked to
PDA’s. The goals of the pilot project were to explore the the curriculum, discussion forums, news groups, calendar, and
feasibility of implementing an electronic curriculum at U of M e-mail functionality. Since this was a pilot project, paper
and to evaluate strengths and weaknesses of the CIS software. resources and student websites were maintained as before.
We envisioned a CMS that would foster and sustain lifelong Therefore, three simultaneous systems existed for the
learning for medical students by providing a solid curriculum during the pilot project: paper-based notes;
undergraduate medical education foundation for future teaching material posted on the student website; and CMS.
growth. An ideal system would be a multi-user, web based Two systems coexisted for scheduling: a printed course
collection of electronic tools designed to assist medical schedule book and CMS.
faculty, students and administrators in the management of all A pilot project evaluation committee was created to
knowledge pertinent to learning objectives defined by the evaluate the feasibility of an electronic curriculum information
curriculum. system for the FOM, U of M, and to assess strengths and
weaknesses of the CMS software. The committee contained
B. Preparation for the Pilot Project Launch representatives from 1st and 2nd year classes, academic and
Prior to the pilot project, we prepared and uploaded CMS representatives, 1st and 2nd class year program
learning resources and teaching objectives to populate the coordinators, course directors for Block III and VI, an
system. All teachers were responsible for ensuring that their education expert, and administrative coordinators. Feedback
content was current, fulfilling the teaching objectives, and was solicited from the students, faculty, course directors and
uploading content prior to the teaching activity. Most teachers administrative staff through student and faculty surveys,
IMPLEMENTATION OF A CURRICULUM MANAGEMENT SYSTEM 91
Fig.1. A personalized interface for students and faculty opens up as the user logs in.
student and faculty log-in and page view data, and informally both 1st and 2nd year classes. Major increases (p <0.05)
through direct communication with the users. occurred between weeks 1 and 6, and the access remained
We hypothesized that a substantial increase in log-ins and stable thereafter. (Fig. 3) An independent survey conducted by
page views over time will demonstrate success, especially the students was submitted to the evaluation committee. The
since multiple systems were available. Data from week 1 was survey evaluated student satisfaction, and strengths and
compared to data from week 6. The sequential log-in and page weaknesses of the CIS® software.
view data was analyzed using chi-square analysis and
student’s t-test.
IV. RESULTS
Weekly system usage data were extracted from the system
logs and included a breakdown of various components and
modules. While faculty logins remained constant, student
logins continued to increase for both 1st and 2nd year classes. A
statistically significant increase (p <0.01) in student logins
occurred from week 1 to week 6 and then reached a plateau
(Fig. 2). Each student logged into the system approximately
once a day, including the weekends. This data is remarkable
since login was not necessary or required for the courses. Each
student received paper and other resources, including the
scheduling information and course material in advance.
Furthermore, with one login, students could access and
download all teaching material for the day. Curriculum page Fig. 2. Login data for 1st year class from week (WK) 1 to week 6 and both
views indicate that a learning resource was accessed by the 1st and 2nd year classes from week 7 to week 10. A significant increase in
user. These page views dramatically increased over time for log-ins (p <0.01) occurred from week 1 to week 6.
92 PAUL ET AL.
Fig. 3. Curriculum page views for 1st year class from week (W) 1 to week
6 and both 1st and 2nd year classes from week 7 to week 10. A significant
V. RECOMMENDATIONS FROM THE PILOT PROJECT
increase in page views (p <0.05) occurred from week 1 to week 6 and Based on analysis of the pilot data and comprehensive
week 10. consultation with all stakeholders, the evaluation committee
made several recommendations. These lessons learned at U of
A. Analyzing Students’ Responses M may assist other medical schools in their own development
It is important to analyze student evaluations and and implementation of electronic curricula.
comments in an appropriate context. The student acceptance 1) To prepare future physicians to deliver best medical
and adoption that we achieved during this pilot project met our care, the faculty and students at U of M determined
expectations. The project was conducted in the middle of an that an electronic curriculum and an integrated
academic year for the 1st year students after they had adjusted curriculum management system were high priorities.
to the paper based scheduling and curriculum system. The 2nd 2) The pilot project at U of M provided a unique
year students were generally less enthusiastic as this was their opportunity for students, faculty, and administrative
last pre-clerkship block before starting clinical rotations. staff to conceptualize, prepare, launch, troubleshoot,
Students also had ready access to the class website, an and evaluate an electronic curriculum. Extensive
unmonitored site that serves as a resource for communication preparation, free communication, thorough
and planning social events. The class website is also used as a consultation and engagement of all users from an
repository for learning materials. The student website is much early stage and throughout are required for successful
easier to navigate than the CMS which requires training and implementation.
has a learning curve to become proficient. While most students 3) The pilot project at U of M identified weaknesses of
were either supportive or neutral (Table 1), some were the software system and gaps in the wireless
resistant to change and thus were the most vocal opponents of infrastructure and hardware of the FOM. (Table 2)
the pilot project. Therefore, “cultural issues” or inherent This evaluation guided further modifications and an
“skepticism” for change definitely appeared to be an important upgrade of the CIS® software to improve interface,
factor in adoption. Such barriers during the implementation of usability, functionality and speed of the system.
electronic systems are well documented in the literature and 4) To gain faculty-wide engagement and adoption,
gradually decline before widespread acceptance and adoption duplication of the curriculum and schedule should be
[20] [21].
Feedback from the Faculty
A total of 53 instructors from 6 clinical disciplines participated
in this pilot project. Six of the course directors were intimately
involved with the roll-out and evaluation. All course directors
and many faculty members provided formal and informal
feedback on the electronic curriculum and evaluation of the
software. Faculty response was generally positive and
enthusiastic. Many course instructors used the system
extensively for scheduling functions and to update and
enhance teaching resources. Results of a survey sent to all
Fig. 4. Faculty response to the question: “I believe in the concept of an
Electronic Curriculum for the Faculty of Medicine at U of M”
IMPLEMENTATION OF A CURRICULUM MANAGEMENT SYSTEM 93
followed the classic innovation pattern. A few early adopting and has been successful partly because of a substantial private
institutions proceed rapidly while the majority of potential funding grant and strong leadership.
adopters make modifications slowly. The implementation
barriers identified by the survey were: lack of time, skill, and VII. CONCLUSION
incentive to develop the educational software. Another The vision of electronic learning is to foster lifelong
educational study by Riley at al. evaluated the effectiveness of learning for future physicians beginning with robust
the web-based teaching model as applied to perfusion undergraduate medical education that emphasizes strong self-
education [26]. Web-based virtual classroom (VC) learning skills and academic scholarship [30] [32]. E-learning
environments and educational management system (EMS) has been shown to be efficient, effective and has received
software were implemented independently and in addition to acceptance within the medical education community. Its
the live, interactive Internet-based audio and video success is only possible via an integrated, user-friendly,
transmission. Use of Internet-based VC and EMS software versatile and highly functional CMS. Such systems are
was a significant improvement over traditional models. currently in development and not widely available.
Interestingly, students recognized the learning efficiency of Considerable effort and resources are required for individual
on-line information but still preferred the traditional face-to- institutions to customize and implement a CMS. Our
face classroom environment. A Spanish study delivered and experience with a pilot project using an electronic curriculum
evaluated an educational web based cardiovascular course management system (CIS®) established a significant need for
prepared as adjunct educational material for the medical such a system, demonstrated that implementation is feasible,
student class [27]. Frequency of use, impact on in-class and found that adoption by faculty and students can be
attendance and student satisfaction, number of visits and expected. Successful implementation of e-learning in medical
quality of the website were assessed. The quality score rating education mandates a thorough consultation, effective
was a mean of 7.7 (scale: 1-10); 93.4% students attended the communication and engagement of all users.
class, and 88.2% of the students evaluated the web as a useful
or very useful addition to medical teaching. VIII. ACKNOWLEDGMENT
Kerfoot et al. conducted a multi-institutional randomized We would like to acknowledge and thank class of 2011
controlled trial to investigate the impact of an adjuvant web- students, Blocks III and VI instructors and administrative staff
based teaching program on medical student learning during of Undergraduate Medical Education Office (Pat McCullough,
clinical rotations [28]. Web-based teaching cases were Pre-Clerkship Program Administrator; Mark Torchia, Med I
developed covering four core urologic topics. Students were Coordinator; and Vieanna Beck, Administrative Assistant) at
block-randomized to receive web-based teaching on two of the the Faculty of Medicine, University of Manitoba for their
four topics. Before and after a designated duration at each superb enthusiasm and terrific support.
institution, students completed a validated 28-item web-based
test covering all four topics. Compared to controls, web-based REFERENCES:
teaching significantly increased test scores in the four topics at [1] Effective use of educational technology in medical education. AAMC
each medical school (p <0.001, mixed analysis of variance), 2007. https://aamc.org/ Accessed on August 8, 2008.
corresponding to an effect size of 1.52 (95% confidence [2] D. Sandham, A. Paul, S. Sharma. Integration of e-curriculum into
medical education. Gravitas 2008; 41:4-5.
interval [CI], 1.23-1.80). Learning efficiency increased three-
[3] Educating doctors to provide high quality medical care - a vision for
fold by web-based teaching (effect size 1.16; 95% CI 1.13- medical education in the United States. AAMC 2004. https://aamc.org/
1.19). Furthermore, students who were tested a median of 4.8 Accessed on August 8, 2008.
months later and those who received web-based teaching [4] B. Barzansky, S.I. Etzel. Educational programs in US medical schools,
demonstrated significantly higher scores than those who 2002-2003.
received non-web-based teaching (p = .007, paired t-test). This [5] H.B. Slotnick. How doctors learn: education and leaning across the
randomized controlled trial provides convincing evidence that medical-school-to- practice trajectory. Acad med 2001; 76: 1013-26.
web-based teaching can significantly and durably improve [6] J.P. Ward, J. Gordon, M.J. Field, H.P. Lehmann. Communication and
information technology in medical education. The Lancet. 2001;
medical student learning. 357:792–796.
Our experience developing and implementing an e- [7] P. Devitt, E. Palmer. Computer-aided learning: an overvalued
curriculum at U of M is similar to the journey undertaken by educational resource? Med Edu 1999; 33:136-139.
McGill University [2] [29]. David Fleiszer and Nancy Posel [8] T. Greenhalgh. Computer-assisted learning in undergraduate medical
were the leaders in development of an online, multimedia- education. BMJ. 2001; 322:40–45.
enhanced undergraduate medical curriculum. The electronic [9] C.L. Friend, C.L. Cole. Learner control in computer-based instruction: a
current literature review. Ed Tech. 1990; November: 47–40.
curriculum aimed to include: (1) learner-centered, self-
[10] D.J. Steele, J.E.J. Palensky, T.G. Lynch, N.L. Lacy, S.W. Duffy.
directed, and guided learning; (2) inherent interactivity of the Learning preferences, computer attitudes, and student evaluation of
technology; (3) a powerful means for integrated learning; (4) computerized instruction. Med Educ. 2002; 36:225–232.
ability to address a variety of learning styles experienced by [11] C.J. Bland, S. Starnaman, L.Wersal, L. Moorhead-Rosenberg, et al.
the students; and (5) opportunities for horizontal and vertical Curricular change in medical schools: how to succeed. Acad Med 2000;
curricular integration. The experience at McGill is ongoing 75:575–94.
[12] M. Vogel, D.F. Wood. Love it or hate it? Medical students’ attitudes to
computer-aided learning. Med Educ. 2002; 36:214–215.
IMPLEMENTATION OF A CURRICULUM MANAGEMENT SYSTEM 95
[13] M.L. Rice, E.K. Wilson, W. Bagley. Transforming learning with [23] J.D. Hoban, J.B. Schlesinger, R.P. Fairman, M.M. Grimes. Electrifying a
technology: lessons from the field. J Tech Teacher Educ 2001; 9(2):211– medical school course: a case study. Teach Learn Med. 2003
30. Spring;15(2):140-6.
[14] B.M. Karnath, W. Thornton, M. DasCarlos. Pilot study of computer- [24] M. Gotthardt, M.J. Siegert, A. Schlieck, S. Schneider, A. Kohnert, M.W.
based self-teaching in cardiac auscultation. Med Educ. 2003; 37:1048– Gross, et al. How to successfully implement e-learning for both students
1049. and teachers. Acad Radiol. 2006 Mar; 13(3):379-90.
[15] A.J. Lipman, R.M. Sade, A.C. Glotzbach, C.J. Lancaster, M.F. Marshell. [25] W.D. Hendricson, F. Panagakos, E. Eisenberg, J. McDonald, G. Guest,
The incremental value of internet-based instruction as an adjunct to P. Jones, et al. Electronic curriculum implementation at North American
clasroom instruction: a propsective randomized study. Acad Med. 2001; dental schools. J Dent Educ. 2004 Oct; 68(10):1041-57.
76:1060–1064. [26] J.B. Riley, J.W. Austin, D.W. Holt, B.E. Searles, E.M. Darling. Internet-
[16] B.L. Ingerman. Form follows function: establishing the necessary based virtual classroom and educational management software enhance
infrastructure. In: Barone CA, Hagner PR, eds. Technology-enhanced students’ didactic and clinical experiences in perfusion education
teaching and learning. San Francisco: Jossey-Bass, 2001. programs. J Extra Corpor Technol. 2004 Sep; 36(3):235-9.
[17] J. Daniel. Computer-Aided Instruction on the World Wide Web: The [27] X. Gómez-Arbonés, A. Ferreira, M. Piqué, J. Roca, J. Tomás, J.L.
Third Generation. J of Econ Ed. 1999. pp. 163–174. Frutos, et al. A cardiological web as an adjunct to medical teaching:
[18] S. Chodorow. Educators must take the electronic revolution seriously. prospective analysis. Medical Teacher 2004; 26:187-191.
Acad Med. 1996; 71:221–226. [28] B.P. Kerfoot, H. Baker, T.L. Jackson, W.C. Hulbert, D.D. Federman,
[19] H.K. Muhsin. The use of e-learning techniques to improve the medical R.D. Oates, et al. A multi-institutional randomized controlled trial of
education. ICTTA 2008. http://ieexplore.ieee.org/ Accessed on August adjuvant web-based teaching to medical students. Acad Med. 2006 Mar;
2008. 81(3):224-30.
[20] R.M. Palloff, K. Pratt. Lessons from the cyberspace classroom: the [29] D.M. Fleiszer, N.H. Posel. Development of an undergraduate medical
realities of online teaching. San Francisco: Jossey-Bass, 2001. curriculum: the McGill experience. Acad Med. 2003 Mar; 78(3):265-9.
[21] H.S. Wentzer. Technology development and implementation in the [30] S. Schultze-Mosgau, T. Zielinski, J. Lochner. Web-based, virtual course
public health institutions and pedagogical task for the learning units as a didactic concept for medical teaching. Med Teach. 2004 Jun;
organization. Stud Health Tech Informatics 2000; 77:264–70. 26(4):336-42.
[22] M. Broudo, C. Walsh. MEDICOL: online learning in medicine and [31] J.G. Ruiz, M.J. Mintzer, R.M. Leipzig. The impact of e-learning in
dentistry. Acad Med. 2002 Sep;77(9):926-7. medical education. Acad Med 2006: 81:207-212.
[32] K. Masters, T. Gibbs. The spiral curriculum: implications for online
learning. BMC Med Educ. 2007 Dec 21; 7:52-55.
Use of Ontology for Reusing Web Repositories
for eLearning
Punam Bedi, Hema Banati, Anjali Thukral
Abstract— eLearning scenario has witnessed a shift from framework that provides latest updated learning material,
instructional to adaptive learning. The focus is now moving from semantically searched from web repositories for the entered
closed to open scenario. However, learning in such situation query.
relegates to the task of retrieving as the information is largely
scattered across the web. Searching such voluminous information The paper is organized as follows: Section II presents the
is practically infeasible and leads to incomplete or irrelevant background work, briefly defining eLearning, advantages of
results. Hence, a need is felt to automatically integrate the vast Semantic Web (SW) for it and ontology. Section III describes
pool of information present in varying forms on web. This paper the framework and algorithm for reusing web repositories for
proposes a framework that searches for information semantically eLearning. Finally section IV discuss future work and
across all existing web repositories for a given query. The concludes the paper.
approach utilizes WSD (word sense disambiguation) algorithm to
semantically analyze and expand the query. A personalized
II. BACKGROUND WORK
knowledge base for learning purpose is subsequently created by
deploying agents to compile information from various learning eLearning is defined with eight dimensional characteristics
resources. The integrated organization of knowledge provides a as: Pull: Learner determines agenda (delivery), Reactionary:
single point reference for the posted query. It thus increases the Responds to problem at hand (responsiveness), Non-linear:
probability of the learner obtaining complete and relevant Allows direct access to knowledge in whatever sequence
information from a single reference. This framework therefore makes sense to the situation at hand (access), Symmetric:
allows a learner to concentrate on learning than the searching Learning occurs as an integrated activity (symmetry),
process thus enhancing the learning experience. Continuous: Learning runs in parallel to business tasks and
never stops (modality), Distributed: Content comes from the
Keywords-eLearning; eLearning; web repositories; query interaction of the participants and the educators (authority),
expansion; software agents; personalized ontology. Personalized: Content is determined by the individual user’s
needs and aims to satisfy the needs of every user
I. INTRODUCTION (personalization) and Dynamic: Content changes constantly
through user input, experiences, new practices, business rules
In the past practice traditional systems like LMS, LCMS [1, and heuristics (adaptivity) [6,7]). eLearning characteristics can
2] and other intelligent systems like AWBES [3] have played effectively be achieved through the new WWW architecture,
an important role in the evolution of eLearning [4]. These Semantic Web [8] (SW). Research [6, 7, 9, 10, 11, and 12]
systems emphasis on providing an integrated environment for a shows the benefits of using semantic web as the platform for
learner, but within a separate and closed scenario, viz within an eLearning. Conceptually the SW represents a layered
educational organization or within some paid association. This architecture (Fig 1), where each level provides different degree
restriction significantly narrows the educational inputs, of expressiveness. The bottom most layer allow user to provide
depriving a learner from vast pool of information thus a controlled vocabulary and namespaces. Next layer, Resource
precluding from the efficiency of learning [5]. Learning in open Description Framework (RDF) represents information about
scenario is also long-winded. The presence of voluminous and resources in the World Wide Web. It consists of a triplet –
scattered information in varying formats deviates and hinders resource, property and value. Ontology vocabulary, the next
the learning process of a learner. A need is hence felt for a upper layer, adds more semantics to RDF by allowing a better
system that can automatically and dynamically organize the specification of constraints on concepts. The rules for explicit
information from all possible resources to a well formed inference are established by Logic layer which are then
structure. This structured learning material helps a learner for a executed by the Proof layer. The top most layer, Trust provide
better understanding of concept. This paper, therefore mechanisms to imbibe trust in a given proof [13]. The
addresses above mentioned need and proposes an eLearning semantic web as whole, acts as a decentralized knowledge bank
that provides meaning to a learner's search and thus allows
Manuscript submission date: 26th October 2009.
extracting most appropriate learning material.
Punam Bedi,( [email protected]), Hema Banati,([email protected])
Anjali Thukral ([email protected])
Dept. of Computer Science, University of Delhi, India
97
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_17, © Springer Science+Business Media B.V. 2010
98 BEDI ET AL.
Trust
a large scale [18] to facilitate a vast group of learners.
Standards were established to generalize the structure for such
Proof
metadata as Dublin Core [19] and LOM (Learning Object
Digital Signature
Logic Metadata) [20]. However, there were issues in using LOM.
Ontology Vocabulary They represent static repositories and suffer from various
shortcomings [21]. The ontology based metadata provided a
RDF + rdfschema
perfect blend of eLearning systems and ITS (Intelligent
XML + NS + xmlschema Tutoring Systems) [22] in closed scenario, overcoming most of
Unicode URI the earlier shortcomings. Subsequently, the ontology layer was
identified as a standard layer of the knowledge based systems
[23]. Existing work of open learning systems either provides
Figure 1: Layered Architecture of Semantic Web links to the World Wide Web documents, based on refined
query searches or exclusively build on semantic web. Semantic
The middle layer and the backbone of Semantic web, web conveys meaning to the content whereas existing WWW
Ontology Layer, forms knowledge about a certain domain by has evolved as a powerful information portal.
providing relevant concepts and relations between them. Often
cited definition of ontology is “an explicit specification of a
conceptualization” [14] i.e., an ontology defines (specifies) the III. FRAMEWORK
concepts, relationships, and other distinctions that are relevant This paper proposes an open learning framework that
for modeling a domain. The specification takes a form of provides the leverage of both worlds. Unlike other existing
representational vocabulary (classes, relations, and so forth) systems, it dynamically integrates all available resources on
which provides meaning for the vocabulary and formal web to construct ontologies. This integration is established on
constraints on its coherent use [15]. the posted query. Words contained in the query are
Mathematically, a core Ontology is defined as a structure, [16] semantically analyzed for sense based query expansion.
Expanding query in this way improves the probability of
O := (C,≤C , R,,≤ R ) retrieving relevant information drastically [24]. Thus, the
system first formulates the query and then integrates resources
consisting of: to form an exhaustive and updated ontology.
i) two disjoint sets C and R whose elements are called This framework is based on three types of ontologies:
concept identifiers and relation identifiers, respectively.
• NLO (Natural Language Ontology): It contains lexical
ii) a partial order ≤ C on C , called concept hierarchy or relations between the language concepts. They are
taxonomy, large in size and do not require frequent updates. They
iii)a function : R → C called signature, and
+ represent the background knowledge of the system to
understand the query formed in the natural language.
iv)a partial order ≤ R on R , called relation hierarchy, where
The framework uses NLO, WordNet to expand the
r1 ≤ R r2 implies | (r1 )| = | (r2 )| and original query.
i ( (r1 )) ≤ C i ( (r2 )) for each 1 ≤ i ≤ | (r1 )| • DO (Domain Ontology): It captures knowledge of one
particular domain, e.g., operating system, computer
and Knowledge Base is defined as a structure [16]: etc. These ontologies provide detailed description of
KB := (C KB , RKB , I,C ,R ) the domain concepts. In this framework it belongs to
the horizontal upper layer of the two level ontology
consisting of: [25] structures. (Fig 2).
• OI (Ontology Instances): They are the main resource
i) two sets C KB and RKB , of knowledge presented in the web and on the learner's
ii) a set I whose elements are called instance identifiers (or local machine. These instances consist of interlinked
instances or objects for short), web resources (includes web pages etc.) forming a
iii)a function C : C KB → (I ) called concept instantiation, horizontal hierarchy (lower layer of two level
organization - Fig 2) among themselves and also
( )
iv)a function R : RKB → I called relation instantiation.
+
vertically linked to their related concepts in the
Ontologies have the potential to construct a dynamic and domain ontology. These are generated automatically
semantic structure of e-content that can be shared, reused and and updated frequently by software agents. DO and OI
enhanced with trust for learning purposes. Since the invention forms the local knowledge base.
of semantic web in 2000, many researchers have proposed The framework (illustrated in Fig: 2) initializes the process by
various architectures and frameworks that use this technology expanding query through query engine. The expanded query is
to enhance the learning experience of learners [17]. Most of then used to extract contents from web repositories. This
these frameworks initially started with the use of metadata. retrieved information is ontologically organized to form
Efforts were made to convert eLearning courses to metadata on
USE OF ONTOLOGY FOR REUSING WEB REPOSITORIES FOR ELEARNING 99
Multimedia
Relational Ontology files
databases
Human
Text (doc,
s/w agents expert
pdf, html)
Query Engine
Image Query (Qo)
file Formalized
Audio query (Qf)
file Movie
PA Learner
OA Domain Ontology,
Two level organization of domain
ontology and ontology instances Ontology Instances
Compiled
learning
Synthesizer
material
Query based
ontology extraction
knowledge base. The output is subsequently re-framed to a where C x is the best possible configuration of senses for Qo .
presentational format by the Synthesizer.
One of the possible expansions for the above query can be seen
Remaining section explains the process of query expansion as:
and content extraction followed by algorithmic summarization
of the complete process of information retrieval. Q f = {medium, Ethernet, coaxial cable, radio frequency
A. Query expansion signal, wireless local area network}.
The query engine (ref Fig: 2) lexically extract words from The words, medium and Ethernet have been semantically
the query for the expansion. As an example, let the query be: analyzed to deduce new words that are used as medium for
“medium used in Ethernet”. It generates a set of words from Ethernet. Though, the words like co-axial cable etc. does not
original query as, exist in the original query, still the learner benefits by obtaining
the intended information related to these supplementary words.
Qo = {medium, Ethernet}
In general, B. Content extraction from web repositories
Web repositories contain learning material in varying
Qo = { w1 , w2 ,..., wk } formats. IGA (Information Gathering Agent) manages the
searching process to gather information from all possible
where, wk are the concepts extracted from the entered query. resources. This information is organized semantically by OA
The query engine formalizes the query using a mapping (Ontology Agent) to update local personalized knowledge base.
OA deals with each source format using specific procedure to
function, . The function uses WordNet 1 and Word Sense give information an ontological structure. Following are the
Disambiguation (WSD) algorithm [24] to expand the query. different formats handled by OA (ref. Fig. 2).
The expanded query consists of semantically analyzed set of
words that results from the maximum number of intersecting i) Text documents: The automatic generation of ontology from
nodes in the semantic networks, generated from WordNet text (ontology population) [26] is achieved with KAON
(Karlsruhe ontology) based tool TextToOnto [16]. The
synonym sets (synsets) of wk , k = 1, .. , n. Thus, resulting in a agent inputs text as a corpus and generates ontology using
formulated query, this tool. Generated ontology is later pruned or enhanced
through the domain expert, if needed. The automatically
Q f = { w1 , w2 ,..., wk , C x } generated ontology does not produce satisfactory results,
but nevertheless provide some basic structure of concept
hierarchy. Generation of quality ontologies without the
human intervention is still a topic of research.
1 http://wordnet.princeton.edu/
100 BEDI ET AL.
ii) Multimedia files: Ontology instances are used to organize : Qo → Q f is formed where, is a mapping from
multimedia files (audio files, video files, presentations,
images etc.). These files are arranged in ontological form the original query, Qo to a formalized query,
and at the same time linked to its related concept in Domain
Ontology (DO). Q f consisting of a set of semantically expanded words
iii)Human experts: Human experts' knowledge can be (explained in section III/A).
communicated through e-mail, blogs of FAQs (Frequently c) PA then, submits Q f , a formalized query to Ontology
Asked Questions), or chatting. This is achieved by linking
useful urls in the ontology instances and then linking them Agent, OA.
to related concepts in DO. Software agents [27] can be iii)OA performs two tasks simultaneously:
trained to take care of such conversations on behalf of the a) First, it searches the most relevant ontology in its local
passive learner. knowledge base consisting of domain ontologies and
iv)Ontology: Reusing existing ontologies offer a much ontology instances.
cheaper alternative than building new ones from scratch b) Secondly, it instructs Information Gathering Agent,
[28]. It provides one of the valuable repositories on web IGA to explore learning material in the World Wide
and reduces efforts and time to retrieve relevant content. Web and semantic web using Q f (explained in section
However, the problem of determining relevance of the III/B).
ontologies is not much different from finding a relevant c) IGA returns to OA with:
searched document. A brilliant approach for automatic
ontology construction from the existing ontologies on the 1 ) a set of top n ranked links (URLs), Ln after
web was proposed [28]. It extracts the relevant parts of the searching WWW and
top ranked (ranked according to some criteria) ontologies 2 ) top m ranked ontologies, Om after searching
and merge those parts to acquire the richest domain
semantic web.
ontology. The ontologies on the web are searched using the
iv)Steps i) and ii) are repeated by PA till the learner is active
semantic web search engines like Swoogle or Ontosearch.
or wishes to update its knowledge base passively.
This approach requires further exploration of areas such as
Consequently, PA keeps the learner updated by
ontology ranking, segmentation, mapping, merging and
communicating with OA, and OA subsequently keep its
evaluation.
knowledge base updated by frequently communicating with
v) Software Agents: Software agents are used by a system for
IGA.
personalized search and interaction with almost each form
of web repository. As the future vision, one may find s/w v) OA update its knowledge base, KB with Ln and Om to
agents on the web, similar to the human learner. It will be a finally generate the result.
normal scenario to have web full of s/w agents vi)Result is provided to the learner in one or both of the two
communicating to each other and negotiating for accessing forms:
relevant content for its owner. The challenge is to form a a) a set of ontologies ranked according to the relevance,
force of such good behaviored servants on web! b) top ranked ontology, forwarded to the 'Synthesizer', in
vi)Relational databases: Web repository also exists in the order to generate a readable document with all its
form of database. Majority of data on old systems consist of instances as links for the learner.
databases. Such resources of eLearning cannot be The above algorithm represents a complete process of
neglected. This framework maps the relational schemata of reusing information by extracting only relevant content and
eLearning databases found on web, to ontology according providing single point reference to the compiled information
to their underlying intrinsic relationships [29]. Tables, for the learner. By using this algorithm, the framework creates
Fields and records are mapped to the ontology as class an open scenario for learning, where a learner automatically
(concept), property and instances. They are further adjusted gets the relevant and dynamically updated information related
to get accommodated in the structure of proposed ontology. to the required concept without getting deviated from the flow
All of the retrieved information from web repositories by of learning. This allows a learner to concentrate more on the
above methods is organized in two level ontological structures concept understanding instead searching for the relevant
to form a learner's personalized knowledge base. This content, which effectively enhances the learning experience.
knowledge base is then used further to answer queries related
to same concept. The following algorithm generates a IV. CONCLUSION AND FUTURE WORK
personalized knowledge base for a given query.
A framework has been proposed that reduces the chances of
i) A learner enters a query in natural language, Qo . loosing out relevant content available on web significantly to
ii) A learner's personalized interface agent (PA) explores create a personalized ontology for learning purpose with the
query entered by learner to infer the context and semantic. additional vision of agent communication. It takes a query
a) Through query engine, Qo query is formalized by PA. from learner, semantically expands it and gathers content from
web repositories with the assistance of multiple software
b) The query engine extracts most appropriate words by agents. These agents work for the active learner as well as
applying Natural Language Ontology (WordNet) and
searches content for the passive learner. Moreover, a
WSD algorithm [24]. Thus a mapping,
personalized agent is provided to assist the learner at various
USE OF ONTOLOGY FOR REUSING WEB REPOSITORIES FOR ELEARNING 101
learning points. At present, the framework has been partially [14] Gruber, T. R., “A Translation Approach to Portable Ontology
Specifications”, Knowledge Acquisition, 5(2):199-220, 1993.
implemented that uses JADE and a tool TextToOnto for
[15] Thomas Gruber, “Ontology (Computer Science) - definition in
extracting text and transforming it to ontology automatically. Encyclopedia of Database Systems”, Ling Liu, Tamer M Özsu (Ed),
In future, other modules will be implemented and explored Encyclopedia of Database System (2008)
further to provide a more relevant ontology generation for the [16] Erol Bozsak, Marc Ehrig, Siegfried Handschuh, Andreas Hotho,
learner's knowledge base. A more formalized agent Alexander Maedche, Boris Motik, Daniel Oberle, Christoph Schmitz,
communication on the web and setting up a descent behavior Steffen Staab, Ljiljana Stojanovic, Nenad Stojanovic, Rudi Studer, Gerd
Stumme, York Sure, Julien Tane, Raphael Volz, Valentin Zacharias,
will also be considered. “KAON – Towards a Large Scale Semantic Web”. Proc. of the 3rd Intl.
Conf. on E-Commerce and Web Technologies (EC-Web 2002), pages
REFERENCES: 304–313, 2002.
[1] S.R. Robbins, “The evolution of the Learning Content Management [17] Janez Brank, Marko Grobelnik, Dunja Mladeni , “A Survey Of
System”, ASTD's Source for E-learning, 2002. Ontology Evaluation Techniques”, In Proceedings of the Conference on
[2] Brandon Hall Research Article, 2007 [Online]: hhtp://www.brandon- Data Mining and Data Warehouses (SiKDD) 2005
hall.com/ [18] Jan Brase and Wolfgang Nejdl, “Ontologies and Metadata for
[3] Peter Brusilovsky, “KnowledgeTree: A Distributed Architecture for eLearning”, pages 579–598. Springer Verlag, 2003.
Adaptive E-Learning”, ACM 1-58113-912-8/04/0005, May 17-22, [19] The Dublin Core Metadata Initiative, http://dublincore.org
New York, USA. [20] Learning Technology Standards committee of the IEEE: Draft Standard
[4] Anjali Thukral, Richa Sharma, Hema Banati, Punam Bedi, , “Modeling for Learning Object Metadata, July 2002, IEEE 1484.12.1-2002,
Roles in E-learning: MoRel”, ACVIT'07 - International Conference on http://ltsc.ieee.org/wg12/files/LOM_1484_12_1_v1_Final_Draft.pdf
Advances in Computer Vision and Information Technology,612-620, [21] Amal Zouaq and Roger Nkambou, “Building Domain Ontologies from
Aurangabad, November 28-30, 2007. Text for Educational Purposes”, IEEE Transactions on Learning
[5] Michael G, Dimitry V.,Pavel, Denis, “Intelligent Educational Technologies, vol. 1, no.1, January-March 2008
Environments Based on the Semantic Web Technologies”, Proceeding, [22] C. Brooks, J. Greer, E. Melis, and C. Ullrich, “Combining ITS and
IEEE International Conference on Artificial Intelligent Systems eLearning Technologies: Opportunities and Challenges,” Proc. Eighth
(ICAI'02) 0-7695-1733-1/02, 2002. Int’l Conf. Intelligent Tutoring Systems (ITS ’06), pp. 278-287, 2006.
[6] Drucker, P. ,“Need to Know - Integrating e-Learning with High Velocity [23] Neches, R., Fikes, R. E., Finin, T., Gruber, T. R., Patil, R., Senator, T.,
Value Chains”. Delphi Group White Paper. 2000 www.delphigroup.com & Swartout, W. R. “Enabling technology for knowledge sharing”. AI
[7] Yas A. Alsultanny, “eLearning System Overview Based on Semantic Magazine, 12(3):16-36, 1991.
Web”, The Electronic journal of e-Learning, Volume 4 Issue 2 (111- [24] Navigli, Roberto and Velardi, Paola, “An Analysis of Ontology-based
118) 2000. Query Expansion Strategies”, In: Proceedings of Workshop on Adaptive
[8] T. Berners-Lee, J. Hendler, O. Lassila , “The Semantic Web”, Scientific Text Extraction and Mining at the 14th European Conference on
American, May 2001, pp. 28-37. Machine Learning, 2003.
[9] Ljiljana Stojanovic, Steffen Staab, Rudi Studer, “eLearning based on the [25] Antonio De Nicola, Michele Missikoff, and Frederica Schiappelli,
semantic web”, WebNet2001 - World Conference on the WWW and “Towards an Ontological Support for eLearning Courses”, R. Meersman
Internet. et al. (Eds.): OTM Workshops 2004, LNCS 3292, pp. 773–777, 2004.
[10] Rosic, M.; Stankov, S.; Glavinic, V., “Application of semantic web and [26] Alexander Maedche and Steffen Staab, “Ontology Learning for the
personal agents in distance education system”, Electrotechnical Semantic Web”, IEEE Intelligent Systems and Their Applications, Vol.
Conference, MELECON 2002. 11th Mediterranean,Volume , Issue , 16, No. 2. (2001), pp. 72-79, 2001.
2002 Page(s): 542 – 546, 2002. [27] James Hendler, "Agents and the Semantic Web," IEEE Intelligent
[11] Hatem, Muna S.,Ramadan, Haider A.,Neagu, Daniel C., “e-Learning Systems, vol. 16, no. 2, pp. 30-37, Mar./Apr. 2001,
based on context oriented Semantic Web”, Journal of Computer Science doi:10.1109/5254.920597, 2001.
1(4):500-504,2005. [28] Harith Alani, “Position Paper: Ontology Construction from Online
[12] Sylvain Dehors, Catherine Faron-Zucker, “Reusing Learning Resources Ontologies”, WWW 2006, May 23–26, 2006, Edinburgh, Scotland.
based on Semantic Web Technologies”, ICALT 2006: 859-863. ACM 1-59593-323-9/06/0005, 2006.
[13] Bittencourt I.I., Costa, E., Soares, E. and Pedro, A., “Towards a New [29] Zhaohui Wu, Yuxin Mao, and Huajun Chen, “Subontology-Based
Generation of Web-based Educational Systems: The Convergence Resource Management for Web-Based e-Learning”, IEEE Transactions
between Artificial and Human Agents”, IEEE multidisciplinary on Knowledge and Data Engineering, vol. 21, no. 6, June 2009.
education Magazine, vol. 3, March 2008.
Actual and Future Trends of the E-education
and Its Costs Specification in the Czech Republic
M. Hedvicakova, P. Maresova
Department of Economics
Faculty of Informatics and Management, University of Hradec Kralove
Rokitanskeho 62, 500 03 Hradec Kralove, Czech Republic
Abstract – The contribution is focused on the specifics of distance We cannot forget the school system, where a great majority of
education in the Czech Republic. First, reasons of emergence of both universities and high schools, and to a lesser extent also
distance education in the Czech Republic and abroad are basic schools, takes advantage of various education and
analyzed. Further, two possible directions of development of e- instruction software.
education in the Czech Republic are outlined, where the first
direction represents an expansion of capacities of the existing
full-time courses. The second possible direction consists of
customization of educational courses according to specific needs II. FORMS OF STUDIES IN UNIVERSITIES
of students. The contribution includes definitions of individual IN CZECH REPUBLIC
participants of the educational process, typology of costs and
revenues and break-even point modeling. Currently three ways can be applied in the process of
instruction: present lessons, i.e. traditional, face-to-face
teaching and learning; distance education supported by ICT,
I. INTRODUCTION e.g. e-learning, virtual learning environment; and blended
learning, i.e. combination of present and distance elements.
With the arrival of new technologies and varied demands in
In the present form of instruction in economic subjects
society, universities must continuously seek new ways to
teacher faces two main problems: students´motivation and
combine advantages of modern technologies with
interest, and the level of knowledge. The former one can be
requirements of target customer groups. Requirements can be
strongly influenced by the teacher, his/her personality, image,
for example: demographic, locality, time or costs etc. Virtual
professional qualities, computer literacy, etc. And it results
universities offer the way to satisfy these requirements while
into ways how to solve the latter problem. Numerous suitable
continuing to multiply advantages of information and
examples, applications from real life, new technologies
communication technologies and thus they ensure a greater
supporting “boring“ calculations, and last but not least
return of realised investments and thereby also the growth of
didactic aspects – these are the means which can help. If the
a university as such.
teacher is able to use appropriate methods and forms to
There are already many virtual universities in the world.
explain the topic, present examples, practice them, motivate
Especially US universities have become pioneers of this type
students to follow-up activities, organise team and project
of education for the reason of their strong financial position
work – all the mentioned aspects may support the required
and a high level of technological equipment. Also e-education
features mentioned above. Implementation of new
in Australia has broken through with success. Interest in
technologies in the process of instruction calls for both
distance programs has been growing in these continents for
special equipment of the classroom and for the teacher who is
the reason of high commuting distance to reach education.
able to work with them in an appropriate way. Modern
State institutions lead by the European Union put a great
software programmes neither work nor bring increasing
accent on strengthening and changing of the approach to
interest and effectiveness by themselves. Only such a teacher,
education. This can be also taken advantage of by regional
who is a professional in both the subject field and
universities, which can flexibly respond to the demand for job
methodology, evokes the required activities from students.
positions in the labour market. They can either focus on
Having a PC, data projector etc. in the classroom, using
retraining courses for employment bureaus or to create
PowerPoint presentations, links to Internet sources, it is a
“customized” courses for commercial organizations.
common standard these days. Various types of software are
Supranational organizations and large companies have
able to present graphs and their animations, calculate values
already been using e-learning education methods for training
and structure them to tables, everything with the aim to make
of their employees. With regard to the number of employees,
the process of instruction easier for both students and
they have become aware of advantages of on-line courses and
teachers. Teacher’s ability to use these tools appropriately and
they try to use them as their competitive advantage.
support all activities by personal interest and knowledge,
Currently we can observe that e-learning education has
these are the qualities that count for much.
reached all areas, industry as well as energetic, financial
sector, healthcare, telephone market, state sector and army.
103
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_18, © Springer Science+Business Media B.V. 2010
104 HEDVICAKOVA AND MARESOVA
It can be said the same goes for the distance education, but III. TWO PATHS FOR UNIVERSITIES
twice more. As there is no face-to-face contact between the
teacher and students, the organisation is more demanding on One of the paths for universities is the above mentioned
both sides, teacher’s and students´. Although the distance distance studies as another form of education. E-learning can
form of education is applied for tens of years, its main be applied as an effective support of all forms of studies. The
development came in last few years because of immense advantage for the existing “stone universities“ is elimination
implementation of information and communication of space and staff inadequacy of today's education. In
technologies into all spheres of life. The adequate use of the consideration of the fact that there are no time or distance
technologies must be emphasized, just because of the absence barriers, interest in this form of studies has also been growing
of direct contact. The quality of all used tools and methods, among new groups of students (employed students,
including constant motivation, strongly influences students´ handicapped, women on maternity leave, prisoners and the
results and success. Professional teams of teachers, IT like). People have also been choosing this form of studies for
workers, economists and other specialists participate in the reason of interactivity of courses.
successful organisation and running the distance education. The advantage for the university itself lies in a strong
Blended learning combines the best of the above mentioned background, high-quality professional staff and tradition of a
forms, i.e. present lessons are supported by ICT, self-study university as an educational institution. Another benefit for
activities and sources provided by the Internet are frequently the university is reinforcement of its image as a holder of new
used, etc. [2] modern technologies and a wide spectre of courses.
As regards newly emerging virtual universities, the
Unlike other states with a larger population or larger universities which get involved in the project exchange the
distance barriers (USA, Australia etc.), where the tradition of subjects not offered by their faculty and thereby expand the
taking advantage of various education methods is greater, the circle of their students.
market with virtual education is still only developing in the The second path for universities is to create commercial e-
Czech Republic. learning courses. These are courses customized for individual
Only few Czech universities offer a compact study program companies or based on needs of the market or labour bureaus.
by means of the e-education method. These are for example A greater accent is put on economical aspects in these
the University of Economics, Prague, offering a follow-up courses. Another effect is interconnection of universities with
master's two-year program Economics and Management see needs of the given region and therefore again reinforcement
[3], or VŠB – Technical University of Ostrava, Faculty of of the image of university as a core of erudition.
Economics, offering the study field Quantitative Management
Support [4] in its follow-up master's studies. Combined
programs can be found in many universities including UHK. IV. FRAMEWORKS FOR COSTING
In our republic, the transition to this new type of education
has been stimulated above all by insufficient capacities of If we want to calculate costs of a course, first we have to
universities for full-time studies, economic conditions of define key types of costs.
students (studying while employed), pressure of the European
Union on the growth of the number of university educated 1. Categorisation of E-learning Costs
inhabitants etc. Another important motive is increasing In the process of comparing costs and revenues in the
attractiveness of studies by combining study programs of distance and traditional education, a clear definition of the
individual universities, modernity of the topic of information basic terms is necessary, i.e. what they consist of. Not the
society etc. whole virtual university is evaluated, but only one distance
Also university networks have been emerging educational course.
experimentally (e.g. project of Inter-university studies IUS at
the universities: University of Hradec Kralove, University of E-learning study costs
Western Bohemia in Pilsen and University of Tomas Bata in The standard structure for cost and revenue classification is
Zlin, Technical University of Ostrava) [5]. Or Erasmus used as it is in the common sphere of commerce, i.e. the
Virtual Economics & Management Studies Exchange variable and fixed costs will be counted.
(EVENE) – the international project EVENE is supported
from resources of the European Union within the program a) Fixed costs
eLearning and it provides students with the possibility to • Designing study materials – when creating an e-learning
study at foreign universities as well. [6] Nine international course, at the very beginning study materials must be
universities are involved in the project. prepared, including their electronic form in the e-
environment. This is a non-recurring action.
• Authors´ royalties for creating study materials in the text
form – authors must be evaluated in terms of finance.
• Printing study materials
• Costs of designing an e-course
ACTUAL AND FUTURE TRENDS OF THE E-EDUCATION AND ITS COSTS SPECIFICATION 105
c) Total costs
The total costs cover variable and fixed costs.
TC = FC + VC
TC TC
VC
FC
Fig. 1: Fixed costs
2. Profits
The traditional and distance course revenue varies and
consists of two items:
a)Contribution on each student (either from a state or
Q public institution supporting education).
Fig. 2. Stair-step fixed costs b) Contribution from each student (some
universities require full or partial fees from students).
b) Variable costs Total of these sums creates the income per student.
• Virtual environment user license
• Running an e-course – tutors and administrators of each
group are evaluated 3. The Break-Even Point
• Consumer material costs Finding the rate of return comes from a principle of
• Rental service costs paid for tutorial rooms comparing costs and profits and finding a number of students,
• Printing study materials who have to be in the course so that the course is profitable.
• Staff costs In economical formulation it is the break-even point. When
costs and profits are equal, the more students above the
minimal number we have, the higher profit we will receive.
Fixed Costs:
Fig. 3. Variable costs F1 - designing study materials
106 HEDVICAKOVA AND MARESOVA
it would evaluate the financial requirements and it would division but a dynamic model which could be adjusted to the
recommend the amount of “tuition” for the given course. conditions of an institution.
Another possible element of this questioning would be Working in a virtual learning environment brings
searching for the most similar courses and enabling contact advantages to universities, e.g. decreasing costs on operation,
with parent educational centres that realised these courses to travelling expenses, accommodation, effective training,
share their experiences, so that we can avoid problems and continuous updating, etc.
become aware of what can be done better in the case of the Even though the distance and virtual study is running in
given type of the course. some large and rich states (Great Britain, France, Germany,
This all would lead to a relatively exact determination of Ireland), it is financially demanding and cannot work without
financial requirements of the course and thus of the considerable state grants and support. The cost calculations
corresponding needs of financial resources. cover author’s remuneration, salaries, distribution costs,
rental service costs, etc. Starting costs which may include
The main objective of such an interactive model is to establishing (building) administrative and study centres,
increase markedly the effectively of the financial resources preparing LMS, are very high. That is the reason why the
expended on the distance education and to achieve such a virtual study is supported from several sources (foreign
state, when a maximum of collected resources would be supportive programmers, tuition, production of teaching aids
invested into an effective functioning of the educational for other customers, selling the study materials, etc.) [7], [8],
system and a minimum of resources would be expended on [11].
the bureaucratic machinery, that represents an “unnecessary
evil”. An increasingly higher number of educational courses
have been appearing that, after the end of the subsidy REFERENCES
program, “suddenly loose” their financial benefit for the
educational institution (the course is loss-making after the end [1] T. ARCHIBALD, M. NIELSEN, T. TORY NEIWERT, K. OLIVEK,
Y-S. LEE, L. MARLER, Virtual Universities: A Current Snapshot,
of the subsidy period) and represents a financial burden in the Utah State University URL: http://www.kwadeoliver.com/it-
form of negative effectively for the institution. Correct portfolio/research-writing/Virtual%20Universities.pdf [online], [2006-
setting of the costs and profit side of the course already in its 11-8].
beginnings is a necessary key factor for a long-term effective [2] M. HEDVICAKOVA, I. SIMONOVA, Contemporary Ways to
Teaching Economic Subjects, 38th IGIP Conference, September 6 - 9,
functioning of the distance education. [11] 2009, Graz, Austria – in press
[3] The University of Economics, Prague, URL:
http://www.vse.cz/kategorie.php3?IDkat=722,
VI. CONCLUSIONS http://distancni.fph.vse.cz/ [online], [2008-2-4].
[4] VSB TUO, Technical University of Ostrava, URL:
Having found the break-even point, the number of students http://moodle.vsb.cz/moodle/206, [online], [2008-2-2].
[5] Project UIS, URL: http://www.uhk.cz/fim/studium/1188, [online],
is set to create and run a profitable course. The effectiveness [2009-2-2].
of any course can be calculated and compared by using this [6] Project EVENE, URL: http://www.uhk.cz/fim/evene/2102 [online],
simple method. [2008-3-11].
Granted finances were not taken into account in the [7] H. Zlámalová, Úvod do distanního vzd lávání, CODV UP Olomouc,
Andragogika, 2001, ISBN 80-244-0276-9
example, but they are often the principal starting point which [8] Bakešová, M.., Koudelka, F..: Ekonomika distanního vzdlávání
enables the work. (Economic of Distance Education), CODV Olomouc 2001, ISBN 80-
The contribution shows that any profitable and effective 244-0274-2
distance module requires great numbers of participants, [9] M. KADAVOVÁ, Virtual Education Relation to Economic Profit of
Educational Subjects, In E-Activities: Networking the World,
according to foreign researches: at least 300 students to cover Proceedings of the 6th WSEAS International Conference on E-
the costs (including fees and payments for printed and Activities (E-Learning, E-Communities, E-Commerce, E-
multimedia study materials). Correct and general information Management, E-Marketing, E-Governance, Tele-Working / E-
on the provided possibilities is essential for the system ACTIVITIES ´07), Puberto De La Cruz, Tenerife, Canary Islands,
Spain, December 14-16, 2007, s. 384 – 388, ISBN: 978-960-6766-22-
approach to virtual university development. 8, ISSN: 1790-5117
An advantage of distance courses, in contrast with the [10] M. KADAVOVA, A. SLABY, F. MALY, Key Factors Involving the
classic courses, is the fact that average variable costs, but Design of the System of Virtual University, In 7th WSEAS Int. Conf.
above all average fixed costs, are decreasing in the long term. on APPLIED COMPUTER & APPLIED COMPUTATIONAL
SCIENCE (ACACOS '08), Hangzhou, China, April 6-8, 2008,
With increase in the number of participants of the course, Published by WSEAS Press, s.678 – 683, ISBN: 978-960-6766-49-7,
inputs such as the number of teachers do not have to be ISSN: 1790-5117
increased in the long term: it is not necessary to rent new [11] M. KADAVOVA, (M. HEDVICAKOVA) Identifying the Break-Even
classrooms, it is not necessary to print other materials. All Point in Distance Courses, In WSEAS TRANSACTIONS on
ADVANCES in ENGINEERING EDUCATION, Issue 5, Volume 5,
materials are created in an e-learning course that can be easily May 2008, WSEAS Press, s. 282 – 294, ISSN: 1790-1979
updated and students can print current information as needed. [12] G. RUMBLE, The costs and costing of networked learning, JALN
The division of costs and their amount differ in various volume 5, Issue 2 – September 2001, Available at http://www.sloan-
institutions. It is not possible to create and present a static c.org/publications/jaln/v5n2/v5n2_rumble.asp [online], [2008-1-5].
Modern Improvements
in the Digital Logic Laboratory
K. M. Noga, M. Radwanski
Gdynia Maritime University
Morska str. 83
81-225 Gdynia, Poland
[email protected]
[email protected]
Abstract-This paper presents examples of application for implemented in PLDs. Altera offers MAX+plus Baseline and
several models of physical objects controlled by programmable Quartus II packets, Xilinx offers WebPack ISE packet.
logic devices (PLDs) to digital logic laboratory course. During
laboratory sessions students create control systems for models and Besides, both companies offer education starter kits for
realize particular assignments with application of PLDs. With the academic use and release handbooks and guides on their
availability of new electronic components, equipment in websites [2, 3].
laboratory has to change. Exercises based on standard digital Popularity of PLDs forced changes in digital logic teaching
integrated circuits made in TTL or CMOS technology were curricula. Last years digital logic course on Electrical
removed from teaching curriculum. Now the purpose of
laboratory sessions is to familiarize students with programming in Engineering Faculty of Gdynia Maritime University (GMU)
VHDL language and software environments designed for PLD. was conducted in form of lectures (30 hours), exercises (30
hours) and laboratory sessions (30 hours) during first and
I. INTRODUCTION
second year of studies. In new schedule another 15 hours of
Necessity of adding PLDs into digital logic education lectures and 15 hours of laboratory sessions intended only for
curriculum appeared together with growing number of PLDs were added. Special starter kits and models made in
implementations and popularity of these devices. Digital Department of Ship Automation and Altera DE2 Development
control of various devices is common today. Control using and Education Board [4] are used during laboratory classes.
PLDs is new technical solution supplementary to used till now Models described in this paper are connected to these kits. In
control systems based on microprocessors and programmable [5] authors depicted first simple models used in digital logic
logic controllers (PLC). In comparison to microprocessor laboratory, newer models are more complex, interactive and
systems, programmable devices contribute simplification of are closer to reality.
some control tasks and technical realization of control system. II. LABORATORY STATIONS AND MODELS
For example, modern PLDs offer lot of universal inputs and
outputs, where propagation time of signals is exactly fixed. A. Basic station with CPLD device
Microcontrollers have limited amount of inputs and outputs
(ports) where computation time is hard to determine, especially
with utilization of high-level programming languages.
Programmable logic controllers are the best solution in control
of mechanical devices in heavy industry.
PLDs find more and more practical applications. Today they
are generally used in data transmission as coders and decoders
of transmitted signals. The new solution is implementation of
PLDs in control of power electronics devices such as voltage
converters, inverters and rectifiers. Control of these devices
requires realization of specific functions in fixed time.
Research and development concerning this implementation are
conducted in many universities and research centers [1].
Altera and Xilinx corporations are the biggest manufacturers
of PLDs. The most important kinds of existing PLDs are:
Complex Programmable Logic Devices (CPLD) and Field Fig. 1. Basic CPLD starter kit
Programmable Logic Arrays (FPGA). Both companies offer
wide range of mentioned kinds of devices, and regularly The first station allowing running of programmable devices
introduce to market newer models with improved capabilities. available in laboratory of digital logic at Department of Ship
The manufacturers of PLDs offer also free of charge basic Automation was laboratory kit ZL1PLD from BTC. ZL1PLD
versions of computer software intended to develop projects kit was equipped with CPLD devices EPM7128 from Altera
109
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_19, © Springer Science+Business Media B.V. 2010
110 NOGA AND RADWANSKI
and XC95108 from Xilinx. There aren’t universal inputs and of control electronics are placed. Near the printed circuit board
outputs available on board, so connecting external circuits and is placed IDC-40PIN socket for communication with DE2
devices to this kit is impossible. ZL1PLD kit has more board. Entire model is shown on Fig. 4.
technical flaws, like unavailable clock pins of CPLD. By these
reasons new extended kit (Fig. 1) with Altera EPM7128 device
was designed. This kit utilizes almost all capabilities of CPLD
device.
B. Robotic arm
Mechanical construction of robotic arm (Fig. 2) is based on
human anatomy, it contains three pivotal joints [6]. One of
them is placed in the mount and allows full circle rotation,
second allows vertical deviation at the base, third one allows
bending at half length of the arm. At the end of the arm is
placed mechanical grab, which can lift load weight up to 1 kg.
Control circuit of the robotic arm consists of three units: Altera
DE2 board, Atmega16-16PU microcontroller and four DC
bridge controllers.
• open the door and switch on light in hall when movement Eleven signals connected to CPLD are used to control this
detector is activated, close the door and switch of light 6 model. Four of them power the engines and set directions of
seconds after last movement, wheels rotation. Next four signals are wired with rotation
• activate alarm when opened windows reach extreme sensors and three last are connected to barrier sensors. Engines
positions. can be turned on and of at any moment with change of
The purpose of exercises is to acquaint students with Altera direction without pauses. Rechargeable battery Sunnyway
DE2 board, Quartus II environment and VHDL. SW613 (6V, 1.3Ah) placed behind control unit allows to
D. Model of Mobile platform operate model up to one-hour period. Data send to PC are used
Mobile platform laboratory station (Fig. 5) consists of to draw movement trajectory with special software. Connecting
vehicle model (1) with control circuit, radio receiver with characteristic direction-changing points formats trajectory plot.
RS232 interface (3) and ZL11PRG (2) PLD programmer. The Position of the vehicle is transmitted to PC in fixed moments in
vehicle communicates with PC through radio receiver, RS232 time or when barrier is detected. Data obtained from radio
interface and sends position data to application developed in receiver are decoded and stored by software in movement
Borland Delphi environment [8]. The model is based on Xilinx history file. Trajectory is stored in bitmap file at the maximum
CPLD programmed with utilization of WebPack ISE resolution of 600 by 600 pixels. Data are transmitted with 2400
environment. bps bitrate in frames with 8 data bits and 1 stop bit.
Sample project
Student should create program that allows to move the
model along fixed trajectory. Current position has to be
recognized by signals in feedback from rotation sensors.
Turning points should be transmitted by radio to vehicle and
tracked position data should be stored in the file and visible on
the screen. Student during this exercise becomes familiar with
use of radio data modems for communication between PC and
movable robot.
E. Multipurpose vehicle
Multipurpose remote controlled vehicle with AVR
microcontroller and CPLD was built in Department of Ship
Fig. 5. Complete laboratory station
Automation too [9]. Additionally vehicle can be controlled by
Schematic diagram of the model is shown on Fig. 6. radio with remote desktop (Fig. 7). Vehicle can run on almost
Underbody structure is made from acrylic glass with three- any surface and may be used to penetrate hard to access or
wheel motion system. Control electronics and battery pack is unsafe places. Load capacity about 5 kg and moderate size (36
placed on top of the platform. Three barrier sensors are placed x 23 x 8 cm) of the robot makes it perfect platform for another
on front bar. Rear wheel isn’t driven; it works as bearing point control and measuring devices. The main control circuit
only. Separate autonomous servomotors drive two front consists of EPM7128 CPLD from Altera and ATmega8-16PV
wheels, than platform can move in every direction. Distance is microcontroller from Atmel. CPLD controls engines, proximity
taken using wheel rotation. Rotation sensors placed on wheel sensors and indicators, microcontroller controls
axes are made from two optocouplers. communication through radio data modem (Fig. 8).
Left barrier
Left drived sensor
wheel
Left rotation
sensor
Middle barrier
sensor
Right rotation
Free wheel sensor
Right drived
wheel Right barrier
sensor
Fig. 7. Vehicle with remote desktop controller
Fig. 6 Diagram of mobile platform
112 NOGA AND RADWANSKI
MICROCONTROLLER
Temp
MODULE +5V POWER
ADC0 PC0
ATmega8-16PV MODULE
D0-D7
+5V
PIN4/RX
Vehicle motion can be controlled by portable desktop too. Control circuit of the robot contains ATMega8 and ATtiny
Desktop is based on ATmega8 microcontroller working with 2313 microcontrollers and EPM7128SLC84 CPLD. Robot’s
ATC-863 radio modem, LCD and keyboard. ATC-863 radio outfit contains sound, light, and distance sensors, battery
modem includes RS-232 and RS-485 interfaces allowing direct charge indicator, buzzer and 2x16 LCD (Fig. 10).
connection with other standard electronics devices containing
UART (Universal Asynchronous Receiver and Transmitter)
port.
Sample project
The vehicle should find a gap between barriers and scour
room behind in autonomic mode. The task for student is to
create program for CPLD, which uses proximity sensors to
determine location of free space and allows sending pictures
from camera. The pictures should be taken in hidden room
with use of vehicle rotation to show the room in various
directions.
F. Quadruped Robot
Quadruped Robot is another model designed and build in
Department of Ship Automation (Fig. 9). Each leg contains
two joints driven by separate servomotors [10]. Data Fig. 10. Block diagram of control modules
concerning angle of flexion are transmitted in control line of Eight digital servomotors HX12K from HexTronik are used,
the servomotor from ATmega8 microcontroller. Two kinds of two in each leg. One moves leg in horizontal plane, second in
steps for boost of motion capability of robot were designed. In vertical. Servo controller uses simple 3-bytes communication
first one robot lay between steps on the bottom side of body, in protocol. Data are transmitted at standard UART bitrate 9600
the second one robot stands on legs during entire walking bps. Maximum resolution of flexure depends on minimal
cycle. length of pulse on output of ATtiny2313 microcontroller and
equals 7.5°.
Motion control system works in three modes. In the default
one running after power-on movements are autonomous
MODERN IMPROVEMENTS IN THE DIGITAL LOGIC LABORATORY 113
according to the program implemented in programmable located inside the mockup. Gantry crane model contains
device. In the remote mode robot’s movements are controlled running tracks, the tower, jib, jib ties, trolley with hoisting
by user’s orders given with utilization of PC software interface. system and hoisting block with electromagnet. Model is
Third one is service mode allowing setting initial position of powered by three electric motors. DC motor moves the tower
servomotors. on running tracks, bipolar stepper motor placed inside the jib
Robot indicates it’s own status with buzzer, LEDs and LCD. moves the trolley, another servomotor inside trolley lifts and
Status signals are generated by ATtiny2313 microcontroller lovers hoisting system.
programmed with Bascom AVR environment. Radio modem Limitary sensors turn off motors when moving part reaches
CC1000PP modules are used for communication between extreme position and send alert signal to FPGA device. Stepper
robot and PC. Communication is based on FSK (Frequency motor is driven by controller based on ATtiny2313
Shift Keying) modulation, data are coded using Manchester microcontroller programmed in C language. Optical sensors
code. Communication protocol allows bi-directional data detect positions of tower on running tracks and trolley on the
transmission in half-duplex mode with error suppression at jib and send sequence of pulses to FPGA device. Using this
bitrate 9600 bps. data it is possible to define current position of movable parts
and to generate signals controlling next moves.
Sample exercise This model is based on the blueprints of Baltic Container
Using VHDL and MAX+plus II or Quartus II environment Terminal in Gdynia [12] and is made in 1:140 scale. Basic
student should prepare program describing sequence of construction of the model has measurements 165 x 145 x 530
movements with status indication. Buzzer, LEDs and LCD mm, jib is 880 mm long.
should be used to display robot state, like distance to obstacle,
intensity of light or low battery level. Sample exercise
The most typical student task is to program sequence of
G. Gantry crane
actions allowing loading and unloading containers from the
Cargo loading is one of the most important issues in
ship. Status of the model should be displayed on 7-segment
maritime transport. Laboratory station with gantry crane model
displays and LEDs included on DE2 board.
(Fig. 11) allows to familiarize students not only with FPGA
programming but also with logistic issues. Gantry crane model H. Bank vault alarm system
is controlled by Altera DE2 board with Cyclone II FPGA To fulfill requirements of didactic occupations interesting
device. model of bank vault alarm system was built too. Model
controlled by DE2 board equipage consists of professional
sensors, GSM cell phone, door opening drive and alarm system
(Fig. 12).
[13]. The door is opened and closed by stepper motor with the pen. EPM7128 device and basic CPLD starter kit fulfill
smooth regulation of speed. role of the controller (Fig. 14). Students can draw simple
graphics (letters or logos) containing segments of straight line,
Sample exercise there is no possibility of drawing angles. Resolution of the
Using Quartus II environment and VHDL the designer drawing equals to 16 x 16 points. Control circuit can be entered
should develop a project that supports vault control system, as logical diagram with standard TTL 74xxx devices or as text
e.g. allows access after entering proper code from the file written in VHDL.
keyboard. In case of double error in code system should III. CONCLUSIONS
generate alarm signals and send alert SMS through cellular
network to selected phone number. Models and boards with programmable logic devices
presented above vary standard laboratory sessions and allow
I. Control of stepper motors watching directly results of control process. Flexibility of
Stepper motors are in common use in automatics. The PLDs gives us possibilities to build various kinds of models.
method of control of stepper motors forces use of digital logic Knowledge about control of DC, AC and stepper motors, using
in practice. Laboratory station (Fig. 13) contains basic CPLD signals generated by sensors and communication between
kit and two separate motors. Use of programmable logic allows devices is required in professional career of electrical engineer.
control in all work modes of motors (full step, half step, Diversity of laboratory stations allows accurate assessment
microstepping, one-phase, two-phase and multi-phase). of student’s work and knowledge from the wide range of
engineering. Students have to be familiar not only with digital
logic, but must know basics of mechanics, electrotechnics,
electronics, data transmission and signal processing.
Developing interactive project of control system for complex
laboratory station is effortful, but students receive guides
containing several examples of projects. Hard-working and
ambitious persons have a chance to prove their abilities by
presenting complex project with optional solutions.
Fig. 13. CPLD kit with stepper motors
REFERENCES
Students can create projects of control systems that allow
[1] M. Baszynski, “A model of the three-phase bridge rectifier with
generating sequence steps in selected direction or generating sinusoidal source current using FPGA implementation”, Przeglad
fixed number of rotations. Source of the project can be entered Elektrotechniczny, vol. 3/2009, pp. 36-41.
as logical diagram or written in VHDL in MAX+plus II or [2] Altera Corporation website: http://www.altera.com.
[3] Xilinx Inc. website: http://www.xilinx.com.
Quartus II environments. This station is used during basic and [4] Altera Corporation, DE2 Development and Education Board User
advanced digital logic course. Manual, electronic version, 2006.
[5] K. M. Noga and M. Radwaski, “Our Stations for Teaching
J. Control of the plotter Programmable Devices”, Innovations in E-learning, Instruction
Technology, Assessment and Engineering Education, pp. 268-273,
Springer 2008.
[6] D. Kupiec amd S. Murawski, Budowa modelu ramienia robota oraz
implementacja algorytmu sterowania w srodowisku Quartus, 1st degree
thesis, GMU, Department of Ship Automation, 2008, in Polish.
[7] L. Meissinger and S. Krawczyk, Zastosowanie ukladow
programowalnych do sterowania modelem domu mieszkalnego – budowa
i oprogramowanie stanowiska laboratoryjnego, 1st degree thesis, GMU,
Department of Ship Automation, 2008, in Polish.
[8] P. Ptach and, S. Sieranski, Sterowanie wybranym modelem po
okreslonej trajektorii z wykorzystaniem ukladow programowalnych, MA
thesis, GMU, Department of Ship Automation, 2008, in Polish.
[9] K. Gorski and S. Orzechowski, Projekt i budowa wielozadaniowego
pojazdu samobieznego sterowanego mikrokontrolerem i ukladami
programowalnymi, 1st degree thesis, GMU, Department of Ship
Automation, 2009, in Polish.
Fig. 14. Laboratory station with plotter. [10] P. Wozniak and M. Wysocki, Projekt oraz budowa kroczacego robota
sterowanego ukladami programowalnymi, MA thesis, GMU, Department
Plotter is one of the oldest devices purposed to print of Ship Automation, 2009, in Polish.
[11] M. Witek and P. Witek, Projekt, budowa oraz sterowanie modelem bazy
computer graphics. Simple method of control requires only kontenerowej z zastosowaniem ukladow programowalnych, 1st degree
basic knowledge from the range of digital logic. Printing task thesis, GMU, Department of Ship Automation, 2008, in Polish.
is based on sending sequence of two 4-bit numbers to the [12] Baltic Container Terminal in Gdynia website: http://en.bct.gdynia.pl/
[13] W. Laskowski and T. Taczynski, Model systemu alarmowego skarbca z
plotter. These numbers determine position of the pen on X-Y wykorzystaniem zestawu edukacyjnego Altera DE-2 i jezyka VHDL, MA
plane, one additional bit represents writing or high position of thesis, GMU, Department of Ship Automation, 2009, in Polish.
e-Cheating and Calculator-Technology
A Preliminary Study into Casual Implications of Calculator-Technology Usage on
Students’ Attitude Towards e-Cheating
Abstract daily life, yet a technology that has increased student ability
in all science fields, in and out of classrooms, and into offices.
Across the globe, educational institutions are adopting e-learning
tools into their curricula to cater to the ever-growing demand for This paper has been written as a subsequent resultant of an
technology inside the classrooms. Where traditional blackboards
on-going research that is being carried out by the author on
have been replaced by virtual whiteboards and library books by
the casual implications of readily-available technologies on
online resources, teachers are trying hard to cope with the growing
competition from technology outside the classroom that has students’ attitude towards e-cheating. At the start of the study,
captured their students’ attention and imagination, with more it looked as though enough research had been conducted on
technology, perhaps increasing student dependency. various aspects of e-cheating. However, as the literature
review has highlighted, distinct gaps seem to exist in the
However, the author’s previous research has shown that engaging literature that highlight the absence of studies showing the
e-learning tools to enhance student learning experience is not impact of certain technologies besides the Internet on
enough as it has negative impact on students’ attitude towards e-
students’ attitude towards e-cheating [8]. One such
cheating, disillusioning their awareness to cyber ethics. So, how do
technology is that of calculators that aid in computations and
teachers ensure that the methods of teaching and the tools
therewith truly enhance student learning and not affect their sense estimations.
of ethics?
This paper looks at the ease with which students can access
As a part of an on-going research, this paper highlights the calculators and how the usage has affected their attitude
exponential usage of calculators in the classrooms and their towards e-cheating.
impact on student’ attitude towards e-cheating, and recommends
further studies to measure the correlation of calculator II. CALCULATORS NOW AND THEN
dependency in the classroom to student attitude towards e-cheating
and professionalism at workplaces. . The first instrument to help making calculations easy and fast
was the abacus. This has wooden beds that are moved along a
I. INTRODUCTION wire or thin wooden stick to represent numbers [9]. The
young French mathematician Blaise Pascal invented the first
Technology in education is not a new phenomenon. It has
adding machine in 1642, a device driven by gears and capable
been around for decades and has become a vital part of
of performing mechanical addition and subtraction [1].
teaching and learning in the twenty-first century. Inside and
out of classrooms, technology usage has gained popularity to A revolution in calculating machines took place between the
feverish heights. Almost any and every technology is now early 1960s and the late 1970s. It was during this time period
used as a part of e-learning in order to enhance the overall that electronics for calculators was at the forefront of
experience for students and teachers [6]. Calculators are one electronics research. ‘Calculators evolved from large,
such technology that users have taken for granted as a part of expensive, mechanical machines to cheap, electronic, credit
115
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_20, © Springer Science+Business Media B.V. 2010
116 KHAN
card sized devices. The development of micro-electronics for in the process of researching for this paper, the author has
calculators was an important phase in the history of found many websites and other sources that seem to entice
technology, which included the development of the students into using calculators in the exam halls, for their
microprocessor.’ [2] term papers and homework in order to achieve higher grades.
Sites such as Online 2 College and Teachapolis that is
III. CALCULATORS – BOON OR BANE supposed to be a ‘virtual city for teachers’ are free to visit by
any user sitting on a PC connected to the Internet. With such
Calculators are very much a part of technology that have
topics as ‘Using Technology to Cheat • Calculator:
upgraded over the last decades to become part of e-learning.
programmable calculators can hold text, formulas, even
As mentioned in the previous study, calculators have been a
pictures’ [3] it is giving students hints on how to use the
breakthrough way before computers had become common
calculators in something other than simply aiding in quick
place at every home [8]. It was definitely a technology above
calculations. Another article by Andrew Kantor in
the use of booklets with pre-calculated tables and slide rulers
USAToday.com relents about ‘add-on memory [that] lets you
[10] As highlighted before, in 1965, the first pocket calculator
store software, turning your calculator into a pocket
was introduced to the market; by 1974 it had achieved
notebook… the company [Texas Instruments] is happy to
providing four functions with LED screen, and although it
point out that the TI-83 Plus is "allowed for use on the PSAT,
was well over a decade before school children had their own
SAT I, SAT II Math IC and IIC, AP Chemistry exam, AP
pocket calculators, the technology advanced rapidly [10]
Physics exam, and AP Calculus exam.’ [4]. In 2005, Texas
Now-a-days, calculators range from simple scientific to
Instruments (a company that pioneered in the development of
graphic to programmable with large amounts of memory
calculator technology in the 1970s) recalled thousands of
space, data wires to allow sharing of information, formulas
calculators issued to students in the United States of America,
and so on among friends.
after a student discovered that pressing certain keys converts
However, studies have shown that where calculators make it decimals into fractions [5]. As schools expected students to
easy for students and adults to make quick calculations, they have the knowledge and skill to perform this task, the
are ‘becoming a mental crutch for students, rather than a tool calculators being used for the same purpose was labeled as
that promotes higher order learning’ [11]. It has been seen cheating.
that most often than not academic institutions ban the use of
Research into existing literature has shown that calculator
certain types of calculators in examinations to ensure students
technology has definitely had an impact on students’ attitude
are able to work out problems upon their ability rather than
towards e-cheating. However, the author has found little or no
with the aid of technology. Such technology in the classroom
evidence of actual studies carried out to establish a casual
is feared to ‘result in an over-reliance on technology to
relationship between the two.
provide solutions, thereby stifling a student’s educational and
creative growth’ [11]. The author has found little or no V. THE STUDY
literature to show the actual effects of allowing calculators in
the classroom. In many schools and universities, teachers are Based on the literature reviewed, and the corresponding study
on high alert in examination halls, keeping an eye out for being carried out on readily-available technologies and their
programmable calculators that students can bring in with effects on students’ attitude towards e-cheating, the following
uploaded formulas and pre-sketched graphs that would study was proposed: investigate if there are any casual
constitute cheating. But, there does not seem to be any study implications of allowing calculator-technology in the
to actually register if there is any casual implications of classroom to students’ attitude towards e-cheating.
allowing high-end calculators, or any computational devices
that are affordable and readily available in stationary shops, To carry out this study, the sample size chosen consisted of
on students’ attitude towards cheating [11]. about 100 students of mixed ethnic and educational
(curricula) backgrounds. Of these, 28 were discarded as they
IV. IS IT AFFECTING STUDENTS’ ATTITUDE TOWARDS E- had not completed the surveys. A survey model was
CHEATING? developed to gather student feedback on various topics related
to academic integrity and specifically to the usage of
Although the author has found no direct studies showing
calculator and other technology. Students interviewed were
affects of calculators on students’ attitude towards e-cheating;
E-CHEATING AND CALCULATOR-TECHNOLOGY 117
4000
Series2
3000
2007
Series1
2000
2006
1000
0
s
t
e
ry
ts
h
cs
s
ec
lu
nc
at
at
tic
ist
un
i
cu
bj
ys
tM
M
na
is
em
co
Su
Ph
at
Ca
sic
Fi
ee
Ac
St
Ch
Ba
cr s
Di
Subjects
[5] Block, R. (2005). Texas Instrument calculators help students cheat. For instance, for the question ‘It is okay to download MP3 or movies from
(Online). Available URL: http://www.engadget.com/2005/06/09/texas- peer-to-peer websites’, the weights start from 1 to 5 for each of the Likert
instrument-calculators-help-students-cheat/ items ‘strongly agree’, ‘agree’, neither agree nor disagree’, ‘disagree’,
‘strongly disagree’ respectively. This is because if a respondent chooses
[6] Khan, Z. R. (2006). E-learning techniques in IT programs – are they ‘strongly agree’, he/she demonstrates he/she has chosen a wrong answer. For
helping my sense of cyber ethics?. Conference paper presented at the IEEE another question, ‘Cheating can be defined as violating accepted standards or
Second International Joint Conferences on Computer, Information, and rules’, the scale is reversed such that if a respondent chooses ‘strongly agree’,
Systems Sciences, and Engineering. December 4-14. the response receives 5 point on the scale thus showing he/she has chosen the
right answer. Once the weights have been placed on each response, the
[7] Khan, Z., Al Qaimari, G. & Samuel, S. D. 2007, 'Professionalism and average is calculated by totaling response for each question and diving by the
ethics: is education the bridge?', in G. R. Lowry & R. L. Turner (eds), total number of respondents for each question (72 students) in each case.
Information Systems and Technology Education: From the University to the
Workplace, Information Science Reference, New York.
[11] Bsmarte. (2006). Managing Technology: Is technology in the classroom Basic Math 468 401 650
good or bad?. Bsmarte.com [Online] Accessed February 2009. Available
Discreet Math 238 265 300
URL: http://www.bsmarte.com/issues/managingtechnology.pdf
Calculus 210 365 365
APPENDIX A Statistics 200 200 255
Physics 165 165 165
Calculating the weighted average:
Chemistry 100 55 290
In Table 1, for each Likert Item, there is a ‘weight’ placed depending on how Accounts 110 100 300
close the choice is to being right. Finance 100 100 144
Neither
Strongly agree or Strongly
Answer Options agree Agree disagree Disagree disagree
Neither
Strongly Strongly
Answer Options Agree agree or Disagree
agree disagree
disagree
it is okay :
(a) to share information among friends during tests or exams 5 10 16 29 12 249.00
Weighted
Weight 1 2 3 4 5 3.46 Average
(b) to copy from the website that has the required information for an
3 17 18 28 6
assignment 233.00
Weighted
Weight 1 2 3 4 5 3.24 Average
(c) to copy from a text book that has the required information for an
4 21 22 20 5
assignment 217.00
Weighted
Weight 1 2 3 4 5 3.01 Average
(d) to write in the information from what someone else says for an
1 13 19 33 6
assignment 246.00
Weighted
Weight 1 2 3 4 5 3.42 Average
(e) to copy from another friend who has the information for an
2 7 15 37 10
assignment 259.00
Weighted
Weight 1 2 3 4 5 3.60 Average
All of the above points a-e but with due citations and reference list 28 30 7 4 0 289.00
Weighted
Weight 5 4 3 2 1 4.01 Average
It is okay to install a copy write software given to me by a friend 9 25 24 10 3 186.00
Weighted
Weight 1 2 3 4 5 2.58 Average
It is okay to download MP3 or movies from peer-to-peer websites 18 35 9 7 3 158.00
Weighted
Weight 1 2 3 4 5 2.19 Average
It is cool to buy pirated movies from vendors on the streets for AED
13 21 15 17 5
5/- instead of the original for more than AED40/- 193.00
Weighted
Weight 1 2 3 4 5 2.68 Average
‘Anything goes’ is a sure attitude to success 4 9 30 22 4 220.00
Weighted
Weight 1 2 3 4 5 3.06 Average
Abstract-This study set out to examine the online involves constructing the situation and supporting the
mathematical communication and measure the progress of critical reasoning behind a solution. It focuses on the process of
thinking (CT) skills of students that occurred in collaborative
groups, during two collaborative problem solving sessions via learning rather than just attaining information. It involves
online Discussion Forums (DFs), in a first year Engineering discovering how to analyze, synthesize, make judgments, and
Mathematics unit which spans 14 weeks. It was found that the create and apply new knowledge to real-world situations. A
calculated CT scores have improved from the first to the second simple standard definition of critical thinking adopted in this
DF session. The CT scores were connected with moderation study is “the ability to use acquired knowledge in flexible and
techniques associated with Socratic Questioning and reflection
scaffolds, during the first session, but not in the second. Students’ meaningful ways, through understanding the problem or issue,
perceptions on the DFs were analyzed and was found that evaluating evidence, considering multiple perspectives, and
majority of the students found the DF sessions quite challenging taking a position” [3], p. 275.
to their thinking skills and profitable for developing their
problem solving skills. B. Socratic Questioning
Yang, Newby & Bill [4] has used Socratic Questioning to
I. INTRODUCTION promote critical thinking skills in asynchronous discussion
forums. A powerful method of cultivating critical thinking in a
Critical thinking is both a process and an outcome [1]. As an
group, Socratic Questioning uses the technical strategy of
outcome, it is best understood from an individual
regressive abstraction and develops a rigorous inquiry into the
perspective—that is, the acquisition of deep and meaningful
ideas, concepts and values held by people with regard to the
understanding as well as content-specific critical inquiry
topic. The questions fall under the six categories of Socratic
abilities, skills, and dispositions. Judging the quality of critical
questioning prompts [5], namely, questions of clarification,
thinking as an outcome within a specific educational context is
questions that probe assumptions, questions that probe reasons
the responsibility of a teacher as the pedagogical and content
and evidences, questions about viewpoints or perspectives,
expert. As a product, critical thinking is, perhaps, best judged
questions about the question and questions that probe
through individual educational assignments. Critical thinking
implications and consequences. Instead of directly providing
(CT) and mathematical problem solving have existed side by
statements as clues for students to advance in their process of
side and have attracted research over the decades.
solving the problem, questions are posed with a reasoning
The research questions answered through this study, done on
approach, from which the students are led to thinking into the
online Discussion Forums (DFs) for a batch of Engineering
different aspects/stages of the problem.
Mathematics students were: (1) What was the level of CT skills
of the participants, individually and group wise, in the online C. Asynchronous Discussion Forums and CT
DFs? (2) What was the effect the moderation techniques have There are two approaches [6] to teach CT using content
on the level of CT skills in the online discussion forums? (3) disciplines: (1) The “infusion” approach where critical thinking
What were the participants' perceptions about the online DFs skills are taught manifestly using the discipline’s content; and
and the impact of the DFs on their critical thinking skills? (2) The “embedded” approach, where critical thinking skills
are taught in indirect ways without spelling it out to students.
II. REVIEW OF LITERATURE
The asynchronous discussion forums are seen as an attempt to
A. Critical Thinking and problem solving try the embedded approach. Discussion forums were identified
Erwin [2] defines problem solving as a step-by-step process by [7] as one of the effective forms of assessments of critical
of defining the problem, searching for information, and testing thinking. Perceptions of asynchronous communication tools
hypotheses with the understanding that there are a limited have been reported by [8] and [9]. In their study on cross-age
number of solutions. Critical thinking is a broader term peer tutors’ contributions in asynchronous discussion groups,
describing reasoning in an open-ended manner, with an [10] have commented that students’ learning experiences and
unlimited number of solutions. The critical thinking process perceptions about the scaffolding offered on asynchronous
121
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_21, © Springer Science+Business Media B.V. 2010
122 JACOB AND SAM
discussion forums might nourish a prospective qualitative “Modern Engineering Mathematics” [13]. During the course of
design. the discussion, the instructor moderated, or, scaffolded and
III. RESEARCH STUDY DETAILS modeled the process through Socratic questioning.
The proposed model shown in Fig. 1 was based on the
The study involved the first semester Engineering following principles: (1) Learning is a dialogical process in
Mathematics unit of the Bachelor of Engineering programme which communities of practitioners socially negotiate the
offered by the Swinburne University of Technology (Sarawak meaning of phenomena [14],[15],16]; (2) views of knowledge
Campus) during the year 2009. Out of the total number of 115 are constructed by, rather than transmitted to, the learner [17];
students, 60 participants were chosen by voluntary consent. (3) ill-structured problems and Socratic questioning are stimuli
The number of participants was fixed at 60, since the study was [4],[5] for critical thinking; and (4) asynchronous discussion
an in-depth study involving a lot of variables, especially the forum is a collaborative mind tool [17] which contributes to the
analysis of the online forum postings of individual students. A development of critical thinking and effective communication.
similar study on online forums in mathematics done by Kosiak The model integrates the principles of the model presented by
[11] has used around 50 participants. The participants filled in [15] for the analysis of collaborative knowledge building in
the Consent Form agreed by the Swinburne University Ethics asynchronous discussions.
Committee, as an agreement to their participation in the study.
The participants consisted of a mixture of Malaysian (83%)
and international (17%) students. The 10 international students,
out of the total of 60 participants, comprised students from
Indonesia (1), Iran (1), Iraq (1), Kenya (2), Mainland China (2),
Saudi Arabia (2), and Sri Lanka (1). Generally the students of
this unit are predominantly male. The sample reflects this
typical scenario – it was an uneven mixture of males (85%)
and females (15%). The first year university students go
through a transition stage from school into university. Hence
they need a lot of support in terms of coping up with new
learning environments and taking ownership of their learning.
The online mathematical forums were used as a means of
establishing good rapport between the students and the lecturer,
and between the students themselves. All these students were
new to the experience of using discussion forums in the
educational setting. Figure 1. A proposed model for designing an online mathematical discussion
The unit was supported by the online Learning Management forum environment.
System (LMS) called the Blackboard Learning System
(http://www.blackboard.com), or BBLS. The Discussion The model conceived an ill structured problem or, its
Forums were available on the BBLS. Two problem solving problem space as the focus of the discussion. As seen in Fig. 1,
sessions were planned on these Discussion Forums - that is, ill the fundamental unit of the whole activity is the problem space
structured problems were posted at two different times (in [18], [19] constructed from the ill structured problem by the
Week 4, DF1 and in Week 10, DF2) during the 14 week long students. The creation of a problem space involves mapping
course. The goal of these sessions was for students to work the problem statement from the ill structured problem onto
collaboratively to solve and understand the ideas on these ill- prior knowledge and constructing a personal interpretation of
defined problems [12]. The 60 students were divided into 11 the problem. The process of solving the problem would follow
groups of 5 or 6. The first forum carried a weightage of 7 %, the iterative and cyclical process [19] that ill-structured
and the second carried a weightage of 8%. 15% of the course problems follow. The process would expect the participants to
work marks was assigned for the Discussion forums to articulate goals, verify the problem, relate problem goals to
encourage students into active participation. The instructor problem domain, clarify alternative perspectives, generate
encouraged these collaborative sessions by moderating problem solutions, gather evidence to support/reject positions,
(scaffolding) the discussion thread in order to stimulate both construct arguments, implement and monitor solution. In the
mathematical learning and mathematical understanding. In this process, the participants generated clarification questions,
study, during the first three weeks of the unit, when the assessment questions, inferences, strategies (viz., the four
discussion forum sessions have not yet begun, the students phases of critical thinking stated in the model shown in table 2)
were given access to examples of ill structured problems and and took part in active and collaborative learning. The
solutions. The ill structured problems were chosen by the moderator would pose scaffolding questions [12] to clarify key
lecturer, who is also the moderator of the forum, from a mathematical concepts and to advance/direct the discussions
standard book of engineering application problems called towards the solution.
PERSPECTIVES ON CRITICAL THINKING THROUGH ONLINE DISCUSSION FORUMS 123
Data was collected in this study using two modes: (1) The section 3 consisting of one open ended question asking for
Discussion Forum postings; (2) Survey Questionnaires- qualitative comments.
perceptions on discussion forums and critical thinking
questionnaire. TABLE II
TABLE I MODEL FOR IDENTIFYING CRITICAL THINKING IN MATHEMATICAL PROBLEM
SCORING TABLE FOR CT CATEGORIES SOLVING SESSIONS OVER DISCUSSION FORUMS
Clarification
Category No. of postings Weightage
Formulates the problem precisely and clearly.
Clarification 1-2 1
Analyses, Identifies one or Identifies Defines or
3-4 1.5 negotiates or more underlying relationships criticizes the
>4 2 discusses the assumptions in among the definition of
Assessment 1-2 3 scope of the the parts of the different parts relevant terms
3-4 3.5 problem problem of the problem
>4 4 Assessment
Inference 1-2 5 Raises vital questions and problems within the problem.
3-4 5.5 Gathers and Provides or asks Make value
>4 6 assesses relevant for reasons that judgment on the
information. proffered assessment
Strategies 1-2 7 evidence is valid criteria or
3-4 7.5 or relevant. argument or
>4 8 situation.
Inference
An ill structured problem was posted on the discussion Reasons out based on relevant criteria and standards
forum by the instructor, students were briefed in detail about Makes Arrives at well Makes Frames
appropriate thought out generalizations relationships
the sessions and given the deadline of one week to solve. The deductions from conclusions from relevant among the
postings were downloaded by the instructor after the sessions discussed results. different parts
were over. The postings were coded into the four categories, results. of the problem.
using the indicators in the model as guides. In cases where Strategies
Thinks and suggests open mindedly within alternative systems of
more than one critical thinking process appeared within a thought.
posting, only one code was associated, which seemed to be the Propose specific Discuss possible Evaluate Predicts
most important in the context. steps to lead to steps. possible steps. outcomes of
The model for content analysis of the postings in the current the solution. proposed
steps.
study was adapted from two models- the model proposed and
tested by Perkins and Murphy [20], and the framework for
assessing critical thinking developed by Paul and Elder [21]. IV. RESEARCH STUDY RESULTS
The new model is shown in Table II [22], with the indicators
The first Discussion Forum, DF1 occurred during Week 4,
and description to each phase/category. The higher levels of
and the second Discussion Forum, DF2 occurred during Week
critical thinking are associated with the stages of Inference and
10 of the 14-week mathematics unit. During the course of the
Strategies.
one week participation, the total number of posted messages in
Also a weighted critical thinking score (CT1 and CT2) was
the discussion forum sessions- namely DF1 and DF2 was 616
associated with every student for DF1 and DF2 respectively,
and 539 respectively, as it occurred in the 11 groups. Fig. 2
based on the classification of the postings, as shown in Table I.
shows the mean number of postings across the 11 groups. As
The score has been developed for the purpose of quantitative
seen from the Fig. 2, participation varied among groups and
analysis based on the CT model. The table was developed with
individuals. The descriptive statistics in table III are compiled
the following principles: (1) The limit of the number of
from DF1 and DF2, hence each variable carries a ‘1’ or ‘2’ in
postings has been set, assuming the average number of
its name, to denote it comes from DF1 or DF2. The arithmetic
postings of a student, per category, for one discussion forum
means in table 3 show that the number of postings was more in
was around three. (2) Higher weights are associated with
the lower phases of critical thinking, namely clarification and
Inference and Strategies, to indicate the higher levels of critical
assessment, than in the higher phases. But the standard
thinking. (3) The maximum weightage of one category differs
deviation (S.D.) and the quartiles (Q1, Q2 and Q3) reveal
by one from the minimum weightage of the next category.
substantial variability in the levels of critical thinking of the
Hence the maximum score possible is 20, if a student has more
students. The quartiles give a better picture of the distribution
than four postings in all categories.
of the variables, since the distribution was not normal.
The perceptions of students were measured using the Critical
Research Question 1a: What is the level of CT of
Thinking and Discussion Forum Questionnaire (CTDFQ). The
individuals in the online discussion forums? Shapiro-Wilk’s
questionnaire had three sections – section 1 on CT, consisting
test for normality done on the variables – the number of
of 4 questions, section 2 on DFs, consisting of 9 questions, and
124 JACOB AND SAM
postings in the four CT categories, and the CT model scores results showed that CT skills measured in the mathematics
(CT1 and CT2), showed that the data values are not normally problem solving sessions were very varied. But, generally CT
distributed (p < 0.05) for all these variables. But the parametric skills were exhibited in the lower phases of critical thinking,
calculations and tests were valid on the data set, since the throughout the two discussion forums. A Pearson correlation
sample size of 60 is large (n >30) enough for the normal proved significant between CT1 and CT2 ((Pearson correlation
assumption of distribution of the data. = 0.644, p < 0.01) at 1% significance level. Or, the CT model
The distribution of the data for the CT categories and the CT scores for the students move linearly between DF1 and DF2.
model scores looked very widely distributed among the batch As students progressed from DF1 to DF2, the students’ CT
of students. One fourth of the students posted no postings in skills exhibited in the problem solving sessions over the DF
the higher phases of CT – namely, inference and strategies, but also showed an improvement.
another one fourth posted more than 4 postings.
Figure 2. Line chart showing the mean total number of postings across the
groups in DF1 and DF2.
TABLE III
DESCRIPTIVE STATISTICS OF THE 60 STUDENTS
Category Mean S.D. Q1 Q2 Q3
found for the phases, namely clarification (F = 2.743, p < 0.01) DF1 to DF2. The students have positively agreed to the fact
and assessment (F = 2.593, p < 0.01) in DF1 and only for that their attitude to mathematics has been improved as a result
assessment (F = 2.173, p < 0.05) in DF2. A one-way ANOVA of the DFs. The students have disagreed to the comment that
done for the CT model scores, CT1 (F = 1.647) and CT2 (F = moderation did not play an important role.
0.975) between the different groups showed the differences TABLE IV
were not significant at 5% significance level. SAMPLE COMMENTS OF STUDENTS
Research Question 2: What is the effect of moderation Positive comments Negative comments
techniques on the level of CT of individuals in the online
discussion forums? MANOVA was employed to test the Questions connected with daily
Unsystematic posting of threads
life applications boosted my
effects of moderation (scaffold) on the participant’s phases of appreciation towards mathematics.
causes discomfort.
CT and the CT model scores, in DF1 and DF2. The moderation
types were: Type 1- Question scaffolds (Socratic questioning) I get to see different opinions over All group members do not think
only and Type 2- Question scaffolds (Socratic questioning) one problem. Thus I find out more and post at the same pace; So
than one way to approach the some group members might be
plus reflection scaffolds.6 out of the 11 collaborative groups, given problem. discouraged to post any further.
had the type 1 scaffold, and the rest of the groups had the type
2 scaffold. Thus the moderation type, which is a combination It is an easier way to get marks for
Not all members are active.
of Socratic questioning and reflection scaffolds, produced Besides we sometimes could not
the course work.
care less about the Disc. Forums.
significant differences in student postings in the first forum
DF1. The differences were noticed in the phases of It’s good to have it online; Saves No active participation from the
clarification, assessment and strategies. But in DF2, students time/hassle free. group members.
seemed to have become more independent that the postings
didn’t vary much between the two groups. Or, moderation We get idea from others, even if I
But two Discussion Forum
seemed to have less effect as time progressed and students don’t know how to solve the
sessions is too much.
problem.
were used to the pattern of discussions over the DFs.
Research Question 3: What are the student perceptions You explore and think outside the Lack of participation from group
on CT and DFs as they experienced during the 14 week box. It gives a wider picture than members is like a cold blanket,
course? Section 1, covering CT, consists of 3 questions rated what you see and hear in the and disables us from exchanging
classroom. views.
on a 4-point Likert scale: Strongly Agree (1) to Strongly
Disagree (4). The 4-point scale was chosen in favour of the
The questions are great and help
usual 5-point scale since the researcher does not believe in engineering students to see how
I am not a fan of team work for
“neutral” people [23] walking about with no positive or problem solving. I prefer to work
simple mathematics could be used
individually than in a team.
negative response towards the issue in hand. There was a in applications.
fourth question asking for the student’s self evaluation of
his/her own CT rated on a scale of Excellent (4) to Poor (1). It makes me excited, and I will My group members did not co-
not give up without getting the operate well. It looks like I am
Student perceptions have favoured the fact that CT was answers; I like critical thinking. discussing with the lecturer.
understood as a result of doing the particular mathematics unit.
The time frame is too short. I need
Group work was also seen as helping CT skills. The argument It increases my group work skills,
time for solving any question
critical thinking skills. Can share
that CT skills became better from DF1 to DF2 showed a mean knowledge with friends.
because the main problem with
closer to 3. Majority of the students were echoing the fact that me is how to begin.
their CT skills (as they understand) became better from Forum
1 to Forum 2. It is noticed from the results of research question But they have agreed that the moderator modeled by posing
1 that the CT model scores also reflected the same result- there appropriate discussions. Nearly 40% of the students sounded
was seen a general progress in CT skills from DF1 to DF2.The that they participated simply because of the course work marks
students’ rating of their own CT was an average (the mean allotted for participation in the DF sessions, and that
value is close to 2) on a scale of 1-4. Majority of the students participation was a demanding task. This was reflected in some
saw themselves as belonging to the class of average critical of the negative comments to the open questions listed later. But
thinkers. the majority of students have participated with a more positive
Section 2 of the CTDFQ dealt with the DFs and the problem mind, and enjoyed the online sessions which made them think
solving sessions. There were 9 questions in this section – 5 out of the box. The rating of the DF sessions on the Black
positively worded statements, 3 negatively worded statements Board was also a pleasant mixture of ‘Good’ and ‘Average’.
and 1 question to rate the DF problem solving sessions But the mean score pointed to the fact that the sessions were
activated on the Blackboard Learning System. Majority of the rated good.
responses were positively favouring the DF sessions as an Section 3 consisted of an open ended question asking for
excellent way to promote CT. They did agree to the fact that positive and negative comments on the experience of students
the discussion and pace of the DF sessions became better from with the DFs.
126 JACOB AND SAM
Table IV shows a set of positive and negative comments [5] Y-T.C Yang and H-A. Chou, “Beyond critical thinking skills:
Investigating the relationship between critical thinking skills and
made by a few students. It was clear from the responses that dispositions through different online instructional strategies,” British
these students have never been exposed to such problem Journal of Educational Technology, vol. 39 (4), pp. 666-684, 2008.
solving sessions in mathematics. Team work skills, innovation, [6] H. Innabi, H. and O.E. Sheikh, “The Change In Mathematics Teachers’
Perceptions Of Critical Thinking After 15 Years Of Educational Reform
sharing with friends, and the fun of being free to think beyond In Jordan,” Educational Studies in Mathematics, vol. 64, pp. 45 – 68,
limitations were seen as the positive sides of the DF sessions. 2006.
But, being first semester students who were new to the [7] G. E. Perry, “Critical Thinking and Web-Based Education in Maryland
Community Colleges: How the Medium Promotes Development”.
university environment, some of them felt that the sessions Unpublished doctoral dissertation, University of West Virginia, in Bers,
required a lot from them on top of their packed schedules. Also T. H., Assessing Critical Thinking in Community Colleges, New
lack of response from their team mates, preference for Directions For Community Colleges, vol. 130, Wiley periodicals, Inc.,
pp. 15-25, 2005.
individual work, short time frame were the reasons for the [8] M. Guzdial, P. Ludovice, M. Realff, T. Morley, K. Caroll, and A. Ladak,
negativity of the students. The challenge of collaborative learning in engineering and math.,
Frontiers in Education Conference, 2001, 31st Annual, pp. T3B-24-29,
V. CONCLUSIONS 2001. Retrieved January 2008 from http://coweb.cc.gatech.edu/cs1.
[9] T. Karsenti, “Teacher Education and Technology: Strengths and
The idea of measuring critical thinking (CT) skills of students Weaknesses of Two Communication Tools,” Proceedings of the 2007
through the online Discussion Forums, proved to be very useful to Computer Science and IT Education Conference, 2007. Retrieved
the instructor, especially in the context of a large class (more than January 2008 from http://csited.org/2007/83KarsCSITEd.pdf.
[10] M.D. Smet, H.V. Keer, and M. Valcke, “Cross-age peer tutors in
50 students), where it is difficult to assess the thinking skills, asynchronous discussion groups: A study of the evolution in tutor
personally or one to one. The instructor could classify the support,” Instructional Science, 37(1), pp. 87-105, 2009.
students’ postings based on their thinking skills and thus give [11] J. J. Kosiak, “Using Asynchronous Discussions To Facilitate
Collaborative Problem Solving In College Algebra”, Doctoral
feedback to the students. In general, the majority of the postings Dissertation, Montana State University, 2004.
fell into the lower categories of CT, as was the case noticed by [12] M-S. Chiu, “Approaches to the teaching of creative and non-creative
Lee [1] and Kosiak [11]. But the CT skills have improved from mathematical problems,” International Journal of Science and
mathematics education, vol. 7(1), pp. 55-79, 2009.
the first session to the second, as shown by the CT model scores. [13] G. James, Advanced Engineering Mathematics (3rd Edition). Pearson
But there were students who were ahead of the others, considered Prentice Hall, 2004.
as outliers. The moderation process was very much necessary in [14] M. Mabrito, “A Study of Synchronous Versus Asynchronous
Collaboration in an Online Business Writing Class,” The American
the beginning stages, but during the second session, students Journal of Distance Education, vol. 20(2), pp. 93-107, 2006.
seemed to be confident to move on by asking, assessing and [15] S. Schrire, "Knowledge building in asynchronous discussion groups:
discussing among the group members. The students perceived the going beyond quantitative analysis", Computers & Education, vol. 46(1),
pp. 49 – 70, 2006.
forums as a boon to work on their CT, problem solving skills and [16] L. S. Vygotsky, Mind in society: The development of higher
majority of the students appreciated the innovation and challenge physiological processes. Cambridge, MA: Harvard University Press.
offered by the DF sessions. Students also gave the researchers [17] S. Schrire, “Interaction and cognition in asynchronous computer
conferencing,” Instructional Science, vol. 32, pp. 475-502, 2004.
enough feedback to work on. [18] A. E. Campbell, V. M. Adams, and G. E. Davis, “Cognitive Demands
and Second-Language Learners: A Framework for Analyzing
REFERENCES Mathematics Instructional Contexts,” Mathematical Thinking & Learning,
[1] K.S. Lee, “Effects of Individual versus Online Collaborative Case Study vol. 9(1), pp. 3-30, 2007.
Learning Strategies on Critical Thinking of Undergraduate Students. [19] S. Sendag, and H.F.Odabasi, “Effects of an online problem based
Doctoral Dissertation,” The University of Texas at Austin, 2004. learning course on content knowledge acquisition and critical thinking
[2] T. D. Erwin, “Definitions and assessment methods for critical skills,” Computers & Education, vol. 53, pp. 132-141.
thinking,problem solving, and writing,” in T. Bers, Assessing critical [20] C. Perkins and E. Murphy, “Identifying and measuring individual
thinking in community colleges, New Directions for Community Colleges, engagement in critical thinking in online discussions: An exploratory
vol. 130, pp. 15-25, 2005. case study,” Educational Technology and Society, vol. 9(1), pp. 298-307,
[3] S. VanderStoep and P. Pintrich, Learning to learn: The skill and will of 2006.
college success. Upper Saddle River, NJ: Prentice Hall, 2003. [21] R. Paul and L. Elder, Critical Thinking: Learn the Tools the Best
[4] Y-T.C.Yang, T.J. Newby, and R.L. Bill, “Using Socratic Questioning to Thinkers Use. Prentice Hall, 2006.
promote critical thinking skills in asynchronous discussion forums in [22] S. M. Jacob and H.K. Sam. “Effect of Online Discussion Forums to
distance learning environments,” American Journal of Distance Improve Mathematics Problem Solving and Critical Thinking – A
Education, vol. 19(3), pp. 163-181, 2005. Comparative Study,” Proceedings of the 5th International Conference on
Innovations in Information Technology (Innovations 2008), United Arab
Emirates, ISBN 978-1-4244-3397-1, pp. 366-370, 2008.
[23] M.G. Lodico, D.T. Spaulding, and K. H. Voegtle. Methods in
Educational Research: From Theory to Practice [E-Book], 2006.
Digital System Description Knowledge Assessment
K. Jelemenská, P. iák, M. Jurikovi and P. Pištek
Institute of Computer Systems and Networks, Faculty of Informatics and Information Technologies
Slovak University of Technology Bratislava
Ilkoviova 3
842 16 Bratislava, Slovakia
[email protected]
Abstract— The paper presents a knowledge assessment approach worse when distant learning is considered, where teachers have
developed for support of Digital system description course. In the do not have the possibility to check out the students
course the students should comprehend the methods and understanding of the solutions.
techniques used in digital system design, and gain the skills in The second part of the course assessment is usually formed
digital systems modelling using hardware description languages by an exam, sometime supplemented by some midterm tests.
(HDL). For many years the course exams used to be done only in The tests and exams are often performed in a traditional way –
pen and paper form. The students hated writing HDL programs using pen and paper. Although this type of exam gives
on paper and the teachers disliked correcting the programs on minimum chance for cheating, there are many reasons why this
paper as well. What is more, this type of exam has not allowed
type of exam is not suitable for practical skills assessment.
the verification of the student’s ability to debug the created
First, the students have no chance to check the correctness of
model, which is the substantial part of the student’s practical
their programs or HDL models before handing them out for
skills. That is why, some years ago, we concentrated our work on
evaluation. Secondly, this type of exam does not allow
designing a knowledge assessment system that will enable reliable
evaluation of practical skills in the area of digital system
verification of the student’s ability to debug the created
description without substantial teacher involvement. Several
program/model, which is a substantial part of student’s
applications have been developed and tested in the course. One of practical skills. Finally, hand written works are very difficult to
the best will be presented here. read, correct, and evaluate. It is very demanding and time
consuming work for teachers to evaluate programs/models
written on paper and the result is often uncertain. In order to
I. INTRODUCTION make the course assessment process more objective, fair and
effective, there was an urgent need for redesign of the
Knowledge and skills assessment has always been problem assessment. One of the solutions assumes that the substantial
in the courses devoted to programming languages, whether it part of the course assessment will be shifted from the
was assembly or higher level programming languages [3,4]. assignments to the skills based midterm tests. At the same time,
The problem is very similar in the case of hardware new technology will be developed to perform the exam and
description languages (HDL). To master the programming or midterm tests. The new applications, specially designed for this
modeling language, the students should solve various problems purpose, will enable the testing of not only the knowledge, but
using the language. It is useless to learn language constructs or also the skills used, all in an automated manner.
exact syntax by heart without writing any program. That is why
the programming courses assessment process often adopt the
II. LEARNING MANAGEMENT SYSTEMS CAPABILITIES
following schema.
To ease the work needed for knowledge and skills
The lab exercises are based on more or less individually
assessment in the course, various Content Management
allocated assignments that the students can solve at school or at
Systems (CMS), Learning Management Systems (LMS),
home. The solutions are evaluated by teachers and usually form
Computer-Based Training (CBT) or Web-Based Training
a part of the overall course assessment. Unfortunately, there are
(WBT) systems can be used which support automated
always some students who try to find the easiest way to get the
knowledge and/or assignments assessment. Nowadays there
lab work results (e.g. search for a complete solution on the
are a vast number of such systems, but most of them are unable
internet, copy the solutions from their classmates, etc.). The lab
to evaluate a student‘s skills using an online test.
teachers usually have no chance to find out whether the
students really did the programming work themselves, or not. Various systems supporting knowledge testing and/or
They can just check up, whether the students understand their assignment assessment have been reviewed in [2]. The
solutions and whether they are able to perform some minor emphasis was placed on supported questions, test types and
changes in them. But it is not an easy task to do. Another automated assessment. The existing systems can be divided
problem is that different teachers might use various scales for into two groups.
assignments evaluation. As a result, many students do not get Test Creation Tools: These tools represent standalone
the required knowledge and practical skills, even though they solutions. The tests created using these tools can not usually be
get the highest evaluation for lab assignments. This part of the connected to learning content and must be used en bloc. The
course assessment is often neither objective nor realistic created tests conform to e-learning standards and therefore can
although requiring a lot of teachers’ work. The situation is even be integrated to LMS. Another advantage is that these systems
127
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_22, © Springer Science+Business Media B.V. 2010
128 JELEMENSKÁ ET AL.
Macromedia Authorware 7
There are three types of users in the module – student
Toolbook Instructor
teacher and system.
Program Autor
HotPotatoes
WebToTest
1) Student: Student is the user with minimum rights. He or
QUIZIT
Moodle
Dokeos
eDoceo
ATutor
Ilias
she can do only actions allowed by users with higher rights
and therefore cannot modify functionality of the system.
Authentication – Students must log in to the system using
the Moodle built in authentication system. They will have
access to the Digital system description course, where a VHDL
Assignment management - - + - - - + - + - +
test is created as a special activity.
Question types Source code editing – A test consists of test assignment and
Program + - - - - - - - - - - VHDL source code editing windows, which can contain a
program core created by a teacher. Comfortable work during
Multi-choice single the test will be provided by VHDL syntax highlighting, similar
- + + - + + + + + + +
answer
to that used in ModelSim.
Multi-choice multi
- + + - + + + + + + + Source code compilation – Students can compile their source
answer codes unlimited number of times. The compilation errors are
True/False - + + - + + + - + + + returned to students by the system.
Gap filling - + + - + + + + + + +
Source code simulation – The compiled model can then be
simulated, using the test bench, prepared by the student or by
Matching - + + - - - - + + + - the teacher. The student should receive the signals waveforms
Slider - - + - - - - - - - - as an output.
Arrange objects - - + - - - - - + + - Source code submission – The students have the possibility
Place object
to upload their partial solutions to the server any time they
- - + - - + - - - + -
wish. After the final submission, the system will further
Identify correct picture process the student’s solution to get it evaluated automatically.
- - - - + + - - + - -
region The final submission of the source code is activated
Identify correct object - - - - - + - - - - - automatically in case the time allocated for the test elapses.
Crossword - - - - - - - - - + - Results review – Students can see the results of all their
+ =YES - = NO VHDL tests in the separated module.
2) Teacher: The teacher can administrate his/her teaching
module and adapt it to his/her own needs (i.e. adding teaching
are better elaborated and support more question types, and test aids, activities, tests, creating forums etc.).
creation is easier and flexible. The VHDL test will be created adding the special new
Learning Management Systems (LMS): These systems activity into the course. There are several activity attributes that
support e-learning and provide assignment management can be set during this process.
subsystems as well as testing subsystems. All considered LMS Instance name – Using this name the teacher can locate and
support the e-learning standards SCORM, ADL and IMS. download or review the results of the respective activity.
The capabilities of the reviewed systems are summarized in General instructions – These instructions are optional and
Table I. As we can see most of the reviewed systems support can be used to display any remarks made concerning the test.
several types of test questions but only WebToTest enables Start of the test – This attribute can be used to schedule the
students to edit, compile, and debug a program and submit it test at any date and time. The test will not be accessible to
using a special form. We were not able to find any system students before this time.
supporting skills based assessment in the area of HDL
modelling. This fact inspired us to develop several skills Test duration – The teacher has to set up the time that the
assessment systems supporting languages like VHDL and students can spend working on the test. After this time the final
SystemC. source code submission, followed by test evaluation is
activated automatically.
The standalone testing application presented in [5,12] and Test assignment – The teacher creates a text form of
the Drag&Drop Moodle module [13] that have been used for assignment saying what the students should design.
student’s knowledge and skills assessment in the last academic
year are two examples of them. The special VHDL testing Program core - The teacher can copy parts of the solution to
module, presented in this paper, is one of the latest solutions, the core which will be displayed in student’s VHDL source
suitable for online skills based examination in the area of code editing window. Typically at least the entity declaration of
digital systems modelling using VHDL. The module integrated the designed circuit is provided to make the student’s design
into the Moodle LMS is already a part of the set of assessment consistent with the reference. Also, the teacher’s test bench
tools used in the Digital system description course. can be provided here, in case it should be made visible to
students. Although this part is voluntary, if the teacher doesn’t
DIGITAL SYSTEM DESCRIPTION KNOWLEDGE ASSESSMENT 129
TABLE III.
TABLE MDL_VHDL_ODOVZDANE: STORES SUBMITTED SOLUTIONS
Name Description
ID Identifier of submission
the results are calculated. The algorithm for the evaluation of similar to that used by ModelSim (used during the term), is
student solutions is given in Fig. 2. used to make the editor familiar to students.
D. Test Creation Shell Script
After the instance of the testing activity is created by the
teacher, a reference waveform in the form of .wlf file is
automatically generated. A shell script is used for this purpose,
which is called after the test creation form is sent. This script
compiles and simulates the reference solution entered by the
teacher. The output of the simulation, a .wlf file, is then stored
together with the other files associated with this testing activity
instance (like e.g. image). They are all stored in the directory
moodledata/courseid/instanceid.
VI. RESULTS AND EXPERIENCE disposal a source code editor with syntax highlighting, a
We started to use the new testing module in the course VHDL language compiler, and simulator outputs. The
Digital systems description to evaluate students’ practical skills application gives the teacher a means to comfortably manage
during midterm tests. In the course, the students will pass three the testing system. Finally, the students have the opportunity to
midterm tests and one final exam. On each midterm test, the verify their designs before sending them for evaluation.
students are divided into three groups based on the testing
environment - TabletPC, Drag&Drop, or VHDL Moodle test. ACKNOWLEDGMENT
Each student will pass each type of test once. However, the test The support by Slovak Science Grant Agency (VEGA
assignment is the same, regardless of the test type. This 1/0649/09 “Security and reliability in distributed computer
constraint ensures that the severity of all the types of the test systems and mobile computer networks“) and HP Technology
should be more or less the same. Fig. 5 shows the results of the for Teaching grant 2007 “Mobile Education Center” is
first and the second midterm test - the distribution of Average gratefully acknowledged.
Test Attainment. As we can see, the students reached the best
attainment in the Drag&Drop test. The reason is that in this REFERENCES
case the correct solution is already given in the test and the [1] M. Mayer, “Distance Learning Support for Digital Systems Description
students just have to find the correct statements and place them Course,” master theses, FIIT STU Bratislava, 2009.
correctly into the VHDL model. [2] P. Polako, “Learning Support for Specification and Description
Languages Course,” master theses, FIIT STU Bratislava, 2007, pp. 86
[3] Ch. Douce, D. Livingstone and J. Orwell, “Automatic Test-Based
Assesssment of Programming,” ACM Journal of Educational Resources
in Computing,Vol. 5, No 2, September 2005
[4] C. A. Higgins, G. Gray, Pavlas Symeonidis and A. Tsintsifas,
“Automated Assessment and Experience of Teaching Programming,”
ACM Journal of Educational Resources in Computing,Vol. 5, No 3,
September 2005
[5] Š. Beleš, K. Jelemenská, M. Jurikovi, P. Pištek, T. Polák, J. Zeman and
M. Žucha, “Mobile Testing System,” Košice : Elfa, 2008, In: ICETA
Fig. 5. The first and the second midterm test results 2008 : 6th International Conference on Emerging e-Learning
Technologies and Applications. High Tatras, Slovakia, September 11-
After the test students were asked several key questions. The 13, 2008. - Košice : Elfa, 2008. - ISBN 978-80-8086-089-9. - pp. 133-
answers to these questions are shown in Fig. 6. As we can see, 138
most of the students think that the Drag&Drop test, composing [6] Š. Beleš, M. Jurikovi, P. Pištek, T. Polák, J. Zeman, and M. Žucha –
a model from given statements, is the best way of testing “Learning Support for Specification and Description Languages,” Team
practical skills. Project, FIIT STU Bratislava, 2008, pp. 83
[7] J. Kytka, “Learning Support for Specification and Description
Languages Course,” master theses, FIIT STU Bratislava, 2006, pp. 85
[8] N. Pollock, J. Cornford, “Theory and Practice of the Virtual University,
report on UK universities use of new technologies,”
http://www.ariadne.ac.uk/, Jun-2000
[9] Bruggeman, J., Long, H., M., Misra, C.: Effective IT Security Practices
(ID: EDU06165), Presented at EDUCAUSE Annual Conferences
(10/09/2006),
[10] Brown, M., B., Smith, J., M., Windham, C.: From Today's CMS to
Tomorrow's Learning Support System, Presented at EDUCAUSE
Annual Conferences (10/11/2006), WINNER: 2006 EDUCAUSE
Fig. 6. Survey results Catalyst Award.
[11] Hemmi, A., Pollock, N., Schwarz, C.: If not the Virtual University then
what? Co-producing E-learning and Configuring its Users, 8th European
VII. CONCLUSION Conference on Media in Higher Education, (Best Overall Paper), GMW,
University of Duisburg, 16-19 September 2003.
The first results prove that the kind of online test described [12] K. Jelemenská, M. Jurikovi, P. Pištek and P. iák, “Automated Skills
previously provides more realistic image of student knowledge Assessment in Mobile Environment,” in: Research, Reflections and
of digital systems modelling, compared to tests based on the Innovations in Integrating ICT in Education, Vol.1, V.International
types of questions commonly used in online tests. Because the Conference on Multimedia and Information & Communication
Technologies in Education (m-ICTE 2009), Lisbon, Portugal : Vol. 1. -
VHDL Moodle test requires creativity by students, the test Badajoz : Formatex, 2009. - ISBN 978-84-692-1789-4. - pp. 487-491
results are worse compared to the Drag&Drop test. [13] K. Jelemenská, E. Tomalová, and P. iák, Pavel, “Knowledge
The presented testing solution brings number of advantages. Assessment Improvement in Moodle Environment,” in: Research,
First, there is a substantial reduction in the demanding and time Reflections and Innovations in Integrating ICT in Education, Vol. 2,
consuming work of the teacher, relative to paper exams V.International Conference on Multimedia and Information &
Communications Technologies in Education (m-ICTE 2009, Lisbon,
correction. At the same time, the level of difficulty of the exam Portugal. - Badajoz : Formatex, 2009. - ISBN 978-84-692-1790-0. - pp.
is preserved. Second, it enables the teacher to check the skill of 979-983
students in the area of model debugging, which was previously [14] J. Geni, “Knowledge Assessment – Practical Example in Testing,” In:
not possible. All participants in the examination have at their Journal of Communication and Computer, Vol. 6, 8, 2009, pp. 65-69,
ISSN 1548-7709
Wikibooks as Tools for Promoting Constructivist
Learning in Higher Education: Findings from a Case Study
Ilias Karasavvidis
Department of Preschool Education
School of the Humanities
University of Thessaly
[email protected]
Abstract - The present paper focuses on wiki-related pedagogies all sorts of content. In fact, most of the typical web 2.0 tools do
in higher education and examines the potential of wikibooks in a not only rely on user participation but would actually be incon-
blended learning context. A wikibook was integrated as a com- ceivable without it. For example, web sites such as Wikipedia
pulsory class assignment in an undergraduate course in a pre- and YouTube would not exist without user-generated articles
school education department. Twenty seven students partici-
pated in the study which investigated (a) the extent of student
and videos respectively. Consequently, the first generation of the
participation in the wikibook and (b) students' perspectives on web, now called Web 1.0, has been transformed into a new, par-
the contribution of the wikibook to their learning. In stark con- ticipatory, read-write web which is known as Web 2.0. The most
trast to findings reported by other studies, results indicated that distinctive feature of Web 2.0 is that it relies on the wisdom of
the rate of student participation was very high as suggested by crowds [3] which in turn leads to crowdsourcing and wikinomics
indices such as page views, page edits, and number of students [4].
involved per page edit. Moreover, the students rated positively
wikibooks' contribution to the learning of the course concepts B. Web 2.0 and E-Learning 2.0
and the active mode of learning required. However, the students The impact of Web 2.0 technologies has been substantial [5].
also mentioned significant problems with the wikibook activity E-learning is one of the areas of education where Web 2.0 has
such as difficulty understanding certain sections and concerns
with the validity of their writings. The paper is concluded with a
had a major impact [6]. Terms such as Education 2.0, School
discussion of the findings and implications for further research. 2.0, Classroom 2.0, and Learning 2.0 are becoming common-
place.
Keywords – Wiki, Wikibook, Web 2.0, Collaborative The importance of Web 2.0 for enhancing learning in higher
Learning education is being increasingly recognized. Its significance lies
in that it can address some of the problems which characterized
I. INTRODUCTION several E-learning implementations. One of the most often
voiced criticisms is related to the adopted e-learning pedagogy.
A. The second generation of the world wide web More specifically, it has been argued that many E-learning envi-
Web 2.0 has attracted considerable attention over the past ronments are content-driven and are being mostly used to deliver
few years. Currently, the emphasis is on collective intelli- course materials on-line ([7], [8], [9]). Such transmissionist ap-
gence, the wisdom of the crowds, wikinomics, distributed proaches, which are dominant in E-learning, lead to the replica-
asynchronous collaboration, social networks and networking, tion of traditional modes of teaching and learning on-line [10].
social software, massively multiplayer on-line games, and As opposed to traditional E-learning, the role of students in
virtual worlds. According to [1], Web 2.0 can be defined in a the new E-learning 2.0 era is more active and participatory. The
number of ways. On a technical level, Web 2.0 tools such as students are not consumers of material which has been compiled
blogs, wikis, podcasts, social bookmarking, photo sharing, by instructors ([9], [5]). One of the major tenets of Constructiv-
instant messaging, VoIP applications, RSS feeds, mashups etc ism, the current dominant learning paradigm in higher education,
afford more possibilities for communication and collaboration is that learning can only be achieved through the active student
compared to Web 1.0 tools. For example, wikis have consi- participation in the learning process ([11], [12], [13]). This ac-
derable advantages over threaded discussions [2] and blogs tive participation by learners is perfectly compatible with the
have more affordances compared to Content Management notion of active participation by users in the context of Web 2.0.
Systems (CMS). Regardless of technical differences, howev- Thus, the affordances of Web 2.0 technologies are in line with
er, what sets apart Web 2.0 from the first generation of the the active approach to learning which constructivism requires,
Web is the role users play. which constitutes the added value of E-learning 2.0 from a learn-
In the first generation of the world wide web users were ing perspective. Web 2.0 technologies can help transform tradi-
simply browsing information which was provided on web tional higher education practices which are based on objectivist
pages. In the case of Web 2.0 the users are not simply con- notions of learning and are characterized by learning metaphors
sumers of information available on web sites. Instead, given that are transmisionist in nature.
the architecture of web 2.0, the users are actively contributing
133
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_23, © Springer Science+Business Media B.V. 2010
134 KARASAVVIDIS
C. Wikis ically, it has been reported that wikis can facilitate group work
Much like all other Web 2.0 tools, wikis have attracted [16], collaboration [17], collaborative learning [18], inquiry
widespread attention. ''WikiWiki'' is a Hawaian word that learning and co-reflection [19], and peer assessment [20]. Re-
means quick. According to [14], a wiki is a “freely expanda- garding performance gains, there is some evidence to suggest
ble collection of interlinked webpages, a hypertext system for that the use of wikis can lead to improved course performance
storing and modifying information – a database, where each (see e.g. [21], [22]) and improved writing and critical thinking
page is easily edited by any user with a forms-capable Web skills [23]. Finally, students appear to have positive attitudes
browser client” (p. 14). Essentially, a wiki system constitutes towards wiki use ([24], [25]) and find their wiki experiences to
a collection of webpages with information which can, in prin- be valuable for their learning ([22]; [17]).
ciple, be edited by anyone. Users can add content to existing Despite positive findings, significant challenges have been
pages, restructure the information on existing pages as well as identified: (a) minimal student participation ([26], [27], [28],
create new pages and create links between existing and new [21]), (b) limited on-line interaction ([23]; [29], [30], [22]), (c)
pages. The fact that everyone can create, edit and delete a preference for individual work [17], and (d) negative student
page means that wikis have a very open-ended nature. Cun- perceptions of wiki's contribution to their learning ([26]; [30];
ningham developed the first wiki, WikiWikiWeb, in 1995 and [22]). While this brief literature review does suggest that wikis
there are more than 200 clones today. hold great promise for higher education, one of the critical issues
The popularity of wikis is due to their simple and flexible which remain to be addressed is related to pedagogy. There is a
nature. Firstly, wikis are characterized by a non-linear hyper- knowledge gap regarding how wikis can be best incorporated
text structure: a wiki enables the creation of associative into the teaching and learning practices of higher education. The
hypertexts where each page contains links to other pages and evidence from existing research clearly suggests that the suc-
the user can choose which links to follow. Secondly, wikis cessful wiki integration into a course requires mildly put a tho-
are characterized by ease of access since all a user needs to rough redesign ([21], [28]).
use a wiki is a web browser. As a web application the wiki E. Focus of the study
does not require any specific browser or operating system
software. Thirdly, the technical skills required for participa- The present study aimed to fill this knowledge gap in wiki-
tion are minimal as users do not have to learn HTML to create related pedagogies in higher education. The specific focus in this
a webpage. Thanks to wiki markup, a simplified version of study is wikibooks, one of the many possible uses of wikis in
HTML, and the built-in WYSIWYG editors, the burden for university courses.
creating and formatting a webpage is significantly reduced. Wikibooks can be considered as class assignments whose
Finally, a wiki system also provides other functions such as main objective is to engage students in the collective writing of
(a) tracking of edits through a logging and notification sys- an electronic textbook. This effectively combines the active ap-
tem, (b) comparison between different versions of edits, (c) proach to learning which is required by constructivism and the
rollbacks to earlier versions, (d) discussions per wiki page and unique potential of wikis to be used both (a) as process tools to
(e) customizable read and edit access rights as well as pro- facilitate collaboration and (b) product tools to enable the crea-
tected pages. tion of a shared artifact such as a textbook.
Wikipedia, the prime example of wiki success, constitutes The potential of wikibooks has been researched in a number
an indication of how popular wikis have become. of recent studies in higher education settings (see e.g. [31], [22],
[32], [33], [34]). The typical use of wikis in those studies in-
D. Wikis in Education volves the writing of textbooks by students as part of the stan-
Wikis can be used to support a wide range of activities in- dard course assignments. As a rule, the findings reported by
cluding but not limited to documentation, reporting, project those studies are favourable in terms of academic performance,
management, on-line glossaries and dictionaries, discussion level of student engagement, and student attitudes. It should be
groups, and information systems. Wikis can be used to pro- stressed, however, that the aforementioned studies are either
mote learning, collaboration, communication, interaction, case studies or quasi experimental ones and, as such, are mostly
sharing, meaning making, and reflection. [15] outline three exploratory. Consequently the evidence is far from conclusive.
main types of uses for wiki: In fact, despite positive findings significant problems also
(a) distributing information: the wiki becomes the medium emerged. More specifically, some students in the [22] study ar-
to disseminate information as well as the place to store the gued that they did not learn anything new from the wikibook
information activity because they only wrote what they already knew. Along
(b) collaborative artifact creation: text is the main artifact the same lines, [34] argued that it is naïve to anticipate that the
produced, be it in the form of glossary or other formats wikibook activity would necessarily lead to collaborative writing
(c) discussion and review activities: anchored discussion, and learning without taking any other measures.
project case libraries, and peer review. The present study examines the potential of wikibooks in ter-
Even though wikis can be very important for education, tiary education in a blended learning context. A wikibook activi-
there is a knowledge gap in the literature. Research on wiki ty was incorporated in an undergraduate course in a preschool
applications shows that wikis can be very useful tools for education department as part of the typical course assignments.
supporting learning in higher education settings. More specif- Given the wikibook use in such a context, the study aimed to
address the following research questions:
WIKIBOOKS AS TOOLS FOR PROMOTING CONSTRUCTIVIST LEARNING 135
(a) what is the rate of student participation in the wiki- ments. Using the wiki log files, a number of measures were
book? computed: total number of pages created, number of page views,
(b) what are students' perceptions of their experiences with number of user views per page, number of edits, and the number
the wikibook? of users involved per page edit. The two survey instruments
were developed ad hoc for the purposes of the study. The first
II. METHOD instrument measured student background knowledge and atti-
tudes towards ICT. The second instrument measured student
A. Subjects
views about their wikibook experience. It contained 15 state-
27 students participated in the study. The students enrolled ments on several aspects of the wikibook and a typical 5-point
in an undergraduate course on the educational applications of Likert scale was used. Qualitative data collection involved 8
the Internet which was taught by the author in his parent insti- semi-structured interviews which aimed to provide a richer un-
tution. The course was elective for fourth semester students derstanding of how students viewed the wikibook.
and this cohort of students constituted a convenience sample.
H. Procedure
B. Course
The students were introduced to the wiki at the beginning of
The course aimed to introduce (a) common internet tech- the course. All students received an extensive introduction in
nologies and services and (b) a set of pedagogical principles using the wiki: in the sandbox provided students experimented
which could support educational uses of the internet. The with all technical aspects of the wiki such as creating new pages,
course comprised two main components: technical and peda- revising existing ones, linking, rolling back, and discussing. Af-
gogical. The main objective of the technical component was ter each class lecture or lab session the students selected a set of
to introduce students to information, communication, file, and concepts which they deemed both important and interesting,
application services. The main objective of the pedagogical searched and gathered relevant information, wrote up the text
component was to use the internet services to support a broad and contributed the information to the wikibook.
range of teaching and learning activities. In the beginning of the course the students were surveyed re-
Given that wikis were among the Web 2.0 tools introduced garding background knowledge of and attitudes towards ICT. At
in the context of the technical component and that the new the end of the course the students were surveyed about their
learning approaches were also presented in the context of the learning experiences with the wiki and eight students were ran-
pedagogical principles, the wikibook afforded a unique op- domly selected for interviews.
portunity to put theory into practice. The students could not
only familiarize themselves with the wiki technology and the III. ANALYSIS
associated pedagogical ideas but also experience it firsthand. To address the first research question, data analysis involved
F. The Wikibook assignment frequency counts and descriptive statistics. To address the
second research question descriptive statistics indices were
The wikibook constituted the main course assignment
computed for the two surveys while the interviews were content-
which was compulsory for all students. This assignment re- analyzed to determine the main themes the students reported.
quired students to (a) gather information on a number of top-
ics from various sources (e.g. printed textbooks, electronic IV. RESULTS
textbooks, wikibooks, web sites etc), (b) process the informa-
tion (e.g. restructure, summarize, compare, contrast etc), and A. Rate of student participation
(c) write the information to a wiki page thereby contributing The students created 66 pages which were viewed 6909 times.
to an existing chapter or sub-chapter. The contents page (i.e. the wiki start page which is the equiva-
Regarding participation, no specific guidelines were pro- lent of the table of contents) is given in table 1.
vided: the students could freely decide on how much informa-
tion to contribute to the wikibook. Moreover, there were no Internet
specific requirements set for covering a minimal number of World Wide Web
topics or chapters: the students could contribute information The social impact of the Internet
File Sharing
to topics which they found interesting. To encourage partici- Secure Data Transfer
pation and avoid social loafing, the wikibook assignment Networks
amounted up to 60% of the total course credit. Distance Education
MediaWiki (http://www.mediawiki.org), the software that pow- Hacking & Cracking
Portals
ers Wikipedia, was the software used in the study. To avoid Operating Systems
spam, vandalism, and all sorts of extraneous distractions, the Important Internet Applications
wikibook was used in a closed-study group. Internet & Schools
Table 1: The chapters of the wikibook
G. Measures
For the purposes of the present study both quantitative and The main wiki statistics are presented in table 2. As can be
qualitative measures were employed. Quantitative measures seen in the table, some pages were more popular than others. In
included the wiki system log files and two survey instru- fact, the wiki contents page amounted to 88% of the total num-
136 KARASAVVIDIS
ber of views. On average, each page was read 104 times and For the most part the pages were reasonably interlinked. The
the variability was very high. Log file analysis, however, in- contents page (initial wiki page) included links to 12 wiki pages
dicated that 55 out of the total 66 pages were created by one through which the majority of all other pages were accessible.
student and were characterized by a small number of views Only 9 pages (13,64%) were not linked to any other pages in the
(typically 1-3). While technically these constituted wiki pag- wiki.
es, they contained little information -usually an image or icon
and a couple of sentences describing the theme of the page. B. Student perceptions
As this student case clearly represents an outlier, the descrip- The student responses to the questionnaire administered at the
tive statistics reported in table 2 should be interpreted with end of the course are given in table 3.
caution. To address this issue, we examined the page views
for the 13 most viewed pages, that is, the contents page (main Variable Mean SD
wiki entry page) and the 12 other pages listed in that page. a 3,85 0,66
This subset of wiki pages accounted for 88,57% of the total ease of use
page views and did not include any of the wiki pages created a 4,07 0,92
by the aforementioned student. After this correction, the mean ease of access
general b 4,22 0,58
number of reads increased by a factor of five (M=470,69,
SD=578,97). This finding clearly suggests a very high rate of overall impression
access. c 3,89 0,58
The wiki pages were edited 1065 times. Again, the statis- understanding contribution
tics reported in table 2 are subject to the same distortion con- d 4,19 0,40
sidering that most of the 55 pages that one student created had writing contribution
not been edited at all. Following the same procedure adopted d 3,81 0,62
above to deal with this outlier the descriptive statistics were reading contribution
d 3,59 0,64
recalculated. The contents page and the pages listed as the contribution
edits made contribution
main wikibook contents were edited 757 times (71,08% of the d 3,59 0,80
total edits recorded by the system). The average number of
edits received contribution
edits per page was 58,23 which is considered to be very high
d 3,59 0,80
and indicates that all pages underwent extensive revisions. It
discussion contribution
should be noted, however, that even for this subset of wiki
pages the variability was rather high (SD=75,20). e 1,85 0,91
The number of users involved per page edit is an important extent of revisions made
edits
measure as it shows how many students were involved in the e 1,74 0,86
authoring of a wiki page. Table 2 shows that 2,7 students extent of revisions received
were involved per page edit. However, as discussed above, a. Very difficult-Very easy b. Very negative-Very positive. c. Not at all
this figure is heavily biased because 83% of the pages were important – Very important. d. Strongly disagree-Strongly agree. e. Mi-
created by a single student. After correcting for this outlier, nor-Major
the mean number of students involved per page edit rises to
8,69 (SD=6,21). This suggests that the wiki pages were essen- Table 3: Descriptive statistics of students' views regarding the wikibook
tially co-authored as on average more than 8 students were
involved in the writing of every wiki page. This finding is As table 3 shows, the students did not experience any difficul-
very promising as the number of joint edits reflects the degree ties in using or accessing the wiki and rated their overall wiki
of on-line student collaboration. Finally, in addition to obtain- experience very positively (4,22). In fact, the mean for this vari-
ing evidence of on-line collaboration among students, table 2 able reflected the highest positive rating in the questionnaire.
also shows that the levels of individual student participation Moreover, the students evaluated positively wiki's contribution
were also very high: the average number of edits per user is to their understanding of the course concepts. While students
about 40 which suggests that students' contributions were rated positively the four aspects of wiki work, the contribution
extensive. of writing to learning clearly stands out. This is quite interesting
as the students appeared to favor writing over reading, editing,
Variable Total min max Mean SD
and discussing. Finally, table 3 shows that students reported that
they neither made extensive edits to the writings of other stu-
Page Views 6909 1 2045 104,68 308,8 dents nor were their texts extensively edited by fellow students.
Page Edits 1065 1 262 14,3 39,1 The students were also asked to rate the most difficult aspect
Users per Page n.a. 1 22 2,77 4,03 of wiki-related work. The final rating is as follows: writing:
Edit 33%, reading: 11%, revising: 44%, discussing: 11%. Thus, re-
Edits per User 1065 3 184 39,44 44,94 vising was found to be the most difficult aspect of the wikibook
Table 2: Descriptive statistics for the main Wikibook variables
assignment.
The qualitative analysis of the interviews indicated that while
students expressed positive views about the wiki, they also men-
WIKIBOOKS AS TOOLS FOR PROMOTING CONSTRUCTIVIST LEARNING 137
tioned considerable difficulties. An outline of the pros and where the students collaborate on-line and create an electronic
cons is presented in table 4. textbook. A number of recent studies have began to explore how
wikibooks can be used in the context of higher education ([31],
Pros Cons [22], [32], [33], [34]). Although the findings from those explora-
Facilitation of conceptual under- first to write
tory studies are rather positive, a number of important issues
standing as the first contributors usually related to pedagogy remain unexplored. The present study aimed
the wiki task helped students master summarized the main points pre- to fill this knowledge gap regarding wikibook applications in
factual information sented in class, the remaining blended learning contexts. The study examined both the extent
students reported difficulties in
Information gathering & ex- finding what to write
of student participation in the wikibook activity and students'
change perceptions of the wikibook activity.
the wiki task necessitated informa- extent of contributions Regarding the first research question, the findings of the study
tion gathering for the students because no specific guidelines indicate that the students created several pages which were both
were given, the students were
group work unable to decide how much infor-
viewed and edited several times. As the analysis indicated, each
the wiki task required a collabora- mation in each book section was wiki page was viewed on average 470 times. Moreover, each
tive effort enough page underwent extensive edits as it was edited on average 58
times. Finally, more than 8 students were on average involved in
revising structuring
the task afforded the opportunity to the students considered the open-
the editing per wiki page. As a whole, these findings are very
revise not only one's texts but also ness of the wiki as somewhat limit- promising as most wiki-related studies typically report that stu-
the texts contributed by others ing and requested more structure dent participation in the wikis is limited ([26], [27], [28], [21])
(e.g. providing a detailed table of and the same holds for on-line interaction ([23], [29], [30], [22]).
writing contents)
the students considered the wiki
Thus, unlike former studies, the outcomes of the present one
task to be a genuine text-production understanding clearly suggest that a compulsory wikibook activity for which
process and not a simple reproduc- the students complained that some students can get a considerable portion of the total course credit
tive one sections of the wikibook were can effectively alleviate the problem of limited student participa-
unreadable which obfuscated un-
mode of learning derstanding
tion which is reported in the literature.
the wiki task exclusively relied on Regarding the second research question, the quantitative anal-
student contributions and this pro- validity of interpretations ysis indicated that the students rated their wikibook experience
moted a more active approach to the students were uncertain about very positively as the mean of 4,22 on a 5 point scale suggests.
learning in the context of the course the validity of the information
which they added to the wikibook
As for the contribution of wikibook to their learning of the
course concepts, the students gave higher ratings to writing and
technical issues reading than to the remaining aspects.
some students mentioned that they The qualitative analysis of the student interviews revealed a
were sometimes unable to commit
the edits they had made to a wiki
number of positive and negative points of the wikibook activity.
page (probably due to concurrent On the one hand, the students favored the wikibook activity for a
edits) number of reasons. It should be noted that all aspects of wiki-
Table 4: Students' perceptions of the pros and cons of the wikibook activi- book which were viewed as favorable by the students are in line
ty with the theory and research regarding the potential contribution
of wikis to learning. As far as the contribution of the wikibook
V. DISCUSSION
to the learning of the course concepts is concerned, the facilita-
The impact of Web 2.0 on many domains of human activi- tion of conceptual understanding and the active mode of learning
ty has been substantial [5]. One of the areas where this impact that the wikibook required stand out from a learning point of
is most evident is education and more particularly E-learning view. On the other hand, the students also expressed considera-
[6]. Considering that many web-based education and e- ble difficulties. Two of the problems students reported, namely
learning approaches have tended to replicate traditional mod- understanding and validity of interpretations, deserve closer fur-
es of teaching and learning which are transmisionist in nature, ther examination because they are related to learning and, to a
Web 2.0 is expected to enhance learning in higher education large extent, contradict the facilitation of conceptual understand-
because, by default, it requires the active participation of us- ing that the students mentioned as one of the most important
ers. Unlike Web 1.0 tools, the corresponding Web 2.0 tools wiki contributions. The remaining difficulties identified are
would not be at all possible without user participation. This more procedural in nature and can be more easily addressed (e.g.
specific feature is a perfect match to the active learner partici- devising specific rules for wiki contributions or providing more
pation which according to Constructivism characterizes all structured versions of the task such as a detailed list of contents
human learning. Thus, Web 2.0 holds great promise for pro- with chapters and sub-chapters). To a large extent, the outcomes
motive active learning in higher education. Wikis are amongst of the qualitative analysis replicate other literature findings re-
the Web 2.0 tools which have great potential for contributing garding the difficulties students experience when using wikis
to learning because they are uniquely suited to support both (e.g. [26], [30], [22], [35]). Thus, the fact that students face im-
process and product approaches. There are many types of portant challenges when engaged in wiki activities remains and
wiki uses in education and the specific focus in this paper was should be addressed in future research.
wikibooks. Wikibooks can be used as class assignments
138 KARASAVVIDIS
Abstract— Information Quality has become an ever The workstation model was predominant in the late 70s
increasing problem for many organizations relying on and early 80s [2], followed by widespread deployment of
information systems for many of their core business the client server model [3]. The rapid expansion of
activities. Despite advancements in technology and internet services soon followed [1, 3]. This expansion has
information quality investment the problem continues to continued with the pervasive availability of mobile
grow. The context of the information systems over the last devices and wireless technologies [1, 6, 7]. Virtualization
number of years has also evolved. This has resulted in a has also made an impact especially with the widespread
diverse number of architectures accessing multiple implementation of unplanned solutions [8].
information systems. The context of the deployment of an
The significant changes and increase in complexity of
information system in a complex environment and its
IS context has in many cases occurred independently of
associated information quality problems has yet to be fully
the underlying databases accessed [1]. Applications may
examined by researchers. Our research addresses this
have been designed, built and tested with a mature
shortfall by examining the relationship between the context
of the information system and information quality. The
software development methodology for a particular
research is conducted within the context of strategic
context, yet within a very short period of time it may be
information quality within the higher education system in accessed and predominantly employed from a different
Ireland specifically the Institute of Technology Sector. A context [9]. Data models have also evolved [2],
mixed methodology approach is employed to tackle the nonetheless a considerable number of IS in use today have
research question of the impact of the context of an data models designed prior to the contexts that are
information system on strategic information quality. employed to access them. There are as a result multiple
Research to date has examined different information accesses from diverse and complex contexts.
systems’ context factors and their effect upon information The quality of information that strategic decision
quality dimensions both objectively and subjectively. Thus makers require is heavily dependent on these IS. An
far our research puts forward a novel information quality examination of the Institute of Technology Sector in
methodology that is context related; that is it takes the user, Ireland reveals that Management Information Systems,
task and environment into account. An initial experiment Decision Support Systems and User driven Office
has been conducted. Results of preliminary experiments Information Systems are the primary source of this
indicate that context affects the perception of information information. In light of the evolving nature of both the
quality. technology and the higher education sector it is imperative
to measure the impact of these contexts on the strategic
Keywords ü Information Quality; Higher Education; information quality (IQ).
Strategy; Information Systems
139
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_24, © Springer Science+Business Media B.V. 2010
140 ELWOOD AND FOLEY
research. Teaching nonetheless continues to be seen as the research, regional development, links with industry, social
primary mission of the institute of technology sector, with inclusion and meeting skills needs” [14]. The response of
research as an interesting, but ultimately costly, adjunct. the institute of technology sector to these emerging
The Regional Technical Colleges Act [11] also established metrics of performance, demanded nationally, yet
the following bodies; Governing Body, Academic Council assessed locally, is the introduction of frameworks of
and the designation attaching to the principal officer of appropriately designed management information systems
each institute, namely Director (re-designated to that of which seek to;
Institute President under the Institutes of Technology Act Improve organisational effectiveness (the determination
[12]. Consequently, three distinct groupings now exist in of unit costs, the quality, retention and destination of
each college: (1) The Governing Body, which oversees the graduates, including attrition and retention reporting.).
work of the College Director, (2) the College’s senior
Understand and minimise the risks facing the respective
management team, headed by the College Director, which institute of technology (the generation of strategy
oversees the day to day management of the College, and
supportive data, the prioritisation of data characterised by
(3) the College’s Academic Council, charged with reliability and timeliness to inform management decision
protecting, maintaining and developing the academic
making).
standards of the courses and activities of the college. Two
of these bodies; Governing Body and the senior Comply with statutory requirements and corporate
management team, are essentially managerial in focus, governance requirements (the generation of statutorily
requiring data which facilitates the discharge of their required financial information and annual reports.).
managerial duties. The mid 1990s saw further pressure on
Ireland’s higher education institutes to act as levers in III. OVERVIEW OF INFORMATION QUALITY
helping Ireland’s economy transition to a knowledge Currently many of the IQ frameworks are employed
economy. The White Paper “Charting Our Education post deployment of the IS. A number have been
Future” [13] signposted a role whereby all higher implemented as a direct result of IQ problems,
education institutes were given responsibility, through highlighting accuracy or completeness problems in the
research, for the development of new ideas, new main. The use of IQ frameworks allows for an
knowledge and the application of existing knowledge. examination of the environment without the need for prior
This devolution of accountability and transparency to all knowledge of the development methodology [17].
higher education institutes has also served to heighten the The area of IQ as a field of research has developed
requirement for timely, accurate and relevant information considerably within the past decade, predominantly
within Ireland’s higher education institutes. stemming from reliance by enterprises at all levels on IS
This altered higher education landscape, coupled with for the completion of their core business activities. IS
Ireland’s economic transformation, provided the backdrop context has also evolved with widespread deployment of
to the OECD Review of Higher Education in Ireland in web services. Enterprises are now increasingly
2004 [14]. Whilst lauding Ireland’s success in increasing characterized having mission critical systems based on
participation rates, and its growing recognition of the diverse contexts, yet have no measure of the quality of the
importance of research to sustainable economic growth, information that they derive their business and profit from.
the review nonetheless highlighted that Ireland’s tertiary This is a wide and varied problem affecting nearly every
education system is at “a significant point of departure” major type of business activity and enterprise. Al Hakim
Over the next decade higher education in Ireland will [6] provides numerous examples of these anomalies from
develop against the framework for higher education many areas, outlining the reason along with the particular
outlined in the Programme for Government [15] and the IQ dimension affected. The examples indicate how the
National Development Plan [16]. The policy objectives generation of information from disparate sources can
outlined for Ireland’s higher education system is to be at impact on many aspects of an enterprise, often not even
the top rank of performance within the OECD [14]. A considered when the IS was initially designed. The impact
common feature pervading most of these policy of these IQ problems has prompted researchers to examine
documents is the increasing use of the funding whip-hand such aspects as IQ dimensions, IQ assessment and IQ
to encourage the higher education system to both agree to management.
and implement centrally determined policy initiatives.
IV. DESIGN AND DEVELOP AN IQ METHODOLOGY
More specifically the following development needs are
specified as key metrics in the delivery of this policy An examination of the literature [5, 6, 17-22] suggests
objective; (1) increased participation and improved access, that the issue of context has not been adequately
(2) greater flexibility of course offerings to meet diverse addressed. Consequently we believe that accessing the
student population needs in a lifelong learning context, (3) same information from diverse contexts will be
enhancement of the quality of teaching and learning, (4) problematic and inconsistent with existing frameworks
significantly increase the number of PhD numbers and and IQ measurement.
research activity, and (5) effective provision for Our construction of the IQ methodology examines and
technology transfer [14]. The Higher Education Authority builds on work completed by Knight & Burn [20] in the
(HEA) envisions a critical role for the Institute of area of algorithm choice for Internet search engines and
Technology sector, that is broadly in line with the OECD Naumann and Rolker’s [23] analysis of information
[14] recommendations, and which recognises: “that the sources. Our dimension selection were chosen by
Institutes of Technology and the Universities have applying Leungs [22] Importance, Urgency and Cost
developed differently and to establish greater clarity on metric. The methodology proposed is flexible in its
the diverse role of Institutes of Technology and the application and caters for the employment of other
Universities in such areas as, teaching and learning, usability metrics such as Doll and Torkzadeh [24] The
THE IMPACT OF THE CONTEXT OF AN INFORMATION SYSTEM 141
Context-specific IQ Measurement
In order to overcome these context deficiencies we Ser DB
Context-specific
propose a novel IQ methodology that takes the context of IQ Requirements
the IS into account, allowing us to analyse the perception Assessment
Objective
Ri
IQ Dimensions Technique RRi
of the same information from different contexts. Knight Subjective Ri i
& Burn [20] point out that despite the sizeable body of
literature relatively few researchers have tackled the Priority Testing
difficult task of quantifying some of the conceptual Survey
Dimension
Instrument
definitions. Naumann and Rolker [23] further suggest that Dimensions Score
Application
N REFERENCES
IQ = ¦ α i IQCi where ∀αi :0 ≤ α
N
i ≤ 1
i =1
¦i =1
α i = 1 [1] Adelman, S., L. Moss, and M. Abai, Data Strategy. 2005:
Addison-Wesley.
Equation 1 Aggregate measure of information quality
[2] Date, C., Databases in Depth. 2005: O Reilly.
The aggregated value should define the quality level [3] Sangwan, R., et al., Global Software Development Handbook.
that characterizes information source. The approach to 2007: Auerbach Publications.
use the average as an aggregation functions may not be [4] Rosenblum, M. and T. Garfinkel, Virtual machine monitors:
current technology and future trends. IEEE Computer, 2005.
suitable among heterogeneous dimensions since 38(5): p. 39-47.
dependencies introduces bias that negatively affect the
[5] Lee, Y., et al., Journey to Data Quality. 2006: MIT.
reliability of the assessment procedure. Change in the
[6] AlHakim, L., Challenges of Managing Information Quality in
context (accessibility service) will have an impact on other Service Organizations. 2007: Idea Group Pub.
dimensions in a framework and as a consequence the [7] Fisher, C., et al., Introduction to Information Quality. 2006: MIT.
aggregate score. The research provides a methodology to [8] Gilen, A., B. Waldman, and J. Humphreys, Technology
measure this impact. This it is argued should be Assessment: The Impact of Virtualization Software on Operating
completed prior to applying the aggregate, because it will Environments. 2006.
provide a more accurate result with respect to IQ. [9] AlHakim, L., Information Quality Management Theory and
Applications 2007.
IX. LIMITATIONS [10] Mulcahy, Report of the Steering Committee on Technical
Although our research revealed interesting results, this Education. 1967.
was a pilot study concentrating only on a subset of [11] Regional Technical Colleges act. 1992: Ireland.
dimensions and functionality used by strategic decision [12] Institutes of Technology Act. 2006: Ireland.
makers in higher education. However we believe that the [13] Education, D.o., Changing our Education Future, Education,
principles and methodology proposed have the potential Editor. 1994, Stationary Office.
for wider application across IS in higher education. The [14] OECD, Review of Higher Education in Ireland. 2004.
application of the full experiment will address this [15] Taoiseach, D.o.A., Programme for Government 2007-2012,
D.o.A. Taoiseach, Editor. 2007.
limitation. Correlation has been used to examine the
relationship between IQ dimensions. Analysis of [16] Taoiseach, D.o.A., National Development Plan 2007-2013,
Taoiseach, Editor. 2007, Stationary Office.
Variance will be completed for all sections of the research,
[17] Batini, C. and M. Scannapieco, Data Quality Concepts,
thus allowing the general application of our findings. Methodologies and Techniques. 2006: Idea Group Publishing.
Finally the research to date has only employed [18] Eppler, M. and P. Muenzenmayer. Measuring Information in the
experiments along with survey instruments as its research Web Context: A survey of state of the art instruments and an
approach. The other data collection methods which will application methodology. in 5th International Conference on
form an integral part of the research such as diaries, semi Information Quality. 2000.
structured interviews and focus groups will be expanded [19] Helfert, M. and M. Ge. A Review of Information Quality Research.
and form part of the next phase of the project. in 11th International Conference on Information Quality. 2006.
[20] Knight, S. and J. Burn, Developing a Framework for Assessing
Information Quality on the World Wide Web. Informing Science,
2005. 8: p. 159-172.
X. CONCLUSIONS AND FUTURE WORK [21] Lee, Y., et al., AIMQ: a methodology for information quality
This research contributes to the analysis of IS context assessment. Information and Management, 2002. 40: p. 133-146.
and IQ. In recent times companies have invested heavily [22] Leung, N., Quality Metrics for Intranet Application. Information
in IQ programmes in an effort to improve IQ [9]. Our and Management, 2001. 38: p. 137-152.
research demonstrates that a relationship does exist [23] Naumann, F. and C. Rolker. Assessment Methods for Information
Quality Criteria. in 5th International Conference on Infprmation
between IS context and other dimensions in an IQ QualityIntetnational Conference. 2000.
framework. The traditional method of measuring IQ [24] Doll, W. and G. Torkzadeh, The Measurement of End User
dimensions will require re-examination as relationships Computing Satisfaction. MIS Quaterly, 1988. 12(2): p. 259-274.
between the context and other dimensions become more [25] Wang, R. and D. Strong, Beyond Accuracy: What Data Quality
clearly established. By demonstrating clearly that the Means to Data Consumers. Journal of Management Information
context of the IS affects the quality of the information Systems, 1996. 12: p. 5-34.
produced, strategic decision makers can proactively allow [26] Galliers, R., Information Systems Research: Issues, Methods and
for this when using information generated by these IS. Practical Guidelines. 1992: Blackwell Scientific.
This research also contributes to the field of IQ research [27] Minger, J., The Paucity of Multimethod Research: A Review of the
by providing a methodology and test environment that can IS Literature. 2001, Warwick Business School.
be employed in a context related manner. Upon [28] Greenwald, R., R. Stackowiak, and R. Stern, Oracle Essentials:
refinement this methodology has the potential to allow Oracle Database 11g. 4th ed. 2004: O Reilly.
organisations to measure the impact of introducing new [29] Field, A. and G. Hole, How to Design and Report Experiments.
2006.
contexts post the development of an IS.
[30] Stvilia, B., et al., A Framework for Information Quality
Assessment. Journal of the American Society for Information
Science and Technology, 2007. 58(12): p. 1720-1733.
Can Mobile Computing Help Reverse the Decline in
Informatics Studies?
D.Vitkute-Adzgauskiene, V.V.Fomin
Dept. of Informatics, Vytautas Magnus University
Vileikos 8,
LT-44404 Kaunas, LITHUANIA
Abstract-- In this work we analyse the factors leading to there are also “global” factors explaining the decreasing
the decline of popularity of traditiona l Computer Science / popularity of the Informatics in Europe at large and the U.S.
Informatics study programmes as observed in Europe and the To name a few, the transformation of the hardware (HW) and
U.S. We further examine the possible directions for improving software (SW) markets, job market trends, transformation of
the Informatics curricula to reverse the downturn trend. Finally,
we report on the first positive experience of including Mobile
the computing paradigm, the effect of the dot-com bubble
Computing related topics into the Informatics study programme burst, inadequate curricula for Informatics education at the
with resulting increase in students’ interest. school level, etc.
In this work, we reflect on the factors affecting the
I. INTRODUCTION downturn in Informatics enrollments. We analyze and
synthesize data from Lithuania, EU, and the U.S., aiming to
The last decade saw rapid growth of Internet usage. Our understand how the universities can react to the decline in
young generation today is usually referred to as the “Internet enrollments, and, specifically, whether inclusion of Mobile
generation”, having a spectrum of uses for the Internet, from a Computing in the study programmes can help re-instigate
virtual library, to a place for social networking, to study students’ interest in the Informatics studies.
room, to gaming room, etc.
Along with the Internet becoming an inseparable part of II. THE RISE OF THE MOBILE SOCIETY
our daily activities, phenomena of the “Mobile” came to
witness much stronger presence in the life of the world’s In many countries, mobile communications have entirely
population. With an estimated 20 percent of the world’s changed the overall national picture of ICT usage. Especially,
population being Internet users in 2007, there were twice as it is true of countries, which had lower personal computer and
many mobile phone subscribers, and there is a plenty of room fixed Internet penetration before the advance of the mobile
for growth, except maybe for the developed countries, where era – such is the case for most of the former Soviet Republics,
there is already a 100 percent penetration rate [1]. With Lithuania being one of these.
mobile phones having become a commodity, it shouldn’t Today, Lithuania is a highly saturated mobile phone user
come as a surprise that mobile broadband is becoming a country – its mobile penetration by the end of Q2‘2008 was
popular way for accessing Internet, and short messaging and 148% [2], i.e., every citizen on average had 1.5 mobile
mobile data services are being used to accomplish everyday subscriptions in his/her disposition. Besides the population,
tasks. the number of mobile subscriptions is increasing in machine-
Despite the visible proliferation of ICT (Information and to-machine applications (e.g., remote control or security).
Communication Technologies) in our everyday lives and Demographic categorization reveals that 88 percent of all
business, surprisingly, during the last decade universities in population, falling into the 16–74 age group, are mobile
Lithuania and globally saw a substantial downturn in phone users [3]. For the 16-24 year group this number is even
Computer Science (CS, referred to as “Informatics” in higher – 98%, which let us safely assume that 100% of
Europe) and Management Information Systems (MIS) university students are mobile subscribers.
enrollments. In the United States the trend started earlier, with While fixed Internet user numbers were also growing rather
enrollments in CS falling by half from 60.000 in 1998 to rapidly in Lithuania during the last decade, showing an
30.000 in 2006/7.1 In Lithuania, we saw the same 50 percent impressive 835% growth during 2000-2008, the Internet
decreases in admission numbers for Informatics students from penetration reached only 59 percent of population2 - much
2004 to 2008. While the Lithuanian case can be partly lower than that for mobile.
explained by highly hyperbolized admission numbers in The relatively low fixed Internet penetration may be the
2001-2002 (2-2,5 times higher, compared to 1999) due to the reason why Lithuanian mobile subscribers show high usage of
State policy and Government orders, as well as by the different mobile phone enabled services, as compared to other
possibilities to study abroad after Lithuania joined the EU, countries. For SMS, for example, statistical Lithuanian user
2
As of March 31, 2009, according to the Internet World Stats,
1
CRA Taulbee Survey, http://www.cra.org/statistics/ http://www.internetworldstats.com/stats9.htm.
145
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_25, © Springer Science+Business Media B.V. 2010
146 VITKUTE-ADZGAUSKIENE AND FOMIN
sends 156 SMS per month – more than two times the number steady decline in the enrollments to Informatics programmes
for the French subscribers, and mobile data services are used around the world. When analyzing possible reasons for the
by half of the Lithuanian population [2]. decline, we may single out three major factors:
Globally as well, mobile Internet is becoming a daily 1) Changes in the environment, including technology
activity for more and more people. In the U.S., for example, changes, transformation of the industry, job market’s
the number of people accessing news and information on their image.
mobile devices more than doubled during the last year [4]. 2) Inadequate Informatics education at schools.
Mobile service usage closely follows the habits of modern 3) Conservatism of Informatics study programmes.
society. With the growing popularity of social networking
services, access to these services from mobile devices has A. Transformation of the industry
grown over 152 percent in Western Europe from Q3’2007 to At least since the dot-com bubble burst in 2001, Computer
Q3’2008. For the U.S., this category saw an impressive 427% Science (CS) / Informatics enrollments saw a downturn in
increase, while access to news and infotainment services enrollments, which continues to date. Although slowly
increased by 100-190 percent [5]. For the popular Facebook recovering since, this time a more fundamental challenge may
service, the mobile access rate came to stand even with the deepen the “enrollments crisis” – that of the transformation of
fixed Internet access rate. the managerial and industrial paradigms. This time the new
Interesting to note, the use of mobile services kept to Lithuania programme of “Business Informatics”3, and
increasing even during the economic downturn. The top U.S. world-established concept and Business Schools’ study
sites saw doubled use from wireless pre-paid users over the program of Management Information Systems (MIS) is also
last year [6], indicating that people consider mobile under the pressure. The global trend of flattening of enterprise
broadband a convenient and cost saving access channel to resulted in management functions becoming increasingly
everyday (Internet) services. horizontal. This, in turn, resulted in elimination of traditional
We can find many examples showing that the mobile usage Ford-style vertical integration with its multi-divisional
growth is dependent on the availability of convenient mobile management system. Consequently, where there used to be
services for everyday use. Mobile parking services in Estonia Chief-X-Officers for every division X, in the new enterprise
and Lithuania are a good example – the services gives a paradigm the Chief Information Officers (CIOs) found
possibility of paying for parking, extending parking time, themselves in a situation where they “couldn’t communicate
getting corresponding notifications through SMS or mobile with the top management group, were ineffective change
Internet services. In Lithuania, mobile payment for parking agents, or could not contribute to business strategy” [7].
grew to about 30% of all parking transactions in just 3 years. Transformation of the industrial landscape couldn’t have
For Tallinn, the capital city of Estonia, the number is 63%. left unaffected the job market, and student enrollments in
When looking for what constitutes the “convenience” in these managerial programmes. Secondary school graduates, when
services, in both cases we find that designers of the services deciding on future professions, cannot be encouraged by
were successful in accommodating the habits and preferences numbers showing that CS graduates are up to 10 times less
of local population (reminders, payment schemes, interfaces, likely to be employed, than graduates from any other subject.
etc.) into the service offering. Other successful non-Internet The employment stats published last year by The Guardian
mobile service uses examples exist, too, such as SMS-based [8] place CS graduates the last in the list of 20 different
purchase of public transportation tickets, vending machine degree programmes.
payments, etc. The recent data from the U.S. present a somewhat more
The trends, that we notice already, show that mobile phone, optimistic picture for the future of Informatics than that in
possessed today by the absolute majority of active society UK. For the first time in six years, the number of CS majors
members, enables fast and convenient access to mobile enrolled at U.S. universities increased by 8.1 percent [9]. IT-
services. If the necessary services are available, it is rather related jobs are reported to be still in demand and expected to
easy for a user to start using them, and, moreover, to enter the stay at this level or even experience growth in the years to
state of “can’t do without”. However, the growth potential is come. Especially this is true for those specialities, that require
still high in this area, and challenges exist. With aging more complex IT and interdisciplinary skills, e.g. IT
population in Europe, and the emergence of “smartphones” as management consulting [10].
mobile Internet access device, the majority of mobile phone
users aren’t familiar with many advanced mobile phone B. Inadequate education in schools
features, and would require an adequate training to take The prospects of the job opportunities as drawn by mass-
advantage of the devices’ functionality, as it is getting closer media in Europe are “complemented” by rather narrow scope
and closer to a computer functionality. of CS/Informatics teaching at secondary schools. This
tendency is true both for the U.S. and Europe. For example, a
III. THE DECLINE OF INFORMATICS AT UNIVERSITIES
3
Despite the growing dependence on ICT in daily lives, It is probably correct to say that “Business Informatics” is the
closest proxy of Management Information Systems programme in
work, and communication, for the last decade there was a Lithuania.
CAN MOBILE COMPUTING HELP REVERSE THE DECLINE IN INFORMATICS STUDIES? 147
study by Blum and Cortina [11], as quoted in recent report (ECSS) held in 2005-6, where heads of computer science
“Informatics Europe” [12], identified the following reasons departments from all over the European region joined forces
for the decline in enrollment in CS courses and programs at to define and promote common policies and study common
the high school level in the US: issues. In August 2007 Informatics Europe published a report
• Dot-com bubble burst/not enough high paying careers “Student Enrollment and Image of the Informatics Discipline“
• Increased amount of required courses and tests [12]. The report aimed at answering the following questions,
• Elective status/competition with many other electives among other:
• Lack of understanding of CS by guidance counselors and • Why is student enrollment a problem in
administrators Informatics/Computing, at least in many of the (Western)
• Ill-prepared teachers/lack of interest European countries?
• Difficulty of material/not ‘fun enough’ • Why is enrollment by female students lagging behind?
• Irrelevance to students What are the reasons of it, and what can be done about it?
• Lack of meaningful curriculum and standards. • Do potential students have the right image of Informatics
The “Informatics Europe” report [12] further concludes as a field of study, as a science, as a profession?
that the situation in Europe is not much better but vastly • How should the field be positioned so it is “clearly” as
differs from country to country: “the problem is not only the attractive and challenging as many other disciplines and
very limited time for Informatics in the school schedules as perhaps even more so?
compared to e.g. mathematics or physics, but also the lack of Their main recommendations on this matter were the
good standards for content (‘what should we teach’), following:
textbooks (‘how do we describe it’) and didactics (‘how • Offer multidisciplinary and cross-disciplinary programs
should we teach it’) at this level” [12], p.10. Similar to the • Fix computing science’s image
situation in the U.S., there is a lack of teachers for Informatics • Increase women’s enrollment in CS
in Europe, too. • Train high school computing science teachers
Besides the narrow scope of Informatics programmes in • Make CS courses fun.
schools, there is reported to be a pressure on teachers of
Informatics from their fellow teachers, nurtured by fear that IV. DIRECTIONS OF THE REVISION OF INFORMATICS
with the rise of the Internet and ICT, Informatics will present PROGRAMMES
a threat to the established teaching subjects, and also a threat
to the teachers of these [12], p.11. Some of the recommendations on the revival of Informatics
have actually been implemented already for a while in what
C. Conservatism of university Informatics programmes came to be known as a School of Information (SI / i-School).
While the two aforementioned categories of factors With some 60 SIs established worldwide up to date, the
contributing to the decline of Informatics programmes are enrollments of SIs in the U.S. grew 50% since 2001, from
“out of control” for the universities, the revision of study 15.000 to over 20.000.
programs can and should be done to tackle the trend. The concept of the School of Information suggests the
Looking at the CS/Informatics study programmes at directions towards which the CS/MIS disciplines should be
different universities, we see that they hardly make use of the re-oriented. Specifically, there must be a shift from isolated to
potential of Informatics to cover broad field of study, based integrated, from technology-centric or community-centric to
on the view of Informatics as the “Science of Information”. synthesized inter-relational approaches to informational
The computational aspect in Informatics study programmes studies.
is usually addressed to the largest extent, taking into account Having examined the mission statements of a dozen of the
the large share of the programmes covered by programming, leading North American SIs, the aims and problem areas of
algorithm design, algorithm analysis, data management, the study programmes of different SIs can be synthesized in
operating systems, applied computing disciplines. Cognitive two broad categories, these of information-centric and
aspects are also to a certain extent present in the majority of technology-centric (see Figure 1 and Figure 2). The former
the programmes, covered by the mathematical logics, problematization is not that distant from the traditional
artificial intelligence, neural network disciplines. However, Informatics studies, and hence the lessons of SIs can be
the social aspect issue, becoming more and more important in applied to the revival of the discipline.
modern society, is clearly underestimated.
We can observe numerous attempts, trying to overcome the A. What Can Be Done At The School Level
crisis in Informatics studies. A European initiative aiming at While universities cannot exert direct control over the
taking a timely account of the transformation of the field, is Informatics’ teaching at schools, certain success in
“Informatics Europe“ association4, which was created as a stimulating pupils’ interest to the studies of Informatics can
result of the first two European Computer Science Summits be achieved with effort from the universities. Making
Informatics studies “more fun” starting from the school level
could be a way out of crisis. This could be done first of all by
4
http://www.informatics-europe.org/about.php
148 VITKUTE-ADZGAUSKIENE AND FOMIN
adequately exploiting the new trends inside computer science When analyzing the Lithuanian context, we can see, that
itself: “An unprecedented change in our information there are attempts to both add transdiciplinary focus and
processing practices [that] has resulted from the introduction include Mobile Computing topics in the Informatics study
programmes. Most cases observed are related to introduction
• Illuminate the role of information or development of transdisciplinary Informatics programmes,
namely “Business informatics” or “Management
Human
needs • Understand the transformation
• Synthesize the human-centered and
technological perspectives Informatics”, combining knowledge in the Informatics and
Information
• Illuminate both computational and cognitive
processes
Business Management fields. Completely new BSc and MSc
Social/ org.
structures
IS / IT • Encompass community and communication programmes in Business Informatics were established at
• Expand human capabilities through
information Vytautas Magnus University (VMU) and at Kaunas
• Turn data to information, information to Humanitarian Faculty of Vilnius University more than a
knowledge
decade ago. Mykolas Riomeris University added a “Business
Fig.1. Information-centric approach Informatics” BSc programme in 2008, after analyzing its
perspectives in the context of other similar programs in
• Understand how technology interacts with
Business
business processes, policy, and strategy
Lithuania [14].
The admission numbers for the new business informatics
processes
• How information impacts our lives,
organizations, cultures and societies
• Shape policies that influence how people seek, programs show they were popular among students [15], For
IS / IT use, and share information and create example, “Business Informatics” MSc programme at
Strategy Policy knowledge
Vytautas Magnus University collects twice as many students,
compared to the “Applied Informatics” MSc programme. The
overall admissions, however, continue to decline.
Besides the introduction of new specializations in the
Fig.2. Technology-centric approach
Informatics study programmes, new course modules, related
to new business and societal trends can be developed. Here,
and use of the World Wide Web… [is] being accompanied by Mobile Computing comes to play the central role. When
technological developments which are in turn educating talking about mobile communication or ubiquitous computing
successive generations as a by product of social practice”[13]. issues in the university CS curricula, usually two basic
Informatics today is different. It “changes every science opinions prevail. The first of them is stating, that mobile
that it touches”, “it leads the way to many promising new environment is just a subset of a broader electronic
directions in research”. And this new image of Informatics environment, therefore this topic shouldn’t be considered
must be promoted starting at the highschool / college level, separately. Another opinion places the mobile computing in a
and supported by new educational curricula for universities. special position, that of boosting the attractiveness and up-to-
Reports from the U.S. give promising signs on how the task date characteristics of Informatics study programmes at the
of the raising popularity of Computer Science and universities.
Information Systems can be dealt with. One of the reasons for These two approaches to certain extent are reflected in the
the growing interest in computer science degrees in the U.S. situation we can see at different universities, offering
is teens’ excitement about social media and mobile Informatics study programs. Some have special programmes
technologies. The perception that computer science is cool is related to mobile and ubiquitous computing, others have
spawned by all the interesting things on the Web. The iPhone special mobile-related modules included in standard CS
and Web 2.0 reinforces the excitement, and that attracts the programmes, while many remain conservative regarding this
best students [9]. Such excitement could not be not shared by point, maybe reflecting only selected topics in corresponding
the European teens. However, there may be some directing course modules, such as “network management”, “Internet
required on behalf of Informatics’ teachers at schools. technologies”, etc.
Looking at the cases when completely new mobile-oriented
B. How Inclusion of Mobile Computing in Study Programmes study programmes are introduced, we mostly find technically-
Can Help Reverse the Decline oriented programmes, which are usually categorized as
Taking into account the transformations in the belonging to electronics or electrical engineering fields.
environment, the rise of the mobile society, and the Another observed trend here is the orientation towards
excitement of teenagers about the social media and mobile, application design, related to mobile and ubiquitous
we can expect, that by including the topics of Mobile computing – for example, the MSc Program of the Lancaster
Computing in the curricula, universities could increase the university in UK “Mobile and ubiquitous computing”
popularity of Informatics studies. Such an assumption can be (http://www.comp.lancs.ac.uk/study/pg/ubi/). In this case, two
to a large extent justified by the success of the i-School modern computer science trends are combined – the new
concept, teaching programmes of which specifically target the ways of communicating and accessing the Internet, brought
aspect of the interaction of technology with society. Inclusion by the mobile computing revolution, and moving away from
of Mobile Computing would also reflect the desktop PCs to everyday digital devices and smart
recommendations on the emphasizing the transdisciplinarity environments. Such an approach can truly make CS teaching
in Informatics programmes [13], p.5.
CAN MOBILE COMPUTING HELP REVERSE THE DECLINE IN INFORMATICS STUDIES? 149
[6] comScore Inc. (2009c). Prepaid wireless services attract [11] Blum, L., & Cortina, T. J. (2007). Cs4hs: An outreach
increased online interest during economic downturn. program for high school cs teachers. Paper presented at
Retrieved June 18, 2009, from the The 38th SIGCSE Technical Symposium on
http://www.comscore.com/Press_Events/Press_Releases/ Computer Science Education.
2009/6/Prepaid_Wireless_Services_Attract_Increased_O [12] Leeuwen, J. v., & Tanca, L. (2008). Student enrollment
nline_Interest_During_Economic_Downturn and image of the informatics discipline (No. IE-2008-
[7] Andersen, E. (2007, January 24). Why don't cios become 01). Zurich: The Association of Informatics /
ceos? Retrieved May 11, 2009, from Computing Departments and IT Research Institutes in
http://www.espen.com/archives/2007/01/why_doesnt_ci Europe.
os.html [13] White, S. A., & Irons, A. D. (2006). The shape of
[8] The Guardian. (2008, Friday 8 August). Unemployment informatics: A UK perspective. In A. McGettrick & G.
by degree, 2004-07. Percentage of uk and eu domiciled McAlister (Eds.), Informatics Education Europe.
first degree qualifiers assumed to be unemployed by Montpellier, France.
subject area, 2002-03 to 2006-07. The Guardian. [14] Dzemydiene, D. (2008). The need and perspective for
[9] Marsan, C. D. (2009, March 30). Does a computer an interdisciplinary informatics study programme (in
science degree matter anymore? Retrieved May 11, Lithuanian). Mykolas Romeris University.
2009,from [15] Gruževskis, B., Blažien, I., Pocius, A., Dikius, G.,
http://www.networkworld.com/news/2009/040609-hot- Bagdonas, A., Jonauskis, M., et al. (2008). Systematic
tech-skills.html assessment of the correspondence between supply and
[10] Dawson, C. (2009, March 18). Is computer science the demand of high quality (MSc) specialists (in
one smart major in a recession? Retrieved May 11, Lithuanian). Vilnius.
2009, from
http://education.zdnet.com/?p=2271&tag=rbxccnbzd1
Towards an Academic Social Network for Bologna Process
João C. Ferreira, Porfírio P. Filipe
{jferreira,pfilipe}@deetc.isel.ipl.pt
GuIAA – DEETC - Instituto Superior de Engenharia de Lisboa, Lisboa, Portugal
Abstract — The Bologna Process aimed to build a multi-faceted skills. In this scenario, the student proposals
European Higher Education Area promoting student’s for individual studies plan can be submitted and approved
mobility. The adoption of Bologna Declaration directives by the desired educational institution. The proposed system
requires a self management distributed approach to deal with offers a graphical user interface that is autonomous from a
student’s mobility, allowing frequent updates in institutions particular institution. The data acquisition and data
rules or legislation. This paper suggests a computational system submission can be implemented in multiple ways such as
architecture, which follows a social network design. A set of
web services or simple by file transfer protocols.
structured annotations is proposed in order to organize the
user’s information. For instance, when the user is a student its
annotations are organized into an academic record. The
academic record data is used to discover interests, namely II. BOLOGNA DECLARATION
mobility interests, among students that belongs the academic
network. These ideas have been applied into a demonstrator
As the Confederation of EU Rectors’ and the Association of
that includes a mobility simulator to compare and show the
student’s academic evolution.
European Universities (2000) explains it (CRE et al, 2000
p.3-4):
I. INTRODUCTION: BACKGROUND AND MOTIVATION “The Bologna declaration reflects a search for a
common European answer to common European problems.
With Bologna Process (BP), the social and cultural The process originates from the recognition that in spite of
developments impose a revision of the graduate and their valuable differences, European higher education
postgraduate education systems. Students can personalize systems are facing common internal and external challenges
their studies from a diversity of options. related to the growth and diversification of higher
Furthermore, study programmes have become modular, education, the employability of graduates, the shortage of
giving students the chance of more mobility during their skills in key areas, the expansion of private and
studies. Organizational and legal efforts have been taken to transnational education, etc. The Declaration recognises the
guarantee high compatibility of study programmes and value of coordinated reforms, compatible systems and
modules, e.g. the so called BP in Europe [1]. Resulting from common action.
this process students can personalized their studies and can The action program set out in the Declaration is based
compose different educational modules at any institutions to on a clearly defined common goal, a deadline and a set of
a highly personalized curriculum. This process will change specified objectives:
the role of educational institutions, such as universities [2]. A clearly defined common goal: to create a European
They are to be considered as components of a common, space for higher education in order to enhance the
integrated educational space. employability and mobility of citizens and to increase the
The traditional tasks like offering lectures, seminars, international competitiveness of European higher
materials, and infrastructure for education will be education;
supplemented by new tasks, such as personalized A deadline: the European space for higher education
educational programmes. This scenario requires new should be completed in 2010;
information systems and data integration among the A set of specified objectives:
different institutions. - The adoption of a common framework of readable and
In order to support this new academic reality the current comparable degrees, “also through the implementation of
work proposes a self-management distributed approach to the Diploma Supplement”;
deal with student’s mobility, allowing frequent updates in - The introduction of undergraduate and postgraduate
institutions rules or legislation. The management strategy is levels in all countries, with first degrees no shorter than 3
based on annotations organized in academic information years and relevant to the labour market;
records, which extend the student profile. - ECTS-compatible credit systems also covering lifelong
Such system should help current and prospective students to learning activities;
search for suitable course plans or single course units in - A European dimension in quality assurance, with
order to compose a personal curriculum by analyzing their comparable criteria and methods;
- The elimination of remaining obstacles to the free
mobility of students (as well as trainees and graduates) and
151
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_26, © Springer Science+Business Media B.V. 2010
152 FERREIRA AND FILIPE
teachers (as well as researchers and higher education descriptions supplementing Web Services available in a
administrators).” network of cooperating organizations is a remarkable
alternative to known standardisation issues. Instead of
III. THE SOCIAL NETWORK developing an inflexible standard API and producing a huge
The Social Web represents a set of web sites and overhead concerning its support and evolution, the service
applications in which user participation is the primary driver provider agree to reference the same domain ontology in
of value. The principal of such systems is well described by their Web Service descriptions. The ability of machine read
Tim O'Reilly [3] and this development is around Web 2.0 such descriptions, makes possible the automatic search of
[4] principals. Examples of these types of initiatives are required services. Therefore the interoperability of services
Wikipedia, MySpace, YouTube, Flickr, Del.icio.us, within the targeted network would be achieved. In this work,
Facebook, and Technorati. Discussions of the Social Web one will basically developed reference ontology and the
often uses the phrase "collective intelligence" or "wisdom of generation of a set of agent applications needed for the
crowds" to refer to the value created by the collective search, invocation and orchestration of Web Services
contributions of all these people writing articles for offered by educational institutions.
Wikipedia, sharing tagged photos on Flickr, sharing Semantic web architecture is a functional, non-fixed
bookmarks on del.icio.us. Knowledge sharing have great architecture [5]. Barnes-Lee defined three distinct levels that
potential due to the diversity of information available, incrementally introduced expressive primitives: metadata
diversity of individual cultures and the low costs of layer, schema layer and logical layer [6]. XML and XML
gathering and computing over big network (Internet). schema define syntax, but mean nothing about the data that
Currently we have millions of persons offering their it describes. That means that some standards must be built
knowledge online, which means that the information is on top of XML that will describe semantics of data. This
stored, searchable, and easily shared, so the big challenge is conduct to RDF and a general model in metadata layer RDF
to find the right match between what is available online and schema, which provides mechanisms for defining the
the approach to take the right/useful information from this relationships between properties declared and others
data. A collective intelligence/knowledge can result from the resources. To enable services for the semantic web, one has
data collected from all those people that are aggregated and the logical layer on top. This layer introduces ontology
transformed to create new knowledge. This collaborative languages, that are based on meta-modeling architecture
approach had the spam problem and other fraudulent sources defined in lower layer. This enables the use of tools with
in the mix, but the majority of users have a proper behavior generic reasoning support, independent of the specific
and can handle these disrupt behavior [3]. problem domain. Examples of these languages are OIL
So, our proposal is collaborative system based on a social (Ontology Inference Layer), DAML (DARPA Agent
network design that is able to connect not only at your own Markup Language), OWL and AOBP. OWL is a semantic
institution but to others within the European Higher markup language for publishing and sharing ontologies in
Education Space. This collaboration is achieved by users the web. OWL is developed as a vocabulary extension of
annotations divided into two main description fields: (1) RDF and is derived from DAML and OIL [7, 8].
semantic annotation, descriptions using terms from system
ontology; (2) descriptions performed by free descriptions
Logical Layer
terms introduced by the users. All these stakeholders can OWL
AOBP
provide semantic annotation by a specific metadata Schema Layer RDF Schema Semantics
Transformation Rules
A. Academic Record
The Academic Record (AR), referenced in Figure 3, can
aggregate information items in several languages, namely:
(1) Extra Curricular Course; (2) Qualification; (3) A Figure 4: Course Plan diagram
Contract of Studies (is celebrated between the student and
the institutions to formalize the terms of the temporary
mobility); (4) Work Experience; and (5) Student Data.
The AR includes all student paths during the formation
period. There is no need of often cross checks between AR
and the information that is kept private by the process
stakeholders. For each, information unit updated in the AR,
the student should send, if needed, a confirmation document
to the institution where he belongs. For instance, if the
student completes a professional formation he should update
his AR and after he will deliver the right certificate to his
higher institution education that holds the AR. When the
students do course change or intuition transfer, a copy of the Figure 5: Individual Studies Plan diagram
AR is produced and must be detained by the destination D. Diploma Supplement
institution.
Diploma Supplement follows the model developed by the
European Commission, Council of Europe and
UNESCO/CEPES. The purpose of the supplement is to
provide sufficient independent data to improve the
international “transparency” and fair academic and
professional recognition of qualifications (diplomas,
degrees, certificates etc.). It is designed to provide a
description of the nature, level, context, content and status of
the studies that were pursued and successfully completed by
the individual named on the original qualification to which
this supplement is appended. It should be free from any
Figure 3: Academic Record diagram value-judgments, equivalence statements or suggestions
about recognition. Information in all eight sections should
be provided. Where information is not provided, an
B. Course Plan explanation should give the reason why. Diploma Suplement
The Course Plan characterizes a specific graduation plan by diagram is a detailed description of the activities developed
the means of describing the Figure 4: (1) Curricular by the student in the graduated process based on the follows
Structure; and (2) the Studies Plan. attributes: employee, student data, course unit, qualification,
The Curricular Structure represents the distribution of ECTS scientific area and associated language.
credits, minimum and maximum, of each scientific area. The This UML representation can transform to a RDF; (2)
goal of the Curricular Structure is to define the amount of missing values are replaced with default values when
ECTS, because of the optional course units. The Studies possible, and (3) the RDF is validated against the metadata
154 FERREIRA AND FILIPE
schema and other validation rules. Each metadata producer A. Enhancing an Existing CMS with Bologna Ontology
gets a report of warnings, errors and other problems that Most content providers in SIR use a CMS for authoring,
were encountered during harvesting and validating the publishing and archiving content on their website. A typical
content. If some parts or all of the metadata is unacceptable CMS supports creation of textual metadata about
due to serious errors, the metadata is discarded until documents, such as title and publication time, but not
necessary corrections are made. Otherwise, the metadata is classified annotations. This would require that the system
added to and published in the Semantic Information has functionalities supporting ontology-based annotation
Repository (SIR), see section 6. work, e.g., concept search for finding the relevant concepts
(identified with URIs), concept visualization for showing the
concept to the user, and storing concepts along other
information about the documents. The CMS should also be
able to export the metadata, preferably in RDF format, to be
used by semantic web applications.
Figure 6: Diploma of Supplement diagram Currently, classification systems are typically shared by
downloading them, and each application must separately
E. Processing Rules implement the classified defined schema. To avoid
The skills concept <www.disco-tools.eu> allows the duplicated work and costs, and to ensure that the classified
establishment of a semantic link between: (1) Extra systems are always up-to-date, we argue that one should not
Curricular Course; (2) Qualification; (3) A Contract of only share the classified system, but also the functionalities
Studies; and (4) Work Experience. for using them by centralized mash-up component services.
A specific course plan is completed when the hosting Such services, e.g. Google Maps, have been found very
institution recognizes that the student has achieved all the useful and cost-effective in Web 2.0 applications for
skills that belonging to a course plan. integrating new functionalities with existing systems.
The skills is a matching unit that allows comparison We have applied the idea of using mash-ups to provide
information items that aggregates its owns skills. A trial classified system services for the content producers of SIR
example is the comparison of course unit against course creating the ONKI Ontology Server framework [9]. ONKI
unit. However it is possible to compare course units against provides ontological functionalities, such as concept
work experience or qualifications. This matching can be searching, browsing, disambiguation, and fetching, as ready-
based on historical tables, recommended tables or even to-use mash-up components that communicate
requires human supervision. asynchronously by AJAX (or Web Service technologies)
with the shared ontology server. The service integration can
VI. BOLOGNA APPLICATION SUITE be easily done by changing only slightly the user-interface
component at the client side. For example, in the case of
AJAX and HTML pages, only a short snippet of JavaScript
code must be added to the web page for using the ONKI
services.
The main functionality addressed by the ONKI UI
components is concept finding and fetching. For finding a
desired annotation concept, ONKI provides text search with
semantic auto completion [10]. This means that when the
annotator is typing in a string, say in an HTML input field of
a CMS system, the system dynamically responds after each
Figure 7: BAS architecture system. input character by showing the matching concepts on the
The Bologna Application Suite, is our proposal for a Web ONKI server. By selecting a concept from the result list, the
2.0 system environment and major components are concept’s URI, label or other information is fetched to the
described at Figure 7. A major challenge in the distributed client application.
content implementation of Bologna process is how to Concept browsing can also be used for concept fetching.
facilitate the cost-effective production of descriptive, Study plan only changed in specific time periods, so the
semantically correct high quality metadata. In SIR three content change can be handling in a better way. In this
ways of creating metadata are considered and supported: (1) model, the user pushes a button on the client application that
Enhancing existing web content management systems opens a separate ONKI Browser window in which
(CMS) with ontology mash-up service components for annotation concepts can be searched for and browsed. For
producing semantic metadata; (2) Using a browser-based each concept entry, the browser shows a Fetch concept
metadata editor for annotating web content; (3) Automatic button which, when pressed, transfers the current concept
conversion of metadata. These approaches are outlined information to the client application. ONKI supports
below. multilingual ontologies, has a multilingual user-interface,
supports loading multiple ontologies, and can be configured
extensively.
TOWARDS AN ACADEMIC SOCIAL NETWORK FOR BOLOGNA PROCESS 155
ONKI is implemented as a Java Servlet application running System is running for 3 months and we are now getting the
on Apache Tomcat. It uses the Jena semantic web first results, with the start of a new academic year. The
framework for handling RDF content, the DirectWeb System prove to use well past experience, i.e. once a course
Remoting (DWR) library for implementing the AJAX unit equivalence is performed from the course x to the y, the
functionalities, the Dojo JavaScript toolkit, and the Lucene others request are automatic processed. This reality helps a
text search engine. lot because most changes occur from a small percentage of
10 courses. At this moment we have register a community of
B. Browser-based Metadata Editor
100 users (small number) but we think this number will be
Some SIR content providers can not add mash-up ontology increase with the usage of the system.
support to their CMS due to technical or economical The users when access the system, they need to choose the
reasons. Furthermore, some content providers do not even institution through a menu, see Figure 8. They choose first
have a CMS or they may not have access to the CMS that the country, then the city and finally the institution. Since
contains the content, e.g., if the content originates from a we did not developed the security module the application is
third party. To support metadata productions in these cases, only available in ISEL network, but the BAS system will be
we have created a centralized browser-based annotation prepared for a European usage. Bologna Process makes
editor SAHA [11] for annotating web pages based on sense if a big network is constructed around it. Login at
RDF/OWL-based ontology as a schema (collection of BAS can be performed by a Figure 9 menu. If they are new
classes with a set of properties, described on section 4). An users they can create new information or import academic
annotation in Saha is an instance of a schema’s class that register from other institution.
describes some web resource and is being linked to it using
the resource’s URL (in same cases, URI). We make the
distinction between the annotation of a document (e.g. a web
page) and the description of some other resource (e.g. a
person) that is somehow related to the document being
annotated. In addition to containing classes used to annotate
documents (annotation classes), an annotation schema used
with Saha can also contain reference classes for describing
resources other than documents. In other words, an
annotation schema forms a basis for the local knowledge
base (KB) that contains descriptions of different kinds of
resources that may or may not exist on the web. Instances of Figure 8: BAS System Menu to choose the academic
the reference classes are used as values of properties in institution.
annotations. The schema element fields in SAHA can be
connected with ONKI ontology service components,
providing concept finding and fetching services to the
annotator, as discussed above. In practice, an annotation
project is Jena’s ontology model stored in a database. A Figure 9: User login menu: (1) for a new user; (2) to
model is comprised of the annotation schema and the import the Academic Register from other institution.
instances of the schema’s classes. It can be serialized to Due to space limitation is not possible to give all details
RDF/XML in order to use the annotations in external about the application. For a complete description see [12],
applications. we will same detail about collaboration issues, as illustrated
C. Automatic Conversion on Figure 10 and 11. In the Figure 10, is the menu process
The third content producing method in SIR is automatic associated with the creation of a new qualification. The
conversion of original data to SIR metadata. This method is number two is another collaboration tag where you can
currently used in cases where metadata exists in a CMS, but make a note to a course simply by being classified and
it is in an incompatible format, does not contain ontological selected the desired course.
annotations (URIs) and/or some minor information is Users can create annotations related with transfer process or
missing in the metadata. simply comments about the subject. In Figure 11 we show
an annotation process regarding course unit equivalence.
VII. CASE STUDY Predefine information field’s helps the reuse of the
information by the users.
We perform a real case implementation of our BAS system
to a polytechnic institute of Lisbon (ISEL). ISEL have more A. Simulation of Course transfer
than 5000 students distributed by 7 engineering degree Students simulate a transfer or change of course.
(licenciatura in Portuguese) and 6 Masters. Bologna reform A prerequisite to perform this operation is the creation of an
was introduced at ISEL in the academic year of 2006/2007. individual study plan. To perform the simulation simply
The BAS system was proposed to help academic process select the individual plan of study desired travel destination
regarding students changes (inside institution and outside). and the location will be generated an XML document
156 FERREIRA AND FILIPE
created with all the information of the simulation. The classified system rather than keywords; (3) The problem of
results are represented in a Simulation Menu (Figure 12), outdated and missing links is eased by providing the end-
where the student can see the results of equivalences if he user with semantic recommendations that change
applies that change. dynamically as content is modified; (4) Content aggregation
is facilitated by end-user facets that collect distributed but
related information from different primary sources; (5)
Quality of content is maintained by including only
trustworthy organizations as content producers (academic
institutions); (6) share suggestions and comments of
students regarding the courses.
At the same time, the problems Bologna Mobility process:
(1) Redundancy of content creation can be minimized by the
possibility of aggregating cross-portal content; (2) Reusing
the global content repository is feasible. By using a central
SIR, external applications, can reuse the content; (3) Internal
and external link management problems are eased by the
dynamic semantic recommendation system of the portal and
the content aggregation mechanisms; (4) The tedious content
Figure 10: Creating a new qualification. indexing task is supported cost-effectively by shared
ontology service components; (5) Metadata quality can be
enhanced by providing indexers with ontology services by
which appropriate indexing concepts can be found and
correctly entered into the system.
REFERENCES
[1] CRE et al, 2000. Confederation of EU Rectors’ Conferences and the
Association of European Universities (CRE), 2000. The Bologna
Declaration: an explanation.
[2] B..J. Krämer (2003), CUBER für die Ausbildung á la carte, in: G.
Simonis (ed.), Lernort Bologna Declaration Communiqué of the
meeting of European Ministers in charge of Higher
Education,Prague,May 19th 2001, ec.europa.eu/education/policies/
Figure 11: Example of a anotation performed by a educ/bologna/bologna.pdf.
student using the descriptors available on the ontological [3] T. O'Reilly. What Is Web 2.0: Design Patterns and Business Models
space. for the Next Generation of Software, September 30, 2005.
http://www.oreillynet.com/pub/a/oreilly/tim/news/ 2005/09/30/what-
is-web-20.html>.
[4] Web 2.0 Conferences. http://www.web2summit.com/ Wensing, M.,
Grol, R. and Smits, A. (1994) Quality judgements by patients on
general practice care: a literature analysis. Social Science and
Medicine 38, 45±53.
[5] Tim Berners-Lee, Weaving the Web, Orion Business Books, London,
1999.
[6] van Harmelen, F., et al, “OWL Web Ontology Language Reference”,
W3C Working Draft, http://www.w3.org/TR/2003/WD-owl-ref-
20030331/, 2003.
[7] Miller, J., Mukerji, J. (eds.), “MDA Guide Version 1.0”, OMG
Document:omg/2003-05-01,
[8] Paulo Trigo and Helder Coelho. Agent inferencing meets the semantic
web. In Progress in Artificial Intelligence, EPIA-09, volume 5816 of
Lecture Notes in Artificial Intelligence (LNAI), pages 497-507.
Springer-Verlag, 2009.
[9] Handschuh, S., Staab, S. and Ciravogna, F., S-CREAM -- Semi-
Figure 12: Simulation Menu automatic CREAtion of Metadata in SAAKM from Web Pages in First
International Conference on Knowledge Capture, (2001).
VIII. CONCLUSIONS [10] Hyvönen, E., Viljanen, K., Mäkelä, E., et al.: Elements of a national
semantic content infrastructure—case Finland on the semantic web.
This paper addressed the needs of Bologna Mobility In: Proceedings of the ICSC 2007, IEEE Press (2007).
process, with a proposed of BAS system in a Web 2.0 [11] Valkeapää, O., Alm, O., Hyvönen, E.: Efficient content creation on the
semantic web using metadata schemas with domain ontology services
environment to: (1) Support information exchange; (2) (system description). In: Proceedings of the ESWC 2007, Springer–
Content finding is supported by cross-portal semantic Verlag, Berlin (2007).
search, based on hierarchical structure imposed by a [12] Paulos, J. and Costa, R. Bologna Application Suite. Final Year project
available at www.deetc.isel.ipl.pt/matematica/jf/bas.pdf.
Supporting Knowledge Discovery in an eLearning
Environment Having Social Components
Paola Monachesi and Thomas Markus Vlad Posea and Stefan Trausan-Matu Petya Osenova and Kiril Simov
Utrecht University Politehnica University of Bucharest Bulgarian Academy of Sciences, Sofia,
The Netherlands Romania Bulgaria
Abstract – One of the goals of the “Language Technology for knowledge, as concluded in [1]. In our specific case we use an
LifeLong Learning” project is the creation of an appropriate ontology in the domain of computing, which is employed to
methodology to support both formal and informal learning. provide deep annotation of the learning materials that should
Services are being developed that are based on the interaction facilitate their understanding and reuse.
between a formal representation of (domain) knowledge in the
form of an ontology created by experts and a social component
However, the ultimate goal of the LTfLL project is to
which complements it, that is tags and social networks. It is complement the formal knowledge represented by ontologies
expected that this combination will improve learner interaction, developed by domain experts with the informal knowledge
knowledge discovery as well as knowledge co-construction. emerging from social tagging in order to provide more flexible
and personalized ontologies that will include also the
I. INTRODUCTION knowledge of communities of users. The enhanced ontology
will be connected to the social network of the learners through
In a Lifelong Learning context, learners access and process the tags that they provide, improving thus the possibility of
information in an autonomous way. They might rely on formal retrieving appropriate material and allowing learners to connect
learning, that is they might focus on textual material approved to other people who can have the function of learning mates
by content providers in the context of a course developed by an and/or tutors.In this way, it will be possibleto provide a more
organization or institution. They might also want to rely on personalized learning experience able to fulfill the needs of
informal learning, that is on (non-)textual content available different types of learners.
through the web which is uploaded and accepted by the The paper is organized as follows. In section II, we discuss
community of learners and not necessarily by a content the differences between formal and informal learning. Section
provider of an institution. III presents the design of the Common Semantic Framework,
One of the objectives of the Language Technologies for while section IV focuses on its implementation. In section V,
LifeLong Learning (LTfLL)1 project is to develop services that we present an initial evaluation of the system while VI
facilitate learners and tutors in accessing formal and informal introduces some concluding remarks.
knowledge sources in the context of a learning task. More
specifically, the aim is to support learner’s interaction in order II. FORMAL VS. INFORMAL LEARNING
to facilitate knowledge discovery and knowledge co-creation.
To this end, a Common Semantic Framework (CSF) is being As already pointed out above, in the LTfLL project, we are
developed. The CSF allows the stakeholders to identify, developing services that facilitate learners and tutors in
retrieve and exchange the learning material. More specifically, accessing knowledge sources within the context of a certain
it supports formal learning by addressing course material which learning task. The learning process is considered an attempt to
includes textbooks, articles, slides as well as informal learning make the implicit knowledge which is encoded in the learning
which we identify with (non-)textual material emerging from materials, explicit.
social media applications. Within the CSF, communication is In the LTfLL project, we aim at supporting both types of
facilitated through the use of social networks and new learning - formal and informal one. Under ‘formal learning’ we
communities of learners are established through the understand the process, in which the learners follow a specific
recommendations provided by the system. In order to provide curriculum. For each topic in the curriculum, there is a set of
recommendations, the user’s profile, his interests, his learning materials, provided by the tutor. In contrast to this, the
preferences, his network and obviously the learning task are curriculum in the informal learning process is not obligatory,
taken into account. It is through the access to formal and and the role of the tutor either is absent or is not obvious. The
informal material that new knowledge will originate. learners exploit different sources to succeed in their learning
Knowledge repositories are employed to achieve the above goal, very often relying on the information, available on the
goal. An appropriate way to retrieve formal content might be web, and on the social knowledge sharing networks.
by means of an ontology which can support the learner in the From the perspective of formal learning (i.e. guided,
learning path, facilitate (multilingual) retrieval and reuse of monitored one), learners rely mainly on material which has
content as well as mediate access to various sources of been prepared by the tutor on the specific topic and is
addressed by the curriculum. Learners are expected to access
1 certain pieces of pre-stored information, which would lead
http://www.ltfll-project.org/
157
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_27, © Springer Science+Business Media B.V. 2010
158 MONACHESI ET AL.
them to the required knowledge. In the informal learning logging each day to Facebook.com.3
process, learners have to locate knowledge sources on the web, The communities that emerge through social media
then to select the relevant ones, and finally to investigate each applications such as Delicious, Flickr or YouTube provide two
of them. During this process, learners often need guidance and important elements that can be integrated in the Common
thus they might profit from finding appropriate people (peers) Semantic Framework, in order to support informal learning.
on the web. They provide the “knowledge of the masses” in the form of
The two learning paradigms are complementary from the tags that are being used to annotate resources. On the other
lifelong learner’s perspective. In the LTfLL project, we are hand, the structure of these communities which can be
developing two sets of services that would support both - the represented through a social network can be employed to
formal style of learning and the informal one. These services recommend content and peers. The social networks
are integrated within the Common Semantic Framework. automatically introduce a notion of trust in the search results in
relation to the user's social network.
III. COMMON SEMANTIC FRAMEWORK: DESIGN The knowledge produced by communities in the form of tags
can be employed to enrich existing domain ontologies semi-
One of the aims of the LTfLL project is to build an automatically. More specifically, we have merged the dynamic
infrastructure for knowledge sharing which we call Common knowledge provided by users/learners through tagging with the
Semantic Framework (CSF). It allows for identification, formal knowledge provided by the ontologies by adding
retrieval, exchange and recommendation of relevant learning tags/concepts (or instances of concepts) and relationships
objects (LOs) and of peers. It is ontology driven allowing thus between concepts in the domain ontology. In the CSF, the
for a formalization of the knowledge arising from the various connection between words/tags and concepts is established by
stages of the learning life-cycle. Domain ontologies offer means of language-specific lexicons, where each lexicon
useful support in a learning path since they provide a specifies one or more lexicalizations in one language for each
formalization of the knowledge of a domain approved by an concept. DBpedia [5] is used as a knowledge base to resolve
expert. In addition, ontologies can support formal learning the extracted lexicalization to unambiguous existing concepts.
through the annotation of the learning objects [1]. DBpedia doesn’t provide much information about relations
More specifically, with respect to formal learning, we are among concepts when compared to specially crafted domain
developing a Formal Learning Support System (FLSS) that ontologies, but compensates this shortcoming with the huge
supports the individual work of the learners and tutors in number of available concepts.
manipulating the learning objects, including the knowledge There is an important side effect of this ontology enrichment
which is implicitly encoded in them and in adding their own process: if tags given by learners or emerging from social
descriptions to these learning materials. In particular, an media applications are related to the concepts present in the
annotation service has been implemented which is meant as a ontology, we manage to include not only the expert view of the
way to make the implicit knowledge attested in learning domain, but also the learner’s perspective. In fact, domain
objects, explicit. Thus, through the annotation service, we aim ontologies developed by knowledge engineers might be too
at relating the knowledge, encoded in the text, to the formally static, incomplete or might provide a formalization that does
represented knowledge in a domain ontology. not correspond to the representation of the domain knowledge
Despite the possible limitations of the ontologies, they available to the learner which might be more easily expressed
provide a highly structured way of interlinking different by the tagging emerging from communities of peers via
knowledge sources. From the point of view of the learning available social media applications.
process itself, the domain ontology can be viewed as a way to It is important to help the learner manage the content that is
produce a concise representation of the most important created daily around him in his learning communities. The
concepts in a given domain. Consequently, the learner has to knowledge produced inside the learner’s network is very
interact with them in order to assess his current level of valuable to him. According to the Communities of Practice
familiarity with the domain or to use them to acquire new (CoP) theory [6], a learner acquires knowledge, as he moves
knowledge. from the periphery to the center of a CoP.
While in formal learning, courses and learning objects play a In order to facilitate the informal learning experience, we
relevant role, in informal learning, communities arising from automatically monitor the learners’ activities and his peers on
social media applications are becoming more and more the social network applications. The data that the learner and
dominant, but are poorly integrated in current learning his network create or annotate is indexed in a semantic
practices. Learners prefer to use the internet to search for an repository. This data can then be used to offer learning support
answer rather than asking a tutor or a colleague [3]. Most of the which exceeds that which is currently offered through keyword
documentation for a given learning task is found on the internet search or simple recommendations.
and most learners do not check if the information is accurate or In this way, the CSF supports self-organization and the
reliable [4]. Recently, the usage of the internet has shifted emergence of collaborative knowledge and classification. In
towards social networking applications. 1 billion movies are addition, we aim at connecting learners to other learners in an
seen everyday on YouTube.com2 and 150 million users are
3
http://www.facebook.com/press/info.php?statistics
2
http://youtube global.blogspot.com/2009/10/y000000000utube.html
SUPPORTING KNOWLEDGE DISCOVERY IN AN ELEARNING ENVIRONMENT 159
developed in the "Language Technology for eLearning" project we have implemented two types of search, that is ontology
(cf. [1]). It contains 1002 domain concepts, 169 concepts from based search that exploits the structured information attested in
OntoWordNet and 105 concepts from DOLCE Ultralite. the ontology as well as search based on tags and social
Candidates are generated by applying similarity measures or networks that exploits the social aspects of the knowledge
other sorting mechanisms to derive socially relevant tags. provided. Our plan is to integrate these two types of search into
Implemented examples include coocurrence-based and cosine- a recommendation system. It should exploit the strong features
based measures [9]. This analysis drives lookups and selective of both types of search, that is ontological structure and
integration from DBpedia for the domain ontology enrichment. community based knowledge (i.e tags and social networks).
The various web services provide additional lexicon entries, Ontology based search
socially relevant specific classes linked to more abstract classes The semantic search can be triggered by a list of keywords,
already present in the ontology and additional relationships provided by the user. The query string is analyzed on the basis
discovered between already existing domain concepts. The of a lexicon, mapped to the domain ontology which is enriched
system provides an automatically enriched ontology which with tags. The result of this analysis is a set of concepts from
contains the vocabulary of the Community of Practice that the the ontology. The concepts, related to the keywords, are
user is part of. More specifically, the resulting ontology additionally processed on the basis of the information from the
integrates the socially relevant concepts, but with the structure ontology by adding related concepts, that is query expansion.
of an expert view domain ontology. Extraction methods Query expansion is done via reasoning over the ontology. The
exclusively focused on deriving ontology-like structures from most typical case is the addition of the sub-concepts to the ones
folksonomies cannot provide such a high quality of results due extracted from the ontology on the basis of the keywords. The
to unavailability of the implicit knowledge in folksonomies initial set of concepts is interpreted conjunctively and the
which has been made explicit in domain ontologies. addition from the query expansion is added disjunctively to it.
Multimedia Document Annotation The query, formed in this way, is evaluated over the
The aim of the multimedia document annotation is to repository of annotated documents. The result is a list of
annotate explicitly learning material with concepts from the documents and in the case of the annotated ones it is a set of
domain ontology previously mentioned. The purpose of this pointers to parts of the document which are considered as
annotation is to allow semantic search, based on the ontology. relevant to the query - sentences, paragraphs and pictures. The
The usage of the domain ontology in the search provides possibility of ontology browsing is also offered to the learner.
abstraction from the concrete wording of. The annotation is More specifically, the enriched ontology can be exploited by
done in the document. This type of annotation allows the user the learner to get an overview of the domain specifically
to be directed to a particular part of the document. tailored to some search query, relevant user or specific
Within the CSF both text annotation and image annotation document. This allows the learner to discover resources which
are possible. The text annotation is implemented as a language do not match previously known keywords by supporting
pipe performing tokenization, POS tagging, lemmatization, discovery of new concepts and associated resources through a
semantic annotation and coreferential annotation. The first visualization of the domain.
three steps rely on available state-of-the-art tools. The semantic Search based on social networks and tags
annotation is performed by a concept annotation grammar In addition to ontology based search and browsing we have
constructed on the basis of an ontology-to-text relation as implemented an alternative type of search based on the
described in [13] and [14]. The coreference annotation module structure of the social network and the tags provided to
provides mechanisms to make the concept annotation more resources.
precise and with a better coverage of text (15% improvement). More specifically, in order to return trusted results, the
The image annotation is performed manually. It is search service focuses on documents created or bookmarked by
complementary to the text annotation. The reason to include it the learner’s friends and close contacts. The results found in
is that very often the pictures and diagrams, etc, present in the this way also contain the name of the users who created or
learning documents contain important information for the recommended them. Being in the learner’s network, he can
learner. The user interface consists of a document browser, now contact them and ask for more information or simply trust
image editor and ontology browser. The user selects an image the document returned more as it has the guarantee of a peer
from the document browser. The image is opened within the recommendation. The resources found this way will most
image editor. Then the user can select a region of the image likely be in the learner’s zone of proximal development [10].
and assign one or more concepts to the region. The The search algorithms can also identify the most relevant
interpretation of this type of annotation is that the region peers for a given task. The learner can interact using chat or e-
depicts an instance of the corresponding concept. mail with the peer to ask further questions and the peer might
The document annotation is stored in the semantic repository very well be helpful. He is already part of the learner’s social
together with links to the documents which will be used for network and we suppose that this means there is already a
document search by the ontology search engine. connection established between them.
Search The search services use the FolkRank algorithm [8]. This
The search service takes the learners’ query as input and algorithm is used because it can retrieve any type of content
analyzes the information stored in the semantic repository in that has been previously tagged.
order to provide relevant and trusted results. At the moment, A similar tag-based algorithm is used to search documents
SUPPORTING KNOWLEDGE DISCOVERY IN AN ELEARNING ENVIRONMENT 161
on the basis of the tags provided by users [11]. The algorithm In the first use case, the scenario employed for the validation
computes results based on tag co-occurences, on tag-resource focused on providing support to tutors in the creation of
and on tag-user affinities. These affinities are computed based courses in the IT domain. The idea was to evaluate the
on the frequency of the user-resource and tag-user pairs. The efficiency of the computer-assisting tool when creating a
tag co-occurence is the most used measure for tag similarity course on a certain topic. This evaluation focused on the
and it measures how often 2 tags are used together. After performance of the various functionalities, reflected in users’
computing the co-occurence matrix and the affinities matrices, opinions. The outcome would give feedback to the developers
the algorithm clusters tags and computes an affinity between a on how to improve the supporting system. The evaluated
user and a cluster of tags. Finally, using the affinity between functionalities were the following: relevance of the learning
user and the tag clusters the algorithm returns a number of material corpus wrt a domain topic; relevance of the retrieved
resources of interest. material; suitability of the ontology for structuring a lecture in
Visualization the chosen domain; possible visualization of the interface;
The visualization module offers a graph based view of either combination between ontology browsing and semantic search.
a social network of users and resources centered around a given There were two target groups: five teachers and three
learner or a part of a domain ontology. managers (vice-deans at the Faculty of Slavonic Languages,
The social network view can present the whole network as it Sofia University). Our assumptions were as follows: the
exists in the repository or it can show only the part of the combination among semantic search, ontology browsing and a
network returned by a search query. The search results are concept map based visualization would support the tutors in the
presented as a network of users and resources attached to them. processes of finding the relevant materials, structuring the
The user can thus see which path links him to a specific user or course and representing the content in a better way. Our
resource on the given search topic. In this way, the user can hypothesis was that, on the one hand, the tutors might differ in
identify the best means to contact a user for further questions their preferences of using the functionalities. For some tutors,
and can also rapidly identify who are the most competent users search would be more intuitive, while others might prefer
on a specific topic. The unified relations represented structure (in our case - ontology). On the other hand, the
graphically are from peers across different social networking managers would look at the facilities from a different
applications because our system monitors and integrates all the perspective, such as how this supporting system would
social media applications that the learner is using. communicate with the well-known Moodle, for example. Two
The ontology based graph allows a user to quickly gain methodological formats were adopted: the think-aloud strategy
insight into the important relations which exist between domain for the tutors, and the interview for the managers. The think-
concepts. The ontology visualization is a hyperbolic aloud format included seven questions, related to the specific
hypergraph that contains domain concepts with their preferred functionalities (search, browsing, visualization) as well as the
lexicalisation and (taxonomic) relations to other domain general usability and problematic issues of the system. The
concepts that are drawn as labeled edges. The hyperbolic interview format included three questions, which were related
geometry allows the user to focus on the important topics while to the usability and management issues related to the adoption
still retaining a sense of the larger surrounding context. The of the system.
user can discover, previously unknown, concepts which would The results confirmed our hypotheses. Additionally, four
not have shown up either in the social network or through a teachers reflected in their answers the fact that the
keyword search, by interactively browsing the ontology. functionalities are more efficient when operating in some
Together the visualization services provide a way for the interconnected mode. Regarding the semantic vs. text search
user to navigate through the knowledge sources, focus on three tutors gave preference to the semantic one, while two
specific details, add new information, take notes, etc. were more cautious and balanced between the two.
The visualization can be customized to be included in a Visualization was the most commented functionality, because
number of widgets to show specific information like the teachers had different ideas on how it should be controlled.
visualizing a part of the ontology, the social network, a The main worry of the managers was the requirement of a
definition, relevant peers, etc. These widgets can easily be training course to be able to use the CSF.
recombined or disabled in order to provide a learning In the second use case, we have focused on the support
environment that best fits the user. provided by an ontology enhanced with tags in comparison
with clusters of related tags extracted from Delicious (lacking
V. EVALUATION ontological structure) in the context of a learning task. The
underlying assumption being that conceptualization can guide
The Common Semantic Framework was evaluated on the learners in finding the relevant information to carry out a
basis of three use cases: learning task (a quiz in our case). The hypothesis was that
1. tutor support in the creation of a new course; learners might differ with respect to the way they look for
2. comparison between browsing of an ontology information depending on whether they are beginners or more
enriched with social tags and a cluster of related tags advanced learners. While beginners might profit from the
in order to solve a quiz; informal way in which knowledge is expressed through tagging
3. recommendation of content and peers on the basis of a more advanced learners might profit from the way knowledge
social network and tags. is structured in an ontology. The experiment was carried out
with six beginners and six advanced learners, all with an
162 MONACHESI ET AL.
163
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_28, © Springer Science+Business Media B.V. 2010
164 SILVA ET AL.
construction industry in Brazil also contribute to the lack of Figure 2 below shows the percentages of companies that
information about fabrication processes available in this have only conventional technologies and those with CNC
country. machines. The percentage of firms with CNC equipment is
Therefore, the objective of this paper is to demonstrate significant and supports our hypothesis. It is important to
that the thesis of CNC technology unavailability as the main stress that 0.9 CNC machines per company in the surveyed
cause of it not being used in architecture is unsustainable. We geographic area is close to the national average detected by
will also show that some experiments in architectural [11]. This result is particularly important considering that the
education involving rapid prototyping and digital fabrication Federal District in Brazil is an area predominantly
may provide a lead towards the solution of this problem. administrative where only in the last 20 years significant
industrial activities have emerged.
IV. RESEARCH METHOD AND RESULTS
According to the 3rd Brazilian Inventory of Metal
Forming and Cutting Machines [11] the total number of
equipments in this area increased in 2.7% from 2006 to 2008.
During the same period, the number of CNC machines
increased in 44.4%. This suggests a technological potential in
fast expansion which can be better explored.
A. CNC availability in Federal District construction industry
Most of the data in the aforementioned Inventory (96,5%)
was gathered from the south and southeast regions of Brazil.
In this paper we show the results of a survey with the
construction industry in the specific region of the Brazilian
Federal District. Companies of small, medium and large size
and with different functions in the construction industry were Figure 2. Percentage of companies with CNC technologies.
(Source: authors of this paper)
surveyed.
Figure 1 shows the percentage of participation of different Figure 3 shows the percentage of companies that have
types of industries in the total of firms surveyed. Several CNC technologies taking into consideration only metal
companies were approached to establish a reasonably frames and structures.
representative sample. Fourteen of those firms accepted to
participate in the survey. The metal structure companies are
the majority in the sample. This is due to the fact that this
sector is the most dynamic of the construction industry in the
process of incorporating CNC technologies. This choice was
made under the assumption that this sector should influence
others in the long term. As an example, the steel structure
companies, through the manufacturing of metal molds, should
influence concrete construction systems.
customization (and the benefits it may provide to the rising of potential. However, one of the surprising results was that
new forms in architecture) and rapid prototyping. many students started designing taking into consideration the
The experiment was developed during four semesters and possibilities of the aforementioned paradigm shift even before
rapid prototyping was introduced at the end of the third one, we introduced rapid prototyping and they had started seeing
on a limited range. Most students were requested to design its practical potential.
using handmade scaled models. The design projects with
more complex geometry were selected for rapid prototyping V. CONCLUSIONS
in a three-axe milling machine.
Figure 8 shows a rendered view of one of the students The results of our survey show that thesis of technological
design project for the exhibition pavilion. unavailability is not sustainable and that the presence of CNC
technology in the Federal District of Brazil is already
significant, although its distribution may be considered
heterogeneous.
Our experiments with CNC at early architectural
education show that the educational issue is a strong
candidate for the problem of slow incorporation of digital
fabrication and mass customization into architectural design
practice in Brazil.
The implications are huge for the future of the Brazilian
architecture considering the possibility of incorporating the
mass customization paradigm and thus allowing for complex
and non standardized components to be produced within
reasonable economic constraints.
Figure 8. Rendered view of the exhibition pavilion of the student Nágila
Ramos
(Rendered with formZ RenderZone Plus) ACKNOWLEMENTS
Figure 9 shows the prototype of roof of the design project We would like to thank the management of Gravia
shown above, just before being removed from the sustaining (http://www.gravia.com/), Mega Forte e Ferro e Aço
frame. Badaruco (http://www.badaruco.com.br) for kindly
participating in survey described in this paper.
REFERENCES
[1] Polette, Doug and Landers, Jack M. “Construction Systems”,
Goodheart-Willcox Company, Inc, Publishers, Illinois, 1995, pp. 45-
49.
[2] Kieran, Stephen and Timberlake, James, “Refabricating
Architecture”, McGraw-Hill, New York, 2004, pp. 131-153.
[3] Kolarevic, Branko: Digital Production, in Branko Kolarevic,
Architecture in the Digital Age – Design and Manufacturing, Taylor
& Francis, New York, pp. 31-53, 2005.
[4] Schodek, D.; Bechthold, M.; Griggs, K.; Kao, K. M. & STeinberg, M.
“Digital Design and Manufacturing – CAD/CAM Applications in
Architecture and Design”, John Wiley & Sons, New Jersey, 2005,
pp. 339-344.
[5] Franken, Bernhard, “Real as Data”, in in Branko Kolarevic,
Figure 9. Rapid prototype of the exhibition pavilion of the student Nágila Architecture in the Digital Age – Design and Manufacturing, Taylor
Ramos & Francis, New York, p. 138, 2005.
(Milled in a Roland Modela MDX-40 three-axes milling machine) [6] Stacey, Michael; Beesley, Philip & Hui, Vincent, “Digital
Fabricators”, University of Waterloo School of Architecture Press,
The above example was one of design project specially 2004, p. 6.
selected for rapid prototyping. [7] Schodek, D.; Bechthold, M.; Griggs, K.; Kao, K. M. & STeinberg, M.
“Digital Design and Manufacturing – CAD/CAM Applications in
We observed that the more theoretical background we Architecture and Design”, John Wiley & Sons, New Jersey, 2005,
gave to the students about digital fabrication and mass pp. 3-4, 237-296.
customization and the more information we provided about [8] Kolarevic, Branko: Digital Fabrication: Manufacturing Architecture
the existence of digital fabricators in our region, the more in the Information Age, in ACADIA 2001, Buffalo, New York, 2001,
they felt confident to endeavor outside the boundaries of p. 275.
[9] Schodek, D.; Bechthold, M.; Griggs, K.; Kao, K. M. & STeinberg, M.
orthogonal and serial architecture.
“Digital Design and Manufacturing – CAD/CAM Applications in
The presence of CNC equipment in our school was Architecture and Design”, John Wiley & Sons, New Jersey, 2005,
important to convince the students of this technology’s p. 341.
COMPUTER NUMERICAL CONTROL, MASS CUSTOMIZATION AND ARCHITECTURAL EDUCATION 167
[10] Cunha, Edgard L., Gonçalves, José R., Souza, José R. A., Silva, [12] Schodek, D.; Bechthold, M.; Griggs, K.; Kao, K. M. & STeinberg, M.
Sílvio P.(Editores): Corte e Conformação de Metais, Editora Aranda, “Digital Design and Manufacturing – CAD/CAM Applications in
São Paulo, 2008, número 44, pp. 70-91. Architecture and Design”, John Wiley & Sons, New Jersey, 2005, pp.
[11] Cunha, Edgard L., Gonçalves, José R., Souza, José R. A., Silva, 3-4, pp. 239-244.
Sílvio P.(Editores): Corte e Conformação de Metais, Editora Aranda, [13] Kolarevic, Branko: Digital Fabrication: Manufacturing Architecture
São Paulo, 2008, número 44, p. 71. in the Information Age, in ACADIA 2001, Buffalo, New York, 2001,
pp. 268-277.
Application of Project Based Study
in the Learning Process: A Case Study
Ashfaque A Chowdhury, Mohammad G Rasul and M. Masud K. Khan
169
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_29, © Springer Science+Business Media B.V. 2010
170 CHOWDHURY ET AL.
analysis of incompressible flow in pipes, analysis of buoyancy The graduate attributes mapping was prepared to obtain an
and stability of floating bodies and measurement of fluid flow. indication of how the fluid mechanics course contributed to
It introduced the methods of analysing fluid systems using the achieve course learning outcomes and graduate attributes and
conservation of mass and momentum equations, Navier to understand better how attributes are developed as students’
Stoke’s equation combined with the concept of a control progress from one year to next year. This is done by indicating
volume. To solve the problems in fluid mechanics course, the graduate attributes targeted by each course learning
students familiarised with the theories of incompressible flows outcome. The indicative marks are placed under the headings
in pipe systems, Euler’s equation, Bernoulli’s equation and the of Teaching, Practice and Assessment to illustrate the course
use of similitude and modelling principles and techniques. emphasis (low/medium/high).
Students were required to act professionally in presenting
information, communicating, working and learning
individually and in peer learning teams to achieve the learning V. ASSESSMENTS
outcome of the fluid mechanics course. Initially the fluid
To pass the course students had to attempt all the
mechanics course was offered to on-campus students. Later,
assignments, laboratory reports and complete all examinations
the existing course was adapted for Interactive System-wide
and obtain a grade of at least 50% and a passing grade in the
Learning (ISL) delivery on other regional campuses. An
formal exam. Assessments for the course were on the basis of
online lecture delivery was developed for flexible course
laboratory reports (20% weighting), a fluid mechanics project
offerings with a one week on-campus mandatory residential
plan (20% weighting) and investigation (20% weighting),
school requirement. The restructured course materials can be
workbook submission (pass/fail in competency) and formal
used both as supporting teaching materials and as reference
examination (40% weighting).
materials in project work for research and self study.
A. Laboratory Exercises
There were three components in laboratory exercises – a
III. STUDENTS LEARNING ACTIVITIES formal lab report, video lab questions and virtual lab design
At the beginning of the course, all students were given an and experimentation. The first two components were group
overview of the course to make clear the expectations about assessment items and each group were required to submit a
how the course would operate. Flex students were introduced report on the assessments. Every group member contributed
to online learning practice and were assessed on online to the conduct, preparation and write up of the laboratory
learning skills required for this program. Students were report. The objective of involving students in video lab
encouraged to study and respond to the review questions and questions and virtual lab activities is to develop higher order
solve exercise problems. They were also familiarised with the thinking from the problem statements and thus a deep
real fluid systems and flow phenomena through the laboratory understanding of the knowledge on the subject matters.
exercises. Students were required to study the text and other Practical videos in most chapters of weekly schedules were
course materials as a guide to solve the workbook problems, uploaded on the course delivery site.
virtual laboratory activities and other assessment items. The students were asked to solve the related practical
Students were able to discuss their difficulties with problems questions demonstrating the steps with related background
in the discussion list or email the course lecturer. Students knowledge for a virtual lab problem. In the design and
were expected to follow a self-directed study schedule which experimentation component, the students were encouraged to
met the required deadline. The weekly schedule of activities design an experiment in fluid mechanics from any of the
for on campus students were Lectures (two hours per week), topics covered in the weekly modules that the students would
Tutorials (two hours per week), Laboratory (two hours per like to improve further in future. Students were required to
week). A residential school was conducted for flexible and submit a report on the selected experimentation covering the
other regional campus students after the midterm vacation. necessity for the investigation, a description of the work, the
Tele-tutorials over Skype and/or local tutorials were arranged parameters requiring observation and the safety procedures.
for off-campus students. Students were also required to complete the entire laboratory
exercise including the drawing of graphs and final answers on
issues such as calibration of obstruction meter, flow in pipes,
IV. ADDRESSING OF GRADUATE ATTRIBUTES stability of a floating body and velocity profile and pressure
losses in a smooth pipe.
In the design and development process of a technical course
for the engineering students, there are several objectives need B. Engineering Fluids Project
to consider such as the goals, resources, course learning The aim of an engineering fluids project is to ensure that
outcome, and last but not the least the graduate attributes. The the future engineers will comply with the principles of
targeted graduate attributes were Science and Engineering sustainability though improving their research ability and
knowledge, Effective communications, Technical competency, knowledge of fluid mechanics. The projects related to basic
Problem solving and systems approach, Functionality in a principles of fluid mechanics assist the students to enhance
team, Social, cultural, global and environmental awareness, their learning experiences and to achieve the graduate
knowledge on sustainable design and development, attributes through a team based design approach utilising the
professionalism and ethics and lifelong learning. concept of sustainability to meet the needs of present and
APPLICATION OF PROJECT BASED STUDY IN THE LEARNING PROCESS 171
future generations. The students experienced a productive show the application and understanding of the subject matter
pedagogy in fluid mechanics through improved knowledge & and to use the correct procedures and calculations in their
understanding. workbook answers. Attainments of competency attracted extra
Several project options were available for the students. In a marks to move to a higher grade when short by a few marks.
project proposal, on part of land assumed, the right side was Students were also required to sit for closed book final exam.
assumed to construct an engineering complex. Based on the
knowledge in the course, a group of students were responsible
to design the supply of drinking water and the management of VI. ASSESSMENTS CRITERIA
waste water facilities for the engineering workshop. Another Engineering fluid projects and workbooks were evaluated
two groups of students were assigned to design the same for a based on the evidence of the application and integration of
shopping complex and a cinema theatre. There were several knowledge and understanding of the material studied and
steps suggested in the course profile for the students to follow communication and justification of the subject matters. It was
and report to the campus tutors on the progress of the project. kept in consideration that students demonstrated accurate use
In the first two to four weeks, students were required to of engineering terms, collection and organisation of
define problems and project scope through the fundamentals information in the various forms of presentation. It includes
of fluid mechanics learnt, and then proposes a sustainable well structured solutions, use of engineering mathematical
concept design complying with course learning outcomes. In reasoning and proof to develop logical arguments, recognition
week five, students were required to present the progress in of the effects of assumptions used evaluation of the validity of
groups identifying the contribution of each member. In the arguments and justification of procedures.
next three to four weeks, students prepared the preliminary The laboratory components of the assessment items were
sustainable design with specifications. Often students revisited assessed based on reporting of key elements required to
the preliminary plan and refined the concept design with the undertake the laboratory sessions, clarity of expressions,
feedback from course tutors and lecturer. Later in the term, appropriate referencing of sources, accurate use of fluid
students finalised and documented the sustainable design with mechanics knowledge, presentation of mathematical
a strategy to implement in future. It was required to build a formulation, clarity and logical presentation of ideas and
physical prototype design to have a nice memory of the young arguments by means of data analysis and synthesis. To
engineers’ efforts in future. evaluate the ability of the students on the application and
The amount of traditional homework problems assigned integration of knowledge and understanding - evidence of
was reduced approximately by half. The design project was interpretation, analysis, identification of assumptions,
assigned to groups of four or five students at the start of the selection and use of symbols of strategies and procedures
course. Handing out the project immediately at the outset of while showing initiative to solve the problem, were
the course, where students are largely unfamiliar with the considered. In some cases, it was found that the problem has
material required for the completion of the project, renders the not been solved, however substantial progress towards a
learning process goal driven. This approach is in support of rational solution and shows evidence of appropriate
the life-long learning scenario for which students ought to be interpretation, analysis, identification of assumptions and
prepared and where the learning typically occurs on a needs selection of strategies and procedures. Additionally, the
basis in an interactive and often collaborative learning mode. problem has been solved however, there is a lack of
The submissions of two written progress reports were appropriate evidence was also reported.
required in order to guide students through the design tasks
and enforce progress throughout the entire semester. In
addition, a total of two class periods throughout the semester VII. STUDENTS PERFORMANCE
were allotted for a progress presentation and a final
presentation by each student team. It was found that the Course history showed that, after introduction of the
project based learning concepts enhanced students’ learning integration of project based investigation in fluid mechanics
experiences and assisted in the achievement of key attributes. course, on average (based on three years data), about 4.6% of
The students experienced a dynamic pedagogy in fluid the total enrolled students withdrew from the course after
mechanics through improved content quality, connectedness to enrolment and the failing rate dropped to 6.5%. Most of the
real world problems, a supportive team learning environment students performed very well and their overall performance
and recognition of differences between the team members. was quite impressive. On an average, 28.7% students received
high distinction and 30.6% students received distinction in the
C. Other Assessment Items course. As shown in Fig 1, the comparisons of student
To improve the research capability and gain knowledge and performances over 2007 and 2008 demonstrate that the failing
improve understanding in fluid mechanics, students were rate of the students in the course was reduced from 14.3
asked to prepare a teaching and learning journal in fluid percent to 5.4 percent. The blended assessment mechanism
mechanics through the lectures, tutorials, experiments and incorporating lab components, project works, workbook and
project work. Students were expected to solve higher order final exam assisted students to understand the subject matter
problems and open answers which require researching on the and to develop a higher level of skills in an advanced
weekly topics. The workbook contributed to the overall grade engineering topic like fluid mechanics.
of the course as it supplemented the assignment work and
confirmed students overall competency. They were required to
172 CHOWDHURY ET AL.
IX. CONCLUDING REMARKS [4] A. Jennings, “Viewpoint: the need to upgrade experimental learning,”
International Journal of Engineering Education, vol. 14, pp. 83-88,
It is intended to teach the fluid mechanics course by 1998.
creating a learning centred environment which will encourage [5] H. K. Navaz, B. S. Henderson, R. M. Berg, and S. M. Nekcoei, “A new
the students for creative and critical thinking and lifelong approach to teaching undergradaute thermal/fluid sciences - courses in
applied computational fluid dynamics and compressible flow,”
learning. It is necessary to prepare undergraduate engineering International Journal of Mechanical Engineering Education, vol. 30,
students to design, and analyse real life problems. The paper pp. 35-49, 2002.
describes an integrated assessment methodology which was [6] J. C. Musto, “A project based approach to teaching numerical methods,”
implemented in a fluid mechanics course and demonstrates its International Journal of Mechanical Engineering Education, vol. 30,
pp. 233-247, 2002.
effectiveness in facilitating engineering undergraduate [7] L. R., Mustoe, and A. C. Croft, “Motivating engineering students by
students to become efficient problem solvers of real life using modern case studies,” International Journal of Engineering
problems. The success of the assessment procedures were Education, vol. 15, pp. 469-476, 1999.
evaluated based on the successful completion of the course, [8] V. Gomes, “Consolidation of engineering education through industrial
case studies,” International Journal of Engineering Education, vol. 18,
students’ grades, students’ withdrawal rate from the course pp. 479-484, 2002.
and the students’ overall perception of benefits of the course [9] L. D. Feisel, and A. J. Rosa, “The role of the laboratory in undergradaute
studied. Integrated assessment methods assisted with engineering education,” Journal of Engineering Education, vol. ,
strengthening previous and prerequisite learning requirements. pp. 121 -130, 2005.
[10] E. DeBartolo, and R. Robinson, “A freshman engineering curriculum
Therefore, benefits the subsequent learning process is achieve integrating design and experimentation,” International Journal of
through excellence in research and teaching. The assessment Mechanical Engineering Education, vol. 35, pp. 91-107, 2007.
items offered students experience of real life problems in a
professional manner. Further improvement of the course is in Ashfaque Ahmed Chowdhury obtained his B.Sc. in Mechanical
Engineering degree from Islamic University of Technology, Bangladesh in
progress considering the feedback from all the stakeholders of 2003. In 2008, he completd Master of Engineering degree at Central
the course. Queensland University, Australia. Since 2004, he worked as a lecturer in the
Department of Mechanical and Chemical Engineering at Islamic University of
Technology and in the College of Engineering and Built Environment, Central
ACKNOWLEDGMENT Queensland University. His current research interest include intelligent
control system, manufacturing technologies, computer numeric control
The authors gratefully acknowledge the contributions of Dr manufacturing, computer-aided design and manufacturing, building energy
Ahmed Elgezawy and Mr Fons Nouwens who generously simulation, energy conservation and management, low energy cooling
technologies.
assisted in the development and implementation of the course
M G Rasul is a Senior Lecturer of Mechanical Engineering at the College
in the faculty. Special appreciation goes to Mr Tim of Engineering and Built Environment, Central Queensland University,
Mcsweeney for his remark to improve the manuscript. The Australia. He received his PhD and MEng degree from The University of
authors also acknowledge the support of the Learning and Queensland, Australia and Asian Institute of Technology, Thailand
respectively. He is specialized and an expert in energy technologies, process
Teaching Education Research Centre (Special Interest Group –
industries, thermo-fluid engineering including combustion and process
Future Engineering Education Directions), which is engineering, and light metals. He has published over 100 research articles in
established and supported under CQUniversity’s Research reputed journals and conferences on a wide range of research interests.
Centres Program. M. Masud K. Khan is an Associate Professor and Head of Department of
Infrastructures in the Faculty of Sciences, Engineering and Health, Central
Queensland University, Rockhampton, Queensland 4702, Australia. He received
REFERENCES his MS (Mech) with 1st class Honours from Moscow Institute of Petrochemical
[1] T. Gray, “Integration of case study technical investigations in and Gas Industry in 1982. Subsequently he worked with the oil industry for 2
honours/masters engineering courses,” International Journal of years. He obtained his PhD in engineering from the University
Mechanical Engineering Education, vol. 34, pp. 121-130, 2006. of Sydney in 1990. His teaching, research and consulting activities are in the
[2] D. Harris, “A Case for Project Based Design Education,” International area of non-Newtonian fluid mechanics, thermofluid engineering and rheology
Journal of Engineering Education, vol. 17, pp. 367-369, 2001. of industrial fluids and polymeric materials. He has authored more than 95
[3] J. E. Holt, “On the nature of mechanical engineering - an engineering refereed journal, conference and technical papers. He is a member of American
ethos,” International Journal of Engineering Education, vol. 12, Society of Rheology, Australian Society of Rheology, Engineers Australia and
pp. 332-338, 1996. the Australasian Association for Engineering Education (AAEE).
Interdisciplinary Automation and Control
in a Programmable Logic Controller (PLC)
Laboratory
Jack Toporovsky, Christine Hempowicz, and Tarek M. Sobh
School of Engineering, University of Bridgeport
[email protected], [email protected], and [email protected]
175
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_30, © Springer Science+Business Media B.V. 2010
176 TOPOROVSKY ET AL.
used widely in industry, robotics, and vision A team of PLC students designed an integration of
concepts. different brands of control components into one
In EE 464, students learn basic and advanced application. The components are: SLC5/03 Allen-
concepts and applications for PLC programming. In Bradley PLC, Network Card 1761-NET-ENI,
addition, students learn to integrate various Mitsubishi Variable Frequency Drive, and Cutler
components with PLCs, such as sensors, switches, Hammer PM1700 Touch Screen. For this project,
vision cameras, variable speed drives, etc. students were involved in wiring, communications
Instructional design for each course was more cable design, equipment programming, and
focused on inquiry and problem-based learning troubleshooting.
within the context of classroom community, and less Students produced the layout design and wiring
focused on lecture-style instruction. All are diagrams for the project to integrate the SLC5/03
supported in the literature [13] and [14]. PLC with the Mitsubishi Variable Frequency Drive
(VFD) and Cutler Hammer Touch Screen, and to
A. Equipment Design and Integration establish a network to the PLC. The design layouts
In the PLC Industrial Control Lab, all control and wiring diagrams were created using AutoCAD.
cabinets share the same local network with different Students were able to build a control cabinet with
types of PLCs and Human Machine Interfaces output to an AC motor, programming of the PLC,
(HMIs), so students are able to communicate to any VFD, touch screen, etc. (See Figure 1.)
of them from the class computer stations.
Fig. 1. Any computer station can establish communication to the SLC5/03 PLC through the network hub switch and through the internet
adapter (# 1746ENIW). At the same time, an AC motor (3-phase AC) runs commands from the PLC, transferring from the relay module
to the Mitsubishi Variable Frequency Drive (VFD/FR/S500). The value of the frequency is converted to an analog voltage (0 to 5 volts)
from the VFD; it is then read by an analog card. In this way, a closed-loop communication link is established between the PLC and the
VFD.
INTERDISCIPLINARY AUTOMATION AND CONTROL IN A PLC LABORATORY 177
Abstract - Three-dimensional (3D) measurements of whole plants of plants with 3D spatial information, spectral information can
may provide detailed structure information about plant growth also biochemically characterize plants growth and health. For
patterns and also complement existing X-ray systems for the example, the normalized difference vegetation index (NDVI)
below-ground part. In addition to this structural characterization using two spectral reflectance responses at red and near-
of plants, spectral information may also biochemically
characterize plants' health. A stereo vision technique is a cost-
infrared (NIR) spectral regions has been widely used in
effective and rapid imaging technique for measuring and remote sensing as an index to estimate various vegetation
reconstructing 3D structures. The Normalized Difference properties including chlorophyll concentration in leaves [6],
Vegetation Index (NDVI) requiring measurements of two leaf area index [7], biomass and plant vigor [8], plant
spectral wavelengths in the NIR and red spectral regions has productivity [9], and stress [10]. We integrated both stereo
been widely used in remote sensing as an index to estimate vision and NDVI techniques and developed a dual-band stereo
various vegetation properties including chlorophyll spectral imaging system for chlorophyll and biomass
concentration in leaves, leaf area index, biomass, and plant quantification of plants in 3D space.
productivity. We integrated both stereo vision and NDVI Our main objective was to develop a SSI system that
techniques and developed a stereo spectral imaging (SSI) system
for chlorophyll and biomass quantification of plants in 3D space.
could monitor whole plant canopies in 3D. The specific goals
We used a stereo vision camera system and custom designed a were to reconstruct plant surfaces in 3D space via stereo
dual-wavelength filter system at 690 nm and 750 nm to develop vision, to acquire spectral features via NDVI, and to map the
the SSI system. Calibration techniques for NDVI computation spectral features back onto the 3D data via computer-graphics
using two spectral band images were developed by referencing a rendering.
diffuse reflectance panel. We also developed a texture mapping
technique for rendering NDVI values in 3D space. In this paper, II. MATERIALS AND METHODS
the performance of the SSI system was evaluated with an
artificial plant with spectral properties similar to real plant’s A. Overview of Stereo Spectral Imaging (SSI) System
green leaves. Stereo vision is a cost-effective and rapid imaging
technique for measuring and reconstructing object structures
I. INTRODUCTION in 3D. The fundamental idea of stereo vision is based on the
intersection of rays from two views in 3D, assuming that
Imaging systems exploiting both stereo vision and
camera intrinsic parameters and the baseline of cameras
spectroscopy have been developed for various agricultural and
(relative positions of the cameras) are known (or can be
remote sensing applications and space missions, e.g. rice
estimated). If an object feature in 3D is visible to both camera
seedling row detection [1], wheat plant geometric properties
views, the feature may appear at different locations on the
[2], autonomous vehicle navigation [3], digital elevation maps
images captured by the cameras. The amount of spatial
for land use/cover databases [4] and the Surface Stereo Imager
displacement, called disparity, is an important cue to obtain
for the Phoenix Mars Lander of NASA [5]. In this paper, a
the depth of the object feature. Then, depth information is
stereo spectral imaging (SSI) system collecting and processing
computed directly from disparity values [11]. In practice,
spatial and spectral information from two camera views and
disparity estimation is performed by a search algorithm called
two spectral bands was developed to study plant health.
stereo matching to find correspondence between two separate
3D measurements of whole plant canopies using stereo
images. Stereo matching methods to solve this correspondence
vision may provide detailed spatial structure information
problem can be categorized into two groups: region
about plant health, disease propagation path, or phenotypes of
correlation-based and feature-based stereo matching. The
plants used in biomass production such as cotton, poplar and
region correlation-based stereo matching algorithm finds
switch grass, and also can complement existing X-ray systems
correspondence for every pixel in the image by using pre-
for the below-ground part. In addition to the characterization
determined search windows to compute correlation whereas
181
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_31, © Springer Science+Business Media B.V. 2010
182 YOON AND THAI
the feature-based method finds correspondence only for a holders were designed and attached to the outside of the
sparse set of image features (e.g. edges, lines and corners). camera body such that the filters were placed in front of the
Two common criteria for correlation computation are the sum lenses to configure the 24 cm baseline setup (Fig. 1).
of absolute difference (SAD) and the sum of squared
difference (SSD). The SSI system was based on the region
correlation-based stereo matching algorithm and the SAD
criterion.
Live green plant canopies absorb solar radiation at the
blue and red spectral regions for photosynthesis whereas most
of the radiation at the spectral region longer than 700 nm is
scattered. Hence, images of green plants appear bright in the
near-infrared (NIR) spectral region and dark in the red spectral Fig. 1. Stereo Spectral Imaging setup.
region. The NDVI was developed to utilize the properties of The selection of the NIR filter at 750 nm was to avoid the
the different light energy absorption characteristics of plants at strong signal attenuation due to the oxygen absorption of
the NIR and red spectral regions. The NDVI at each image natural sunlight energy at 760 nm in the consideration of the
pixel is obtained by possibility of deploying the SSI system outdoors and the low
~ ~ quantum efficiency of the CCD sensor beyond 800 nm. Two
I NIR( x,y ) − I RED( x,y ) . (1)
NDVI ( x,y ) =
~ ~ tungsten halogen lamps (AC) were used as the lighting source
I NIR( x,y ) + I RED( x,y )
where ~ and ~ in a laboratory setting for this study.
I RED ( x,y ) I NIR( x,y) refer to the calibrated (or
The software was written in C++ and the C++ libraries for
normalized) relative reflectance values at the visible (red) and stereo image acquisition and processing were provided in the
near-infrared (NIR) spectral regions. The NDVI value itself Point Grey Research’s Software Development Kit (SDK). The
varies between -1.0 and +1.0. In practice, a value less than 0.2 graphical user interface of the software was written in the
usually means unhealthy or sparse vegetation and a value Microsoft Foundation Class 6.0 (i.e. Visual C++ 6.0). An open
close to +1 (more than 0.6) means very healthy green leaves source computer vision library OpenCV 1.0 was used to
or very dense vegetation. NDVI values for vegetation implement image processing algorithms mentioned in the
typically range between 0.2 and 0.7. A zero or less than zero paper. OpenGL [12] codes were written for implementing 3D
means no vegetation. However, actual wavelengths and rendering and texture mapping. Fig. 2 shows a screen capture
bandwidths for NDVI calculations varied with different of the software’s graphical user interface.
sensors and applications.
The SSI system consisted of a stereo vision camera
(Bumblebee XB3, Point Grey Research Inc., Richmond, BC,
Canada), a dual-wavelength filter system at 690 nm (red) and
750 nm (NIR) for NDVI computation, a PC and software. The
Bumblebee XB3 is a 3-CCD monochrome stereo camera with
the IEEE-1394b communication link (800 Mbits/s). The three
lenses were horizontally placed 12 cm apart, thus three Fig. 2. Software Interface Design.
different combinations of baselines were selectable by the user
via software: one wide-baseline (24 cm) and two narrow- B. Stereo Image Acquisition and Processing
baselines (12 cm). The lenses were sealed inside the camera For each experiment, 3 sets of stereo images were
unit. The focal length of each lens was set to 3.8 mm. The captured: raw (i.e. no filter was used), red filters on each lens,
horizontal field of view of each lens was 70°. The camera was and NIR filters on each lens (see Fig. 3). Each set was
factory pre-calibrated for lens distortions and camera acquired one at a time by manually removing or changing to
misalignments. Therefore, there was no need for in-field the proper filters at the filter holders as shown in Fig. 1.
camera calibration and images were aligned within 0.05 pixel
RMS error. The maximum image resolution per lens was
1280x960. Smaller resolutions were selectable by the user via (a)
software in order to speed up the stereo processing and to
increase the robustness of the stereo matching process. The
frame transfer rate was 15 frames per second at the full-
(b)
resolution.
Optical band-pass filters (50.8 mm x 50.8 mm) were used
for the NDVI computation. The center wavelengths of the
near-infrared and red band-pass filters were 750 nm and 690 (c)
nm (Part No. 20BPF10-690 and 20BPF10-750 respectively
from Newport Corporation, Irvine, CA) and two sets of these
filters were used. The bandwidths (full width at half Fig. 3. Raw (a), red (b) and NIR (c) stereo images.
maximum) were 13 nm (NIR) and 11 nm (red). Two filter (Note: the right camera view is located on the left of each stereo image)
STEREO SPECTRAL IMAGING SYSTEM FOR PLANT HEALTH CHARACTERIZATION 183
The camera parameters and lighting conditions were not computation of the NDVI became possible with raw digital
changed during the data acquisition period (except for a flat numbers in the place of normalized values, as defined by
~ ~
field calibration to be mentioned later). The captured three I NIR ( x, y ) = I NIR ( x, y ), I RED ( x, y ) = IˆRED ( x, y ) ,. (4)
stereo images (raw, red and NIR) were used for the following
tasks. First, the raw stereo image was used for stereo matching where IˆRED( x,y) is the intensity value of the red-band image
leading to 3D reconstruction. Second, both NIR- and red- captured after the adjustment of the camera parameter.
filtered stereo images were used for NDVI computation. Stereo processing in the SSI system was implemented by
Third, the NIR-filtered stereo image was used for foreground using the library functions provided in the Point Grey
object segmentation because background clutters were better Research’s SDK and OpenCV. The following steps were
suppressed by the NIR filter. applied to all 3 stereo images:
The NDVI computation required the intensity calibration a) After two images (8-bit monochrome 1280x960)
(or called normalization) to obtain the relative reflectance at was captured, they were interleaved into a single image
each pixel. We implemented a relative reflectance (RR) according to a pre-determined image format (2560x960)
calibration method for the SSI system. This method was to in order to transfer the data according to the DCAM
calibrate every single pixel on the image by using a 99% specification. The acquired stereo images were then pre-
diffuse reflectance panel which completely covered the plant processed by a low-pass filter and spatially rectified to
object in the field of view. The relative reflectance calibration un-warp the lens distortions. Without low-pass filtering,
equation at each pixel required to collect a reference image the image rectification would produce aliasing effects.
and a dark current image, and was calculated by After the rectification, the radial image patterns were
~ Imeasurement ( x,y ) − Idark current ( x,y ) . (2) straightened. This process was applied to all stereo
I m easurement ( x,y ) =
I 99% reflectance( x,y ) − Idark current ( x,y ) images whether the band-pass filters were used or not.
The RR calibration method may be impractical for outdoor b) The rectified NIR-band stereo image was sent to
applications because it is not possible to put a reflectance an image segmentation function. For segmentation, an
panel large enough to cover the area in the entire field of view. intensity threshold was applied to the rectified stereo
Thus, we developed a flat field (FF) calibration technique image for suppressing out background clutters. Then, a
more suitable for outdoor imaging of plants. The original FF morphological open filter was applied to remove small
calibration technique [13] was based on an assumption that the and isolated background clutters. The remaining clutters,
relative reflectance of a calibration target remained unchanged if remained, were removed manually. Then, this
over wavelengths (i.e. red and NIR spectral regions). For the segmentation image was used as a binary mask image to
FF calibration, a reflectance target having very little variation be applied to all stereo images.
in reflectance over wavelengths was placed in the scene and c) The raw stereo image masked by the segmentation
the digital number at each pixel was normalized to the mean map was sent to the stereo matching algorithm to obtain
value of all pixels in the region of interest (ROI) of the 3D information of the object in the scene. The stereo
calibration target. The normalization was obtained by matching was based on the sum of absolute difference
~ I ( x, y ) ~ I ( x, y ) ,. (3) (SAD) correlation method for estimating disparity at
I ( x, y ) = NIR
NIR , I ( x, y ) = RED
RED every pixel in the reference view image. The right camera
I A, NIR I A,RED
view was used as the reference view image. The
where I A, NIR and I A, RED were the average intensity values of secondary (optional) stereo matching was based on the
the ROI (i.e., the target area A) at the NIR and red spectral edge detection that utilized the relative difference in
bands, respectively. If I A, NIR and I A, RED are equal (i.e. a intensity values. The output of the stereo matching
algorithm was a disparity image. When displaying it, each
constant I A ) and (3) is plugged into (1), then the constant pixel on the disparity image was mapped by a look-up
I A will be cancelled out and the un-calibrated digital table representing a close object feature with red and a
numbers, IRED( x,y) and INIR( x,y) will replace the normalized far-away object feature with blue. The disparity image
~ was represented with either an 8-bit (full-pixel
values I~RED( x,y) and I NIR( x,y) in (1). For making the average displacement) or 16-bit (sub-pixel displacement)
intensity values of the ROI equal (or approximately equal) precision.
between the two spectral bands, we implemented an algorithm d) The NIR and red images were input to the NDVI
to adjust either a shutter time or a camera gain of the red-band. computation engine. Again, they were also masked by the
More specifically, we captured the NIR-band image first and segmentation map. A pseudo-color image was generated
then captured the red-band image by changing a camera by a color look-up table in order to display the resulting
parameter (a shutter time or a gain) such that the average NDVI image. The color look-up table of the NDVI values
intensity value of the ROI in the red-band image becomes was made according to the following rule:
(approximately) equal to that in the NIR-band image. The • If NDVI is less than or equal to 0.2, then grey
algorithm found an optimal camera parameter for the red-band value. The black refers to -1.
capture by minimizing the sum of the absolute intensity
difference between the ROIs of two bands. Then, the
184 YOON AND THAI
• If NDVI is greater than 0.2 and less than or equal calibration (Eq. 2). Another 99% diffuse reflectance target
to 0.8, then mixture of pure red (0.2) and pure (SRT-99-050, Spectralon, 5”x5”, Labsphere) was used as the
green (0.8). reference material for the flat field (FF) calibration (Eq. 4). To
• If NDVI is greater than 0.8, then mixture of evaluate the RR calibration technique, we used seven target
green and pure blue (1.0). materials: a 5”x5” multi-step diffuse reflectance target (SRT-
The floating-point NDVI values were also saved for MS-050, Labsphere) having four sub-panels of 99%, 50%,
future numerical quantification. 25% and 12% nominal reflectance values, two vinyl sheets
used for military camouflages (uniform dark and light green)
C. Rendering of NDVI Image in 3D and their back sides (black). The nominal reflectance values of
After stereo processing, each pixel in the right camera the vinyl sheets were measured by a four channel fiber-optic
view (main view – 2D) was mapped into a point in the camera spectrometer (SQ2000, Ocean Optics, Dunedin, FL). For
coordinate system (3D) where its origin was at the focal point evaluating the accuracy of the computed NDVI values, we
of the right camera (please note that by convention this image used the three camouflage vinyl sheets.
was located on the left of each stereo image as shown in Fig. An artificial orchid plant (Fig. 1) was used for the SSI
3). The set of these points in 3D were called point clouds. system development because its leaves showed similar NIR
Obtaining point clouds from stereo matching was the first step spectral reflectance as compared to real leaves. Also it was
for 3D surface reconstruction. Simply, the point clouds were essential to have a reference plant during the system
nothing but a set of 3D coordinates representing surface points development cycles because the artificial plant was not
in 3D. Thus, it was important to spatially connect these perishable and provided good references over time.
surface points in order to create an approximate surface
representation. This representation was obtained by a III. RESULTS
geometric triangular mesh model. Rendering of point clouds A. Calibration and NDVI Computation
and triangular meshes in 3D was implemented by OpenGL The performance of the RR calibration technique was
functions. The quality of the 3D mesh representation was evaluated by using seven test materials placed in the camera
greatly affected by the quality of generated cloud points. field of view: the multi-step reflectance target (99%, 50%,
Typically, point clouds needed to be cleaned and edited to 25% and 12%) and the two vinyl sheets (dark and light green)
remove duplicated and unreferenced points. For this task, and their back side (black). The performance of the RR
when necessary, we removed isolated and tiny surface features calibration was compared to spectrometer measurements. The
by applying a median filter directly to a disparity image. The calibration of the spectrometer measurements was done on the
3D coordinates of the point clouds were also saved as a fly by the Ocean Optics’ software using its own dark-current
standard file format (xyz format) so as to be used in the free & reference (Spectralon) spectra. Calibration values at 689.92
open-source 3D data processing software (MeshLab) for off- nm and 749.99 nm of the spectrometer were compared against
line processing of these point clouds. the calibration values of the SSI system at the NIR and red
Image rendering based on OpenGL was implemented for bands. Fig. 5 shows the scatter plot of calibrated image and
mapping the computed NDVI values onto 3D point clouds as spectrometer measurements of the test materials. Each point in
well as a mesh model in the camera coordinate system such the figure represented an average calibration value of one
that a 3D virtual image was generated at any given viewpoint. particular material at one of the two spectral wavelengths, thus
Also, 3D anaglyph NDVI images were generated in real-time 14 points were obtained. A regression line was overlaid to
to provide a 3D stereoscopic perception when viewed with represent the relationship between the two instruments. The R2
two color (red and green) glasses. Fig. 4 is the data flowchart value (goodness-of-fit) of the linear regression was 0.9912.
of the 3D rendering algorithm of the SSI system. Overall, the measured values from the imaging system were
slightly lower than the spectrometer measurements. But, the
Raw stereo image Stereo matching Point clouds
difference was consistently small across the entire calibration
Background
value range from 0 to 1.
NIR stereo image matching
3D surface
reconstruction
Red stereo image NDVI
compute
3D rendering
use of the camera on small plants was challenging. The most REFERENCES
important parameter for depth estimation was the min-max
range of the disparity values to search for. In order to improve [1] Y. Kaizu and K. Imou, “A dual-spectral camera system for paddy rice
the quality of the 3D rendering, the accuracy of the depth seedling row detection,” Computers and Electronics in Agriculture, vol.
63, no. 1, pp. 49-58, 2008.
estimation needs to be improved by a better stereo matching
algorithm or the point clouds may be cleaned by multiple [2] H. J. Andersen, L. Reng and K. Kirk, “Geometric plant properties by
stereo images taken at different viewpoints. relaxed stereo vision using simulated annealing,” Computers and
Electronics in Agriculture, vol. 49, no. 2, pp. 219-232, 2005.
[3] M. Kise and Q. Zhang, “Creating a panoramic field image using multi-
spectral stereovision system,” Computers and Electronics in
Agriculture, vol. 60, no. 1, pp. 67-75, 2008.
(a) Segmentation mask (b) Disparity map (c) NDVI image [4] F. Paul, C. Huggel, and A. Kääb, “Combining satellite multispectral
image data and a digital elevation model for mapping debris-covered
glaciers”, Remote Sensing of Environment, vol. 89, no. 4, pp. 510-518,
2004.
[6] C.L. Jones, N.O. Maness, M.L. Stone, and R. Jayasekara, “Chlorophyll
(e) Triangular mesh rendering at different viewpoints estimation using multispectral reflectance and height sensing”,
Transactions of the ASABE, vol. 50, no. 5, pp. 1867-1872, 2007.
Abstract –This paper illustrate through a case study the advantages validation. We also adopt the convention that there can be only
of the proposed Transition Network (TN) for modelling discrete one task (state) node following each transition, and that all the
automation on the shop floor. TN is associated with high-level state required resources for each task (state) are available by
diagram The proposed technique can easily simulate and emulate incorporating the inputs related to their availability into the
the automation logic and can be automatically translated to PLC
transition conditions.
(Programmable Logic Controller) code.
II. THE EXAMPLE CASE STUDY
Keywords: Transition Network, PLC, Automation, Ladder- For the case study a drill-press system is chosen to illustrate the
diagram
remaining steps of the proposed technique. The system is
adapted from [6]. Figure 1 depicts the main components and the
I. INTRODUCTION layout of the automated drill-press that will be used throughout
this paper. The drill press consists of an electric drill mounted on
The traditional modelling approach for state diagrams of a movable platform. The platform is attached to a piston of a
discrete automation approach typically characterize a state as hydraulic cylinder. The movable platform moves vertically
any possible combination of inputs and outputs [1,2,3]. The between the up position and the down position. The limit switch
emphasis is on the states. Each change in this combinatory design LS1 senses when the drill is in the up position and the limit
is a transition of one or more inputs and/or outputs [2, 3,4]. switch LS2 senses when it is in the down position. The hydraulic
Typically the logic of the generated models grows rapidly with cylinder is controlled by a hydraulic three-position, four-way
the number of inputs and outputs. Moreover, tracking the logic is valve with two solenoids. Solenoid A drives the hydraulic
a very difficult task [2,4,5]. In this paper we suggest a different cylinder down. Solenoid B drives it up. When both solenoids (A
approach with intuitive representation of cause and effect which and B) are deenergized the platform is stationary.
facilitate tracing of logic. In our model the states represent only A stationary table supports the workpiece and a pneumatic clamp
the output combinations. These are either machine actions or clamps the workpiece against a bracket. Limit switch LS4 is
some machine states of rest. The transitions are invoked by input actuated when the clamp is in the clamped position. Limit switch
changes and result in changes to some outputs. LS3 is actuated when the clamp is in the unclamped position. An
The technique presented in this paper not only relieves the ejector is also attached to the table (oriented 90 degrees from the
automation engineer from dealing with Programmable Logic clamping cylinder). The ejector ejects the workpiece when the
Controller (PLC) code when constructing the logic, but also in drilling is finished. Limit switch LS5 is actuated when the ejector
later stages, when changes are required. This approach offers is retracted and LS6 is actuated when the ejector is in the eject
intuitive action representation facilitating verification and position.
* Corresponding author: Full name: Yuval Cohen, Academic degree: PhD, Working place: The Open University of
Israel, Research field: Automation and production modeling, Post address: 108 Ravutski, P.O.Box 808, Raanana, Postcode
187
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_32, © Springer Science+Business Media B.V. 2010
188 COHEN ET AL.
POWER Sol. A
Hydraulic
OFF cylinder Pressure line Hydraulic
pump
ON
Fluid
tank
Drill Sol. B
motor
up
Movable platform
(drill head)
LS1 Chuck down
Drill bit
LS3 LS4
LS2
A
Bracket
Workpiece
Stationary Table A
Pneumatic
cylinder (clamp)
Sol. C
View A-A
Exhaust
Sol. D
Air supply
Exhaust
Fig. 1. The case study system: an automated drilling machine (adapted from [6])
machines and are therefore related only to combinations of the Next, essential data manipulation is described to enable the
machines outputs. Ignoring inputs dramatically simplifies the generation of the TN. The main input required for the proposed
state diagram. In a later stage only the necessary input scheme, is a specification of the output and timers status for
combinations are considered. The design of actions (legal each state. An example for such input (for the drill-press
states) is facilitated by analysing the steps. machine) is depicted in Table 1. This specification depends
For the case study of figure 1, the eight legal system states for heavily on the specific system at hand, and must be done
the drill press example are as follows: carefully and correctly.
1) OFF
2) Idle (only the hydraulic cylinder is ON) TABLE I
3) Ready - (similar to Idle but the drill motor is ON and the STATUS OF OUTPUTS FOR STATES IN FIG. 2.
clamp is in the clamped position) (Entries representing actions or active timers are shaded.)
4) Descend - (similar to Ready but Solenoid A is activated)
5) Ascend - (similar to Ready but Solenoid B is activated can OUTPUTS
be with or without clamping) Solenoids
6) Release clamp - (similar to Ascend but Solenoid C is State Pump Drill Coo Tim-
deactivated to release the clamp) A B C D motor motor -lant er 1
7) Eject workpiece - (similar to Release but Solenoid D (the 1 off off off off off off off off
ejector) is activated) 2 off off off off on off off On
8) Retract ejector - (similar to Release but Solenoid D is 3 off off on off on on off off
deactivated) 4 on off on off on on on off
At this point a state diagram can be constructed. Figure 2 5 off on on off on on off off
depicts the state diagram for the drill press specification. States 6 off off off off on off off off
are denoted by numbers and transitions by capital letters. 7 off off off on on off off Off
8 off off off off on off off Off
Start/End I
state Each transition in the state diagram has origin state and
A destination state. The status changes between the origin and
1.OFF 2. Idle
destination outputs and timers are the transition effects.
B Therefore, by comparing the lines in Table 1 that relate to the
J
origin and destination states, the effects could be generated
L 3. Ready automatically.
Transition effects are summarized in Table 2. For example, it is
C easy to see that the changes between lines 1 and 2 in table 1 are
in the pump motor and timer. These changes appear in line one
H 4. Descend of Table 2.
8. TABLE II
D THE EFFECTS OF TRANSITIONS OF FIG. 2.
The generation of significant portions of Tables 1 and 2 can be the states of the system. For example, in table III transition C
automated. In addition to the elements discussed so far, there from state 3 to state 4 is triggered by the corresponding ST3 and
are external variables originating in higher control levels. additional input change. Its effects include activating ST4 while
Moreover, different history of past events may set different deactivating ST3.
modes of operation. Such modes are identified by internal
variables. The additional variables for the drill-press example Two arc types used to activate triggers are as follows:
are as follows: 1. An enable arc ( ) the triggers can fire only while the
source place is ON.
External Variables 2. A disable arc ( ) the triggers can fire only while the
1. PROCESS2 signifies that the resources for task 2 are source place is OFF.
occupied. It is turned ON when the workpiece is ready to be
drilled (the robot finished placing the workpiece at the drill- Enable and disable arcs are drawn with dashed lines to denote
press). It shuts OFF when the drill is ready to execute a new that they do not activate or deactivate elements. A tokens is
task. used to denote the current activated state. Two types of arcs
2. FINISH2 signifies that the drilling task is finished. It is used to identify the effects of a transition as follows:
turned ON when the drilling task is finished, and turned OFF 1. Activate arc ( ) turns ON the place when the
after invoking the next transition. corresponding E-Transition is activated.
2. Deactivate arc ( ) turns OFF the place when the
Internal Variables corresponding E-Transition is activated.
1. Timer1 is used to warm up the hydraulic engine and cool it
down (turned ON by transitions A and H). TABLE III
2. Reset variable (signifies that the system is recovering from CASE STUDY TRIGGER EVENTS
power failure). Reset is turned ON whenever power is turned Trigger Trans Source state: Detected Supplementary
ON (transitions J and L) and turned OFF after workpiece event ition Var. (name) Inputs Identifiers
ejection (transition H). Change
3. Done variable signifies that the next time the system will be 1. Start A ST1 (closed) PROCESS2
idle is a cool down period followed by system shut OFF. ↑
4. Eight state-variables: ST1,ST2, ST3, ST4, ST5, ST6,ST7, 2. Reset ascend J ST1 (closed) Reset↑ LS1/OFF
ST8, signifying that the system is at a certain state. 3. Reset L ST1 (closed) Reset↑ LS1/ON
ejection
4. Warm-up B ST2 (Idle) Timer ↓ Done/OFF*
completed
III. TRANSITIONS NETWORK GENERATION 6. Cool Down I ST2. (Idle) Timer ↓ Done/ON**
completed (Internal
The state diagram is further exploded into a new type of a variable
graphical scheme named Transition Network (TN). TNs turned ON by
describe the changes in low level elements such as inputs, Transition H)
outputs, and registers, required for executing the state diagram. 7. Clamp C ST3. (Ready) LS4↑
TNs arrange the elements in a meaningful way that enables 8. Down D ST4. LS2↑
immediate comprehension of a low level code. The TNs operate (Descend)
much like Petri networks. However, instead of tokens, the 9. Up E ST5. LS1↑ Reset/OFF
transition is triggered by change of inputs and its effects are (Ascend)
modeled as turning On and OFF certain outputs. The transitions
10. Reset K ST5. LS1↑ Reset/ON
are composed of the following elements: 1) places, 2) triggers,
(Ascend)
and 3) arcs. These elements are all depicted in figure 3. Each
transition is activated by one or more triggers. The triggers are 11. Released F ST6. LS3↑
denoted by triangles pointing at a thick vertical line that (Release)
symbolizes the transition. Places (denoted by circles) represent 12. Ejected G ST7. (Eject) LS6↑
the inputs, outputs, events, and variables. 13. Retract H ST8. LS5↑
Events are assigned places with additional symbol to denotes (Retract)
the type of event (turn ON, and shut OFF). Places that use non- *Internal variable turned OFF by Transition I
binary data (e.g., timers and counters) are denoted by ** Internal variable turned ON by Transition H
rectangles. Additionally, places are added for logically denoting
TRANSITION NETWORK FOR AUTOMATION MODELING: A CASE STUDY 191
Each trigger is invoked by places linked to the trigger by enable trigger. Since the system starts at an initial state and the current
or disable arcs. Note the usage of the source state (STi) variable state is updated as integral part of each transition, tracking the
of the transition to facilitate trigger’s identification as one of the current state is done automatically. Moreover, the trigger
trigger’s conditions. After the trigger is activated, a transition identification is facilitated as a result. Trigger identification for
from the source state (i) to another state (j) occurs immediately. the case study is elaborated in Table III, and is depicted in
Each transition also resets the source state variable (STi) and figure 3.
sets the destination state variable (STj). Note that each trigger
has only one transition, but a transition may have more than one
ST3 C ST3
State 3. Ready Sol. A State 4. Descend
LS4 ↑ Coolant
ST4
D
ST4
ST4 D Sol. A
Coolant
LS2 ↑ Sol. B
ST5
State 5. Ascend
K E
IV. Trnslating the TN to PLC Code 3. Activation arcs are translated into latched outputs, and Turn-
A Ladder Diagram (LD) is a popular PLC code [3,4,5,6] which off arcs are translated into Unlatched outputs.
was chosen to illustrate the translation of the model to PLC
code. Today, most of the control programs are developed using
LD [7]. The generated LD rungs are arranged in three main V. CONCLUSION
blocks as follows: 1) events identification 2) transition triggers,
and 3) transition effects. The construction of the above three This paper presents a new modeling technique for discrete
blocks is presented next. control modeling technique that is intuitive (based on causes
and effects) and has a potential to be used also for simulation,
A. Events Identification validation, code generation and maintenance. The technique is
illustrated using a case study and shows a great potential. Such
Inputs and outputs change their voltage level when turned ON simplification could reduce the cost of discrete automation and
or OFF. These changes are referred as rising or falling edges. could contribute to progress towards more flexible automation.
The international standard IEC 61131-3 [8] defines special LD The proposed technique could be easily translated to code of
contacts for detecting rising and falling edges. A rising edge existing control equipment (PLCs). For each processing task, a
corresponds to a place with “↓” and a falling edge to a place simple state diagram that describes only the elementary actions
with “↑”. G of the system has to be built. From that point the process of the
B. Transition Triggers model generation could be automated, provided that the
following input is specified: the status of outputs for each state;
Each trigger activates one E-Transition. Each transition is the related inputs of each state, and supplementary trigger
assigned an internal variable in the LD. When the E-Transition identifiers.
is enabled that variable will be turned ON. In order to Future research includes implementation of the proposed
implement this logic, a set of rules is described as follows: technique on an industrial shop floor. This would enable
I. Each trigger forms an LD rung. simulation and visualization of operation; it would assist in real-
II. Each place (in E-Transition) that is input to a trigger forms a time tracking and failure analysis of the control system.
contact: (enable arc forms a normally open (NO) contact, and Currently the construction of a translator (model to LD) is
disable arc a normally closed (NC) contact. under way, to enable such future implementation.
III. The LD rung output is a variable that corresponds to the
activated transition.
REFERENCES
Figure 3 depicts a ladder diagram segment corresponding to the
triggers of transition D. These variables are used in figure 4.
[1] Ferdinand Wagner (Author), Ruedi Schmuki (Author), Thomas Wagner
(Author), Peter Wolstenholme (Author) "Modeling Software with Finite State
ST4 D Machines: A Practical Approach", AUERBACH, 2006
↑LS2 Trigger for
Transition D [2] Peng S. and Zhou M. C., "Sensor-based Petri net modeling for PLC stage
programming of discrete-event control design", Proceedings of IEEE
International Conference on Robotics and Automation (ICRA), Volume 2, 1907
Note: each output is activated for one - 1912, 2002.
scan only - ↑ or ↓ are inputs and they last [3] Manesis S. and Akantziotis K., "Automated synthesis of Ladder automation
one scanning cycle circuits based on state-diagrams", Advances in Engineering Software, Volume
36 , Issue 4, 225 – 233, 2005.
Fig. 4. Ladder Diagram segment for triggering transition D [4] Collins K., "PLC Programming for Industrial Automation", Exposure
Publishing, 2007
[5] Bolton W., "Programmable Logic Controllers", Fourth Edition, Newnes,
C. Transition Effects 2006.
The rules for establishing the ladder diagram portion of [6] Bateson N. R., "Introduction to Control System Technology" (7th Edition),
transition’s effects is as follows: Prentice Hall; 2001.
1. Dedicate a rung for each output place of the transition and [7] Lee G. B., Zandong H., Lee J. S."Automatic generation of ladder diagram
add to it a corresponding LD output. with control Petri Net", Journal of Intelligent Manufacturing, 15, 2, 245-252,
2. In each rung add a contact that corresponds to the relevant 2004.
transition. [8] Karl-Heinz J. and Tiegelkamp, M., IEC 61131-3: Programming Industrial
Automation System, w. CD-ROM : Concepts and Programming Languages,
Requirements for Programming Systems,6th edition, Springer, 2001.
Computational Modeling and Simulation
of a New Energy-Saving Sealing Unit
in Industrial Applications
Zheng (Jeremy) Li
University of Bridgeport, Bridgeport, CT 06604, USA
193
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_33, © Springer Science+Business Media B.V. 2010
194 LI
H * L = - Lg * Φ (2)
U 0 * Ag
Here,
H = intensity of magnetic field of magnet steel
L = length of magnet steel
Hg = intensity of magnetic field in working gap
Lg = length of working gap
Φ = magnetic flux
U0 = magnetic conductivity of vacuum
Ag = cross section area of working gap
The results of computational modeling and Currently the lubrication oil and gaseous media
simulation of this prototype are shown in Table 1 and leakages are common and difficult engineering
2. problems that affect the industrial machinery function
and performance. This new magnetic energy-saving
Table 1. Air leakage at different piston linear speed. sealing system has been developed to reduce the oil
and gaseous media leakages in industrial machinery.
Piston Linear Estimated Air All the theoretical, mechanical and magnetic
Speed (Ft/Min) Leakage (SCFM) analysis, computational modeling and simulation, and
prototype tests verify that this sealing system can
5 0.0005 significantly reduce the oil and gaseous media
leakages in industrial machinery. Its sealing
20 0.007 performance is reliable due to the firmly bonded and
strong forces between the magnetic particles and
35 0.012
reciprocating pistons/shafts. This seal mechanism is
45 0.023 also durable if compared with regular seals including
rubber seal, diaphragm seal, corrugated pipe seal.
60 0.034 This new sealing technology brings energy-saving to
the industrial machines because of higher machine
75 0.046 efficiency with less frictional force between surfaces
of seal and pistons/shafts and decreased internal
gaseous media contamination. Furthermore this
magnetic sealing product will contribute to the future
Table 2. Air leakage at different air pressure. exploitation, popularization, and application of the
rich rare-earth elements/materials.
Air Pressure Estimated Air
(PSIG) Leakage (SCFM)
50 0.001 References
[1]. H. Quelle, G. González-Jorge, L. Domarco, Magnetic field
200 0.005
cancellation on the surface of superconducting rings: Influence
350 0.009
on the flux creep, Cryogenics 48 (2008), pp. 2-5.
450 0.014 [2]. H. González-Jorge, B. Liñares, I. Quelle, E. Carballo and G.
197
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_34, © Springer Science+Business Media B.V. 2010
198 GHASSOUL
Motor drive: rotation whereas the panels are driven through pin
The drives for pilot and panels motors are identical RB5 for the forward rotation and RB6 for the return
except the amount of power delivered to each of rotation. When RB3 is activated, Q5 and Q6 are both
them. The only difference is that the pilot is driven closed, and the pilot motor is energized, and rotates
by the micro controller through its output port B pin in clockwise direction. When RB4 is high, Q7 and
RB3 for the forward rotation and RB4 for return Q8 are energized and the pilot motor rotates in
DESIGN OF AN INTELLIGENT SOLAR TRACKING SYSTEM 199
counter clockwise direction. Unlike RB3, RB4 is Sun rise and sun set capture:
connected to a set software flip/flop (refer to figure2) The sun rise and sun set detection is straightforward.
which drives the system back to the origin by hitting It is done through a light dependent resistor as shown
a proximity switch which resets the flip flop and the in figure (4). When sun rises, the LDR resistance
system comes to a stand still. decreases and the current through it increases. The
The same procedure takes place with the panels. voltage at the transistor base builds up until the
When RB5 is activated, they rotate forward and transistor is forced into saturation. The collector
when RB6 is activated, they rotate backwards. As it current in turn increases and the relay closes
is well known, the flywheel diodes are used to connecting the supply to the micro controller input
protect the micro controller from any current surge PORTA pin RA0. When darkness falls, the LDR
due the coil effects (refer to figure (3). resistor increases and the transistor base current
decreases cutting the collector current, and in turn
disconnecting the relay. The flywheel diode is
connected to protect the transistor against the di/dt
effect. An identical circuit is used for the panels but
the input is through controller pin
RA3
From microcontroller
possible was designed. Unlike what was reported in [2] "Design of PLC based tracking system to maximize solar
energy conversion" Ghassoul, M., proceedings of the 8th
literature, the main advantage of the system is that it IEEE TEM, KFUPM Dhahran,17-18 april 2001
is intelligent enough so that the panels track the sun [3] "28/40 pin 8 bit CMOS flash micro controllers", microchip
only if that contributes to extra energy extraction and datasheet 30292, 1998
at the same time, the energy consumed by panels [4] "RC model aircraft motor control", publisher J.,microchip
technology INC, 2002
driving motor is less than that extracted. The system [5] “Maximum power solar converter” Application AN1211,
can also align itself to perfection either on sunrise or Charais,J. Microchip Technology, 2008
sunset so no drift is allowed. Another main [6] “Low power capacitive sensing with the capacitive sensing
advantage is that it takes the days variation during module” Application AN1103, Floricica,A. Microchip
Technology, 2008
the year into consideration so to make sure that the [7] “Software handling for capacitive sensing” Ferme,T.
panels do not lead the sun during longer days nor lag Microchip Technology, 2007
it during shorter ones otherwise the system will [8] "Speed control of three phase induction
completely go out of synchronization. motor using PIC 18 micro controller"
Yedamale P., Microchip Technology INC, 2002
Finally It is hoped that the system could be further [9] "Brush DC servomotor implementation using PIC 17C756A",
looked at for commercial exploitation. Benling S., Microchip Technology INC,1999
[10] "Servo control of a DC brush motor", Bucella T., Teknic
INC,1997
References: [11] "Programmable logic controllers", Par E.A., 1996, published
[1] "A programmable logic system to control a solar panel to by Newnes
pump water from a well" Ghassou M. and Radwan F., 5th [12] "Programmable logic devices and logic controllers",
IEEE TEM, KFUPM Dhahran, pp 67-70 Manlado E. and Co, published by Prentice Hall, 1997
Adaptive Two Layers Neural Network
Frequency Controller for Isolated Thermal
Power System
Ognjen Kuljaca, Krunoslav Horvat*, Jyotirmay Gadewadikar
Systems Research Institute, Department of Advanced Technologies, Alcorn State University
1000 ASU Drive #360, Alcorn State, MS 39096, USA
[email protected], [email protected]
*
Brodarski Institute
Av. V.Holjevca 20, Zagreb, HR 10020, Croatia
[email protected]
Abstract: - An adaptive neural network control scheme for environment are subject to frequent parameters changes
thermal power system is described. Neural network control which may diminish the quality of control when
scheme presented in this paper does not require off-line nonadaptive controllers are used.
training. The online tuning algorithm and neural network
architecture are described. The performance of the It is thus natural to examine adaptive control schemes.
controller is illustrated via simulation for different changes Conventional adaptive control schemes, like one
in process parameters and for different disturbances. described in [7] can perform well. However, they require
Performance of neural network controller is compared with parameter identification and with fast and constant
conventional proportional-integral control scheme for changes this could pose problems in controller
frequency control in thermal power systems. Neural realization. Thus neural network control may be
network control scheme described here is not linear-in-
parameter. Neural network has two layers and nodes employed with no need of constant parameters
weights in both layers are tuned. identification. NN load-frequency control is described in
[8], [9] and [10]. The results obtained by using NN
I. INTRODUCTION controllers are good. However, the described controllers
require training. We provide here a performance analysis
The paper deals with the neural network (NN) of adaptive NN controller that does not require training.
frequency controller for isolated thermo power system. In The neural network is capable of on-line learning. The
modern power systems stable frequency is one of the controller described here is an improved version NN
main measures of power quality. Frequency control control scheme given for the first time in [30]. Unlike
becomes more and more important as power systems neural network controller given, our scheme here is
enter the era of deregulation. Frequency of the system augmented in such way that both layers of neural network
changes if change in load is not met with the adequate are tunable, not only output layer like in [30].
change in generation. It is becoming very hard, if not The paper is organized as follows. In Section II. are
impossible to schedule loads precisely, thus the load given some mathematical preliminaries. Model of
fluctuations in the power system are becoming more isolated thermo power system is given in Section III.
explicit. Also, it is becoming very hard to have Neural network control scheme is described in Section
determined and constant primary controllers and turbines. IV. In Section V. the simulation results are given and in
Emerging markets of ancillary services mean that primary Section VI. is given the conclusion.
controllers and turbines that are used in secondary control
change constantly. These changes can cause serious II. MATHEMATICAL PRELIMINARIES
problems when conventional control schemes are used,
including problems with stability, unless the primary Let R denote the real numbers, R n the real n-vectors,
turbines and controllers are carefully selected. To avoid R mxn the real mxn matrices. Let S be a compact simply
possible instability with changes of parameters in the connected set of R n . With map f : S → R m , let us define
system, conventional secondary controllers are
implemented using smaller integral gains than optimal C m (S) the space such that f is continuous. Let • be
performance would require. any suitable vector norm. The supremum norm of f ( x )
The literature about frequency and load – frequency over S is defined as:
control is numerous ([1], [2], [3], [4], [5], [6], [7], and
many others). Many non-adaptive schemes are given in
([1], [2], [3], [4], [5] and [6]. However, as described sup f ( x ) , f : S → R m , ∀x ∈ S (1)
before, the modern power systems in deregulated
203
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_35, © Springer Science+Business Media B.V. 2010
204 KULJACA ET AL.
Ks ,
Gs = (6)
y = W T σ( V T x ) (2) 1 + s ⋅ Ts
ΔPL
1 1
ΔPr - ΔPm - Δf
θv,1 Gg Gt Gs
+ +
z1 σ1
¦ w1,1 θ w,1
v1,1
x1 ¦ y1
z2 σ2 Fig. 2: The model of isolated thermo power
¦
system
x2 ¦ y2
z3 σ3
¦
where yg is output of the turbine controllers. These states = Fσ(V T z)Δf − k z FW.
W (13)
w
are physically available, so this representation will allow
for the NN control scheme design.
= Gz(σ(V′T z) T WΔf ) T − k z GV.
V (14)
IV. ADAPTIVE NEURAL NETWORK CONTROL v
protection functions take over in that case and some loads 0.1
have to be disconnected. 0
-0.1
Δf
-0.3
ª1 º -0.4
Let the control signal be given by Fig. 3: The step input PL=0.1 p.u.
-0.1
0.05
-0.2 0
Δf
Δf
-0.3 -0.05
-0.1
-0.4
-0.15
-0.5
-0.2
440 460 480 500 520 540 560 0 100 200 300 400 500 600 700 800 900
t [s] t [s]
0 0.005
-0.5
Δf
Δf
0
-1
-1.5
-0.005
-2
-2.5 -0.01
-3
-0.015
550 560 570 580 590 600 610 0 100 200 300 400 500 600 700 800 900
t [s] t [s]
-0.005
parameter neural network controller are shown. The
simulation results show that controller performs well and
-0.01
adapts to changing parameters. The controller does not
-0.015
require off-line training phase. Analysis of responses
0 100 200 300 400 500 600 700 800 900
t [s] shows that neural network control scheme outperforms
conventional control scheme. By defining the neural
Fig. 6: The responses for fL=0.002
network differently, having output layer weights W
defined as a matrix this scheme can be adjusted to deal
The response of the system for changed turbine with the multivariable systems. The performance analysis
governor and turbine parameters (Tg=0.09s and Tt=0.8s) here shows that is would be worth to continue with the
is shown in Fig. 6 and Fig. 7 . Again, NN control research effort toward two tunable layer neural network
outperforms PI controller. controllers for interconnected systems as well as for the
0.04
systems with generation rate constraint.
PI
NN
REFERENCES
0.03
0.02
[4] Fosha, C.E., Elgerd, O.I., "The Megawatt- [17] C. A. Desoer and S. M. Shahruz, “Stability
Frequency Control Problem: A New Approach Via of dithered nonlinear systems with backlash or
Optimal control Theory" IEEE Transactions Power hysteresis,” Int. J. Contr., vol. 43, no. 4, pp. 1045-
Apparatus and Systems, PAS-80, No. 4, pp. 563-577, 1060, 1986.
1970 [18] B. Friedland, Advanced Control System
[5] Hamza, M.H., Agathoklis, P., Chan, W.C., Design, Prentice-Hall, New Jersey, 1996.
"A Combined Selftuning and Integral controller for [19] B. Igelnik and Y. H. Pao, “Stochastic
Load Frequency Control of Interconnected Power Choice of Basis Functions in Adaptive Function
systems "Regelungstechnik 30. Jahrgang Heft 7 pp. Approximation and the Functional-Link Net,” IEEE
226-232, 1982 Trans. Neural Networks, vol. 6, no. 6, pp. 1320-1329,
[6] Kundur, P., Power System Stability and Nov. 1995.
Control. McGraw-Hill, Inc.,1993 [20] K. Hornik, M. Stinchombe, and H. White,
[7] Liaw, C.M., "Design of a Reduced-Order “Multilayer feedforward networks are universal
Adaptive Load-Frequency Controller for an approximators,” Neural Networks, vol. 2, pp. 359-
Interconnected Hydrothermal Power System 366, 1989.
"International Journal of Control, Vol. 60, No. 6, pp. [21] B.S. Kim and A.J. Calise, "Nonlinear flight
1051-1063, 1994 control using neural networks," AIAA J. Guidance,
[8] Demiroren, A., Zeyneigly, H.I., Sengor, Control, Dynamics, vol. 20, no. 1, pp 26-33, 1997.
N.S., "The application of ANN technique to load- [22] J.-H. Kim, J.-H. Park, S.-W. Lee, and E. K.
frequency control for three-area power system”, 2001 P. Chong, “Fuzzy precompensation of PD controllers
IEEE Porto Power Tech Conference, Porto, Portugal for systems with deadzones,” J. Int. Fuzzy Syst., vol.
[9] Bevrani, H. "A novel approach for power 1, pp. 125-133, 1993.
system load frequency controller design", [23] S.-W. Lee and J.-H. Kim, “Control of
Transmission and Distribution Conference and systems with deadzones using neural-network based
Exhibition 2002: Asia Pacific. IEEE/PES , Volume: 1 learning control,” Proc. IEEE Int. Conf. Neural
, 6-10 Oct. 2002 Networks, June 1994, pp. 2535-2538.
[10] Birch, A.P., Sapeluk, A.T., Ozveren, C.T., [24] F. L. Lewis, C. T. Abdallah, and D. M.
"An enhanced neural network load frequency control Dawson, Control of Robot Manipulators, Macmillan,
technique", International Conference on Control, New York, 1993.
1994. Control '94. Volume 1, 21-24 Mar 1994 [25] F. L. Lewis, A. Yesildirek, and K. Liu,
[11] R. Barron, “Universal approximation “Multilayer neural-net robot controller with
bounds for superpositions of a sigmoidal function,” guaranteed tracking performance,” IEEE Trans.
IEEE Trans. Info. Theory, vol. 39, no. 3, pp. 930-945, Neural Networks, vol. 7, no. 2, pp. 1-11, Mar. 1996.
May 1993. [26] F. L. Lewis, K. Liu, and A. Yesilidrek,
[12] A.G. Barto, "Reinforcement learning and “Neural net robot controller with guaranteed tracking
adaptive critic methods," Handbook of Intelligent performance,” IEEE Trans. Neural Networks, vol. 6,
Control, pp. 469-492, Van Nostrand Reinhold, New no. 3, pp. 703-715, 1995.
York, 1992. [27] F. L. Lewis, K. Liu, R. R. Selmic, and Li-
[13] J. Campos and F.L. Lewis, "Adaptive critic Xin Wang, “Adaptive fuzzy logic compensation of
neural network for feedforward compensation," Proc. actuator deadzones,” J. Robot. Sys., vol. 14, no. 6, pp.
American Control Conf., vol. 4, pp. 2813 –2818, San 501-511, 1997.
Diego, 1999 [28] F.L. Lewis, S. Jagannathan, and A.
[14] F.-C. Chen, and H. K. Khalil, “Adaptive Yesildirek, Neural Network Control of Robot
control of nonlinear systems using neural networks,” Manipulators and Nonlinear Systems, Taylor and
Int. J. Contr., vol. 55, no. 6, pp. 1299-1317, 1992. Francis, Lindon, 1998.
[15] M. J. Corless and G. Leitmann, “Continuous [29] N. Sadegh, “A perceptron network for
state feedback guaranteeing uniform ultimate functional identification and control of nonlinear
boundedness for uncertain dynamic systems,” IEEE systems,” IEEE Trans. Neural Networks, vol. 4, no. 6,
Trans. Automat. Contr., vol. 26, no. 5, pp. 850-861, pp. 982-988, Nov. 1993.
1982. [30] Kuljaca, O., Lewis, F. Tesnjak, S., "Neural
[16] G. Cybenko, “Approximation by Network Frequency Control for Thermal Power
superpositions of a sigmoidal function,” Math. Contr. Systems", IEEE Control and Decision Conference,
Signals, Syst., vol. 2, no. 4, pp. 303-314, 1989. 2004
Optimal Controller Comparison Using Pareto Fronts
David Moore
Department of Electrical Engineering
University of Cape Town
Rondebosch
South Africa
Abstract – New design methods in Control System are The framework proposed in this paper aims to make as
regularly proposed. These new methods are typically compared broad an analysis of controller applicability to a specific
to existing methods in a focussed manner than highlights certain problem as possible. Hence it uses a posteriori method,
criteria, but may neglect other criteria. This paper deals with the Differential Evolution (DE) [8], where as little subjective
use of Level Diagram visualisation for Pareto Fronts as a device
for comparing modern control schemes in a robust and objective
inference as possible is added before the decision maker is
manner. Specifically it uses a comparison of two methods: PI presented with information. DE is a stochastic population-
and Sliding Mode Control as a test case. based multi-objective optimizer. It uses real numbers and the
Keyword – Multiobjective Optimization, Differential evolution based ideas of crossbreeding and mutation to find
Evolution, Pareto Front, Level Diagrams, Sliding Mode Control. solutions in the form of sets that are mutually optimal. Such
solutions are typically represented as Pareto Fronts.
Pareto Fronts are traditionally visualised on a 2-D plot of
cost vs. cost showing the trade-offs that can be made between
I. INTRODUCTION
various costs 2]. When the front exists in more than 2
New methods are constantly being proposed for the design dimensions it can become difficult to analyse [1] and even
of Control Systems (e.g. VSC in 1970s, H-infinity in the more difficult to visualize.
1980s, etc). More often than not, the new methods are Level Diagrams 3] are a novel method for visualization of
compared to previous ones in a focused manner that high dimensional Pareto Fronts that overcomes this limitation.
highlights one or two criteria. This can be misleading in some Most MOO are coupled to a decision maker and attempt to
cases [l] and a more comprehensive comparison is required. find a single or directed set of solutions to a given problem
This paper presents a framework for comparing Control [2] rather than presenting a visualization of the trade-offs.
design methods using high dimensional Pareto fronts, that are
computed using Multi-Objective Optimization (MOO) 2] and II. METHOD OVERVIEW
investigated using the novel Level Diagram [3] visualisation. The success of a control method is typically dependent on the
Specifically it compares the Sliding Mode Control 4] and PI
specifics of its design and implementation. It makes sense to
Control methods in terms of many criteria. compare two methods only when both are very well designed.
To ensure that the comparisons are unbiased, the In order to ensure this, optimization is used to find the best
framework uses the best controllers that each method can controller that a method can produce, based on the costs that
produce. The performance of the controllers is quantified by have been defined for the problem.
cost functions (or costs) that are assumed to identify the best
controllers. These costs are frequently and typically mutually
A. Multiobjective Optimization
exclusive.
Multiobjective Optimization finds the best costs by
Multi-Objective Optimization methods are utilized to find
simultaneously optimizing two or more conflicting objectives
optimal (best) solutions to a given problem [2] where optimal
subject to certain constraints 2]. It can be formalized
is defined in terms of the trade-offs between computed cost
mathematically as follows
functions and the concept of Pareto Efficiency 2].
A Pareto Efficient or Optimal situation is one where any 1
change to make a member better off is impossible without
making some other member worse off [2]. Solutions that are
Pareto Efficient are described as non-dominant 5][6].
Many different types of MOO methods exist; they fall into
four groups depending on when the decision maker adds
preference information. These four groups are priori (before), where is the input or decision vector, is the decision
progressive (during), posteriori (after), and no preference 7]. space and is the cost or objective vector. This
209
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_36, © Springer Science+Business Media B.V. 2010
210 MOORE
Figure 2 gives the Level Diagrams for the same set given in
Figure 1.
An example of Level Diagrams with higher dimension, (in
Figure 1: 2-D Pareto Front this case 7 costs) is shown in Figure 7 and Figure 8.
(4)
(6) (7)
SMC is meant to be insensitive to parameter changes and The motor speed is regulated by the PI and (Integrated)
be able to completely reject matched disturbances [11] due to SMC controllers.
the non-linear term .
Complete SMC design methods are presented in both B. Cost Definitions
[10][4]. The Pareto Differential Evolution algorithm works by
One of the inherent difficulties of SMC is chatter that is optimizing the outputs (i.e. costs) for a given set of inputs (i.e.
induced by non-ideal switching. Minimization of this chatter decision variables). The choice of these costs affects the
is one of the primary focuses of research on SMC. A review outcome of the optimization.
of methods for reducing chatter is presented in [11]. In order to keep the comparison as general as possible, the
For this paper a pragmatic engineering approach, called costs were chosen to be as application agnostic as possible.
Integrated SMC, was used to deal with chatter. The method The first three costs are based on the Integral Square Error
works by augmenting the SMC controller with an integrator (9) of the plant error e (8), resulting from changes in the
on its output. The output of this integrator is the manipulated setpoint r, input v and output d disturbances introduced to the
variable for the plant. There are three main reasons for the system.
integrator’s inclusion: Firstly, SMC design techniques call for
systems with greater than first order dynamics, so first order (8)
systems need augmenting. Secondly, the integrator acts as a
filter, removing the need for any other chatter correction. And (9)
finally it eliminates the finite offset error that would occur
with first order plant models due to input disturbances. Similarly, the next three costs are based on the Integral
One expected result from this modification is that Square Error for the plant control u (ISU). It is modified such
integrating the SMC output reduces the controller’s that the resulting cost is a finite integral. This is done by
effectiveness at rejecting disturbances due to the delayed subtracting u’s final value from u at each time step.
response that the integrator introduces.
(10)
III. COMPARISON DEFINITIONS Both the and costs are abbreviated to
As stated above, PI and SMC control are compared to one and in the diagrams that follow.
another using Level Diagrams that are computed by the The last cost included is a rough measure of the robustness
Pareto Differential Evolution Algorithm. Furthermore, a set of the controlled system to model changes. It finds the largest
of PI controllers that cancel the plant poles and lead to poor positive percentage change for each of the 6 other costs when
internal performance [12] was included for extra comparative the plant parameters A and T are varied by 50%. In order for
purposes. Figure 3 shows the general plant setup with setpoint this measure to be useful on the Level Diagrams, the base ten
r, input v and output d disturbances. logarithm of this value is used.
C. Optimizer Setup
The optimizer was run for 20 generation with a crossover
rate and mutation rate of 0.15.
212 MOORE
(11)
A. General Results
During testing it was found that neither PI nor SMC
dominated the other on any significant portion of their fronts.
This indicates that both Pareto Fronts exist in a mutually
exclusive region of the n-dimensional space as shown in
Figure 6.
OPTIMAL CONTROLLER COMPARISON USING PARETO FRONTS 213
Finally, as the proportional gain Kp for the PI controller While Integrated SMC degrades the disturbance rejection of
increases above around 5, the controller norm worsens. SMC, it still retains the other robust characteristics of SMC.
As such, it is a useful method for chatter avoidance in
V. CONCLUSIONS situations where its inferior disturbance rejection is deemed
A number of general conclusions can be drawn with regard acceptable.
to both the comparison of PI and SMC design methods, and Level Diagrams proved to be a useful tool for visualisation
the use of Level Diagrams. and visual analysis of optimal controller sets. As with other
As expected, SMC proved far more robust in general that visualisation techniques, Level Diagrams would benefit from
PI, and was clearly shown using Level Diagrams. interactive features that cannot be easily shown in print.
Including more costs does not prevent the phenomenon of
pole-zero cancellation; including a set of pole-zero cancelling
controllers in the Level Diagrams does help with identifying
controllers that avoid this phenomenon.
REFERENCES
[1] M. Rapson, "Pareto Analysis of Controller Design Methodologies for [7] Johan Andersson, "A Survey of Multiobjective Optimization in
Intergrator plus Dead Time Processes," CISSE, 2007. Engineering Design," 2000.
[2] J.B. Yang, J.F. Whildborne G.P. Liu, Multiobjective Optimization and [8] Kenneth Price Rainer Storn, "Differential Evolution - A simple and
Control.: Research Studies Press LTD., 2002. efficient adaptive scheme for global optimization over continuous
[3] J.M. Herrero, J. Sanchis, M. Marínez X. Blasco, "A new graphical spaces," 1995.
visualization of n-dimensional Pareto front for decision-making in [9] Cecil L. Smith, Digital Computer Process Control. Scranton: Intext
multiobjective optimization," Information Sciences: an International Education Publishers, 1972.
Journal, vol. 178, no. 20, pp. 3908-3924, October 2008. [10] W. Li J.J.E. Slotine,.: Prentice Hall, 1991, ch. 7, pp. 276-309.
[4] S.K. Spurgeon C. Edwards, Sliding Mode Control: Thoery and [11] V.I. Utkin, Ü. Özgüner K.D. Young, "A Control Engineer's Guide to
Applications.: CRC Press, 1998. Sliding Mode Control," IEEE Trans on Control Systems Technology
vol. 7, no. 3, pp. 328-342, 1999. ,
[5] Carlos A. Coello, "Twenty years of Evolutionary Multiobjective
Optimization: A Historical View of the Field," IEEE Computational [12] P.A. La Plante, Comprehensive dictionary of electrical engineering
Intelligence Magazine, vol. 1, no. 2, pp. 28-36, 2006. CRC Press & IEEE Press, 1999. .:
[6] Ruhul A. Sarker Hussein A. Abbass, "The Pareto Differential Evolution [13] Keith J. Burnham Ali J. Koshkouei, "Control of DC Motors Using
Algorithm," International Journal on Artificial Intelligence Tools, vol. Proportional integral sliding mode,".
11, no. 4, pp. 531-552, 2002.
Interaction with Objects in Virtual
Environments
Jesús A. Verduzco-Ramírez1, Enrique Araiza-Montes-de-Oca2, Daniel A. Gaytán-Valencia1, Nicandro Farías-Mendoza1
and Bruno Raffin3
1
Departamento de Sistemas y Computación, Instituto Tecnológico de Colima, Av. Tecnológico #1, C.P. 28976. Villa de
Álvarez, Colima, México
2
Facultad de Telemática, Universidad de Colima, Av. Universidad #333. Colonia Las Víboras. C.P. 28040. Colima,
Colima, México
3
INRIA, LIG, Avenue J. Kuntzmann, 38330 Montbonnot, France.
Abstract— The applications of virtual and augmented reality them within the environment in which the user can examine,
require high performance equipment normally associated with manipulate and interact with the objects shown [2].
high costs, inaccessible for small organizations and educational
In addition to the VR, there is the technology of augmented
institutions. By adding computing power and storage capacity of
several PCs interconnected by a network we have an alternative reality (AR), which consists of a collection of devices that
of high performance and low cost. The objective of this project is enhance the capabilities of human senses [3]. The AR is a
to implement an augmented reality platform integrated with technology that integrates elements of the environment
items available in the laboratories of an educational institution, typically captured in video format and audio with three-
such as PCs, webcams, Ethernet networking and free software. dimensional computer graphics objects resulting in new
This platform is an attractive and economical alternative for the
consistent, supplemented and enriched worlds [4].
development and implementation of virtual and augmented
reality applications to enable low-budget institutions simulate Both technologies are closely related but have differences.
processes, physical phenomena, weather, etc. In this first phase, VR allows immersion of the user in an artificial world that
we have a prototype that creates a scene consisting of a set of completely replaces the user's real world. The RA maintains
virtual objects, the video stream acquired from a Webcam and the user's real world enriched by the presence of virtual items.
processes the captured image into a graphic with a set of virtual The applications of this technology are found in the following
objects. A user facing the angle of acquisition of the webcam can areas: military, medical, recreational, archaeological and
interact with virtual objects in the scene with only the move of his
hand. Although our prototype is reduced to only two computers pedagogical.
the first in which images are acquired through the webcam and There are several platforms that use this approach
the second where the mixture, processing and visualization of the successfully; one of the most outstanding is Grimage [5],
scene is made, these early results demonstrate the feasibility of which performs 3D modeling of images acquired by multiple
our idea and are a strong motivation for continuing the project. cameras, physical simulation and parallel execution that offers
As future work we have planned to increase the number of
a new immersive experience. On this platform it is possible to
Webcams, to explore the interaction with complex virtual scenes
as well as distributing the processing of the scene among a group capture objects to be instantaneously modeled in 3D and
of computers to increase the size of the scene displayed. entered into a virtual world populated by rigid or flexible solid
objects with possible interaction with them. Grimage requires
Index Terms—Augmented reality, virtual reality, interactive for their operation high-performance equipment. The existing
applications, computer graphics. configuration at the INRIA Rhone Alpes has six high
definition cameras, a wall of images with 16 video projectors
and two interconnected clusters, one with 11 dual-Xeon 2.6
I. INTRODUCTION GHz for the application and another with 16 dual-Opteron 2
215
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_37, © Springer Science+Business Media B.V. 2010
216 VERDUZCO-RAMÍREZ ET AL.
The software used in Grimage, FlowVR [6] is a VR API images and sends them in a video stream to the rendering
that offers support for heterogeneous codes executed in node.
parallel. FlowVR allows the development and implementation As a video server we selected Spook [9], a free software
of interactive applications in high performance computing that allows the delivery of video streams over IP networks.
clusters and grids. FlowVR reuses and extends the data flow Spook takes the images stored in the buffer, compresses the
paradigm commonly used for scientific visualization data in sequential order, and sends them as RTP packets
environments [7, 8]. through the network. In the rendering node, the
One of the limitations of the AR platforms such as Grimage is client starts the MPlayer immersed in FlowVR. Thus,
that they are made up of costly and inaccessible equipment for it is possible to display the images captured by the webcam in
educational institutions that have a small budget. the rendering the node.
Our approach aims to avoid such problems. Our proposal is to The mixture of images captured with the webcam and a 3D
develop an augmented reality platform that integrates low-cost scene component is provided by the FlowVR Render, which
equipment such as Ethernet, PC, Webcams and free software. allows the combination of the video stream to a virtual
One use expected to achieve with the implementation of this environment rendered by FlowVR.
platform is that teachers of an educational institution can
simulate physical models and manipulate virtual objects in III. IMMERSION OF OBJECTS
applications that are intended to support teaching.
A. Description of our Application
II. ARCHITECTURE OF OUR SOLUTION To test the functionality of our platform, we have developed
some examples that demonstrate the recognition of the
To implement a platform for low-cost RA, our solution is contours of the objects captured by the webcam and the
divided into four functional modules, as shown in Figure 2: interaction with virtual objects without using any device of
• One or more webcams to capture the video stream. interaction. For both operations we used the OpenCV library
• A module for the acquisition and distribution of images [10].
captured with a webcam.
• A rendering module to mix the flow of video and
B. Contours detection
interactive 3D scenes.
• A display wall for viewing the results. A contour is a list of points representing the curvature of an
image. This representation may be different depending on the
circumstances. There are many ways to represent them. In
OpenCV the contours are represented by sequences where
each entry in that sequence encodes information about the
location of the next point on the curve.
To get to show the contours of an image, we first need a
method for managing memory. OpenCV has an entity called
Memory Storage, which is a method for managing the location
of dynamic objects in memory. Any object of OpenCV that
requires the use of memory as part of its normal operation,
needs a Memory Storage. The Memory Storage is a linked list
of memory blocks for fast data storage and retrieval.
The type of object that can be stored in a memory storage is
Fig. 2. Interacting with virtual objects in Grimage.
a sequence. The sequences are linked lists of other data
structures.
In this first phase, to verify the feasibility of our idea we have
The first developed example is the recognition of a series of
implemented a prototype consisting of only two nodes. letters and figures drawn by hand. Examples are shown in
A. Acquisition Node Figure 3.
The acquisition node connects one or more web cameras,
which acquire images of the environment. A video server
stores and transmits these images to the rendering node.
B. Rendering Node
The rendering node is running an instance of FlowVR, it
receives the video frames, makes the rendering and subsequent
display through MPlayer immersed in FlowVR.
Fig. 3. Detection of contours.
C. Operation
As described in Section A, the node captures
INTERACTION WITH OBJECTS IN VIRTUAL ENVIRONMENTS 217
V. FUTURE WORK
The results obtained so far encourage us to continue this
project. We want to increase the number of nodes involved in
the rendering scenes, the number of cameras allowing the
acquisition of more than one video stream, to add textures to
Fig. 4. Architecture of the applications developed. objects displayed on the interaction scene and study the
system performance in these applications.
• The interaction scene: represented by a set of images from
the camera. REFERENCES
• The application scene: represented by a set of images in [1] C. Cruz-Neira, D. J. Sandin, T. A. DeFanti, R. V. Kenyon, and J. C.
which there are some features of the scene of interaction. For Hart. The Cave Audio Visual Experience Automatic Virtual
example, contours, corners and colors. Environment. Communication of the ACM, 35(6):64_72, 1992.
• Finally, the virtual objects embedded in scenes of interaction [2] J. Hilera, S. Otón, J. Martínez, “Aplicación de la Realidad Virtual en la
Enseñanza a Través de Internet”, Universidad de Alcalá, septiembre de
and application, which react to events within the application 1999. [Online]. Disponible en:
stage and are represented through images or textures in the http://www.ucm.es/info/multidoc/multidoc/revista/num8/hilera-oton.html
scene of interaction. [Accesado en diciembre 10, 2008].
[3] J. Ahrens, C. Law, W. Schroeder, K. Martin, and Michael Papka, A
Parallel Approach for Efficiently Visualizing Extremely Large, Time-
To test the interaction with virtual objects, we developed a Varying Datasets.
couple of examples intended to show how a user can interact http://www.acl.lanl.gov/Viz/papers/pvtk/pvtkpreprint/.
with them. [4] L. Heras, J. Villareal, “La Realidad Aumentada: Una Tecnología en
The first application shows a set of bubbles that the user can Espera de Usuarios”, Revista Digital Universitaria, vol. 5, no. 7, Agosto,
2008.
exploit to make moves on the interaction scene. The [5] J. Allard, C. Menier, B. Raffin, E. Boyer, F. Faure. 2007. Grimage:
application scene is the difference between a frame image and Markerless 3D Interactions. ACM SIGGRAPH 2007 emerging
its predecessor. Thus, a white shadow is formed as shown in technologies.
Figure 5. [6] J. Allard, V. Gouranton, L. Lecointre, S. Limet, E. Melin, B. Raffin and
S. Robert. 2004. FlowVR: a Middleware for Large Scale Virtual Reallity
The second application example shows a set of balls Applications. In Euro-Par 2004 Parallel Processing: 10th International
bouncing in the interaction scene. When the user touches one Euro-Par Conference, 497-505. http://flowvr.sf.net.
or when they collide with each other, they change their [7] Covise Programming Guide, 2001.
direction. http://www.hlrs.de/organization/vis/covise.
[8] S. Jordà. Realidad Aumentada. Taller de Sistemes Interactius 2003 –
The application scene consists of the contours of the figures 2004. Universitat Pompeu Fabra.
appearing in the interaction scene, which upon contact with [9] Spook Video Live Streamer http://www.litech.org/spook/ “OpenCV
virtual objects, they are running a routine to change its Wiki”, IBM, 2002. [Online]. Disponible en:
direction of motion. http://opencv.willowgarage.com/wiki/ [Accesado en mayo 11, 2009]
Both example applications are shown in figure 5.
Abstract - One approach for training of recurrent neural design of IIR digital filters with given amplitude and phase
network model of 1-D IIR digital filter is proposed. The responses is discussed in [11]. A method for 2-D recursive
sensitivity coefficients method has been applied in the training digital filters design is investigated in [12], where the design
process of the neural network. The set of time domain data is problem is reduced to a constrained minimization problem the
generated and used as a target function in the training
procedure. The modeling results have been obtained for two
solution of which is achieved by the convergence of an
different cases - for 4-th order bandpass IIR digital filter and for appropriate neural network.
partial response IIR digital filter. The frequency domain Some methods for the non-linear digital filters design
behavior of the neural network model and the target IIR filter
using neural networks are considered in [3]. A design method
has been investigated. The analysis of the frequency responses
shows good approximation results.
for nonlinear adaptive digital filters using parallel neural
networks is given in [4].
219
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_38, © Springer Science+Business Media B.V. 2010
220 STEFANOVA
II. STRUCTURE AND DESCRIPTION OF THE NEURAL NETWORK The equations (2), (3) can be used to describe the recurrent
neural network shown in Fig. 1.
The two layer recurrent neural network is considered. The
The following additional vectors have been defined:
structure of neural network is shown in Fig. 1.
• a vector of the first layer inputs of neural network
The recurrent neural network is described with the
recurrent system of equations as follows: z (k ) = [ x1 (k ), x 2 (k ),...x n (k ), u1 (k ), u 2 (k ),..., u nu (k )]T ,
ª x1 (k ) º z ∈ R nz (4)
« » • a vector of the neural network outputs
ª w11 w12 ... w1n ...w1nz º « x 2 (k ) » y (k ) = [ y1 (k ), y 2 (k )... y m (k )]T
ª s1 (k ) º «...........»
(5)
« » « »
« w21 w22 ... w2 n ...w2 nz » « » •
« s 2 (k ) » a vector of the neural network weighting coefficients
«........ »
= «...................................... »
« »
« x n (k ) » ,
«u (k ) » [
p = w11 w12 ...w1nz , w21 w22 ...w2 nz ....wn nz , c1 , c2 ...cm , ]
« » « 1 »
¬« s n (k )¼» « wn1 wn 2 ... wn n ...wn n » p∈R
np
¬ z ¼ «...........» (6)
« » where n p = n.n z + m is the number of neural network
¬«u nu (k )¼»
(1) coefficients and the elements of matrix W ∈ R n × nz are
ª x1 (k + 1) º ª f ( s1 (k )) º introduced row by row.
« » « » Mean square error objective function is defined in
« x 2 (k + 1) » = « f ( s 2 (k )) » k = k 0 , k 0 + 1,..., k f − 1 , following form:
«........... » «........ »
« » « » 1 k f −1 m
«¬ x n (k + 1)»¼ «¬ f ( s n (k ))»¼ J (p) = ¦ ¦ [ yi (k ) − yˆ i (k )] ,
2
(7)
2 k =k0 i =1
where f(.) is the activation function for the first layer of the where { yˆ i (k )} is a set of samples of the target .
neural network
Let be introduced the vector function f(s) as follows: III. ALGORITHM OF SENSITIVITY COEFFICIENTS
f(s) = [ f ( s1 ), f ( s 2 ),..., f ( s n )]T , The main problem in the neural network training process is
the gradient calculation of the mean square objective function
then the neural network description in the matrix form is: (7) with respect to weights.
x (k + 1) = f ( W z (k )), x(k 0 ) = x 0 , (2) For that reason, it is necessary to determine the gradient
components of this objective function.
y (k ) = Cx(k ) , (3)
∂J ( p) k f −1 m ∂y
= ¦ ¦ [ y i (k ) − yˆ i (k )] i , l=1,…,np (8)
∂pl k = k0 i =1 ∂pl
x1 ( k )
For the first layer of the neural network the derivatives
∂y i
w11 s1 (k ) x 1 (k + 1) x 1 (k ) ŷ(k ) have been determined from (9) as follows:
¦ f (s1) z −1 ∂pl
x 2 (k ) x2(k +1)
¦
s 2 (k )
z −1
x 2 (k )
Cx (k )
y( k )
∂yi (k ) ∂x (k )
= ci i , i = 1,, m , l = 1, nw , nw = n.n z (9)
f (s2 )
∂pl ∂pl
x n (k )
xn (k +1)
x n (k ) ∂xi (k )
s n (k )
¦ f (sn ) z −1
The derivatives from (9) can be obtained from
wnn ∂pl
u 1 (k )
the recurrent system:
Fig. 1. Recurrent neural network structure for digital filter modeling n ∂x ( k + 1) f ∂x j (k )
∂xi (k + 1) ∂xi (k + 1) f
where = +¦ i ;
∂pl ∂pl j =1 ∂x j ( k ) ∂pl
x ∈ R n is a state space vector,
; i = 1,..., n ; l = 1,.., n w ; (10)
u ∈ R nu is a vector of neural network input excitations,
∂xi (k 0 )
y ∈ R m is a vector of neural network output, =0 ; k = k 0 , k 0 + 1,..., k f − 1
∂pl
W∈R n × nz
and C ∈ R m ×n
, m ≤ n , C = cij , [ ] where
c i = j ∂xi (k +1) f z j . f ′(si ) ; i = q = 1,, n ; j = 1,, nz
cij = ® i are matrixes of system coefficients, =® (11)
¯0 i ≠ j ∂wqj ¯0 ; i≠q; j = 1,, nz
n - is a number of neurons at the first layer, ∂xi (k + 1) f
nu - is a number of input excitations, = wij f ′( si ) ; i = 1, , n ; j = 1, , n (12)
∂x j (k )
n z = n + nu - is a number of neural network inputs.
ONE APPROACH FOR TRAINING OF RECURRENT NEURAL NETWORK MODEL 221
and the derivatives (11), (12) are determined from the neural V. MODELING RESULTS
network shown in Fig. 1 in the case when network structure
Case 1. Modeling of 4 th order bandpass IIR digital filter
has been considered as a feedforward neural network.
For the second layer of the neural network the The following requirements to the magnitude response of
∂y the target bandpass digital filter are specified:
derivatives i have been determined as follows: passband [300–3400] Hz, passband loss – 3 dB; stopband [0–
∂pl
100] Hz and [3600–4000] Hz, stopband loss – 15 dB;
∂y1 (k ) ∂y (k ) sampling frequency – 8000 Hz.
= x1 (k ) , , m = xm (k ) (13)
∂p nw +1 ∂p n p The training sequences are generated using the impulse
Then the objective function (7) can be minimized response of target IIR digital filter shown in Fig. 2. This
target IIR digital filter has been designed by MATLAB’s
applying the standard optimization procedure.
There are two ways to realize the training algorithm of Filter Design Toolbox.
the neural network. The first one, described above, insists on The model of the bandpass IIR filter is realized as
recurrent neural network structure with 4 neurons.
determining the solution of the recurrent system (1) and
calculating the summary error from (7) for all time The effectiveness of the proposed algorithm is
samples k = k 0 , k 0 + 1,... . Simultaneously with this demonstrated by simulation with neural network model. The
calculation process, the sensitivity coefficients from (10) have simulation is realized applying input harmonic excitations
to be calculated and the gradient components of the objective with frequencies in the digital filter’s passband and stopband.
function (7) have to be obtained using equations (8). After These sinusoidal signals are different from the set of data
that the objective function (7) has to be minimized applying used in the training process of neural network model. The
standard gradient procedure. This training procedure supposes impulse responses of the target digital filter and the neural
the processing of all calculated data i.e. the training process is network model are shown in Fig. 2 and Fig. 3.
off-line. The second way to train the neural network requires
calculation of the error function (7) on the neural network
output by solving the system (1) for every time sample
k = k 0 , k 0 + 1,... . The gradient (8) of this error function has
to be determined using the sensitivity coefficients obtained
from system (10). The next operation is a correction of the
neural network weights using the components of the error
gradient (8). This training process is recurrent or on-line
training which can be realized in the real time. The possibility
of on-line realization of the training procedure is the main
advantage of the sensitivity coefficients algorithm in
comparison with the Lagrange multipliers training method,
described in [16], when the training process can be realized
only off-line.
IV. NEURAL NETWORK MODEL OF IIR DIGITAL FILTER
The IIR digital filter can be considered as a linear discrete Fig 2. Impulse response of the target bandpass IIR filter
dynamic system described in time domain with n – order
difference equation that has been stated using the delayed
samples of the input excitation and the response signal at the
output. Transforming the difference equation the state space
description of the IIR digital filter can be obtained as a linear
system of “n” number of first order recurrent equations. The
advantages of the state space representation are the simplicity
of time domain analysis, the matrix form description of digital
filter impulse and step responses and possibility of parallel
calculations.
The neural network structure shown in Fig. 1 must be
modified to linear structure. In this case the activation
function f(x) is linear. The corresponding system of equations
(1)-(3) is linear recurrent system. The sensitivity coefficients
algorithm modified for linear network is used as a training
procedure of the neural network model. Fig. 3. Impulse response of the neural network model
222 STEFANOVA
The magnitude responses in frequency domain of the The comparison between the input harmonic excitation
target IIR bandpass digital filter and neural network model with frequency 700Hz chosen from digital filter’s passband
are shown in Fig. 4 and Fig. 5, respectively. and the output reaction of neural network model is illustrated
in Fig. 7.
Case 2. Modeling of partial response IIR digital filter MAGNITUDE RESPONSE OF THE TARGET FILTER
MAGNITUDE
interpolation multirate systems. To achieve zero ISI, Nyquist
0.6
filter must satisfies some criteria in time domain that they
should have zeros equally spaced in the impulse response
coefficients except one specified. 0.4
°̄0 , otherwise
Fig 11. Magnitude response of the target IIR partial response filter
where K and N are integers
The transfer function H ( z ) of the (IIR) Nyquist filter can
be expressed as:
Nn
¦ bi z − i
−K i = 0 , i ≠ kN = K 1
H ( z ) = bK z + Nd / N
, bK = (15)
−i N N
¦ ai N z
i =0
where N n , N d are integers, all the filter coefficients ai , bi are
real, a0 = 1 , N d is the multiple of N.
The specifications for the IIR partial response filter are the
following: N n = 10, N d = 2 , N = 2 , K = 6 [17].
The impulse response of the partial response filter, shown
in Fig. 10, is used as target function in the training procedure
of the neural network model. The magnitude response of the
partial response IIR digital filter has a cosine shape which is
illustrated in Fig. 11.
The impulse response and magnitude response of the
Fig 12. Impulse response of the neural network model of
neural network model with 6 neurons of IIR partial response IIR partial response filter
filter are shown in Fig.12 and Fig. 13 respectively.
MAGNITUDE RESPONSE OF THE NEURAL NETWORK MODEL
0.8
MAGNITUDE
0.6
0.4
0.2
0
0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1
NORMALIZED FREQUENCY
Fig 10. Impulse response of the target IIR partial response filter
Fig.13. Magnitude response of the neural network model
of IIR partial response filter
224 STEFANOVA
VI. CONCLUSION [7] Zeng Zhe-zhao; Wen Hui, “Optimal design study of three-type FIR
high-order digital filters based on sine basis functions neural-network
algorithm”, IEEE International Symposium on Communications and
One approach for the IIR digital filter modeling based on Information Technology, Vol. 2, No 12-14 ,pp. 921- 924, October 2005.
recurrent neural network is proposed. The sensitivity [8] X. H. Wang, Y. G. He, H. Li, Y. L. Peng, M.J. Li, “Design of 2-D
coefficients method has been used to develop an effective digital filters using back propagation neural networks”, Proceedings of
International Conference on Machine Learning and Cybernetic, 18-21
training algorithm for the neural network model. The training Aug. 2005, Vol. 8, pp 4684 – 4689.
process also can be on-line and has a real time realization. [9] Wang D. and A. Zilouchian, Intelligent control systems using soft
The partial response IIR and bandpass IIR digital filters have computing methodologies, Editors: Ali Zilouchian, Mo Jamshidi,
been modeling using the recurrent neural network structures. Prentice Hall, March 2001, pp, 93-110.
[10] Chow, T.W.S. Siu-Yeung Cho “An accelerated recurrent network
To investigate the behavior of the digital filter models some training algorithm using IIR filter model and recursive least squares
simulations have been realized using input harmonic method” IEEE Transactions on Circuits and Systems I: Fundamental
excitations with different frequencies. The analysis of the Theory and Applications, Vol. 44, No 11, Nov, 1997, pp. 1082-1086.
obtained results shows a very good approximation of the [11] Bhattacharya, D.; Antoniou, A., “Design of IIR filters with arbitrary
amplitude and phase responses by feedback neural networks”, IEEE
impulse responses in time domain and the magnitude International Symposium on Circuits and Systems, ISCAS ’96, Vol. 3,
responses in the frequency domain. This approach can be 12-15 May 1996, pp. 457 – 460.
extended and successfully applied to the case of modeling of [12] Mladenov, V.M., Mastorakis, N.E., “Design of two-dimensional
nonlinear recursive digital filters. recursive filters by using neural networks”, IEEE Transactions on
Neural Networks, Vol. 12, No 3, May 2001, pp. 585 – 590.
REFERENCES [13] W. T. Miller, R. S. Sutton and P. J. Werbos, Neural Networks for
Control., Cambridge, MA:MIT Press, 1990.
[1] Jin L., P. Nikiforuk, and M. Gupta, “Approximation of discrete-time [14] R. Grino, G. Cembrano and C. Torras, “Nonlinear system identification
state-space trajectories using dynamical recurrent neural networks”, using additive dynamic neural networks – two on-line approaches”,
IEEE Transactions on Automatic Control, Vol.40, No 7, 1995. IEEE Trans. on Circuits and Systems I, vol. 47, no.2, 2000, pp. 150-
[2] H. Cruse , Neural Networks as Cybernetic Systems, Brains, Minds and 165.
Media, Bielefeld, Germany, October 2006. [15] Y. Fang, M. Yagoub, F. Wang and Qi-J. Zhang, “A new macromodeling
[3] M. Pedersen, “Optimization of recurrent neural networks for time approach for nonlinear microwave circuits based on recurrent neural
domain series modeling”, Ph.D Thesis, Department of Mathematical networks”, IEEE Trans. on Microwave Theory and Techniques, Vol.
Modeling, Technical University of Denmark, 1997. 48, no. 2, Dec. 2000, pp. 2335-2344
[4] J. Cao, Yahagi T., “Nonlinear adaptive digital filters using parallel [16] Stefanova S., “One Dimensional IIR digital filter modeling based on
neural networks”, Proceedings of IEEE International Conference on recurrent neural network”, International Joint Conferences on
Neural Networks, Vol. 2, Nov-Dec 1995 pp. 850 – 853. Computer, Information, and Systems Sciences, and Engineering
[5] D. Bhattacharya, and A. Antoniou, Real-time design of FIR filter by (CISSE 08), University of Bridgeport, Bridgeport, U.S.A, December 5 -
feedback neural network, IEEE Signal Processing Letters, Vol.3, No 4, 13, 2008.
May 1996, pp. 158 – 161. [17] S. Pei, Ch. Hsu, P. Wang, “Design of a class of IIR eigenfilters with
[6] Zhe-Zhao Zeng; Ye Chen; Yao-Nan Wang, ”Optimal Design Study of time and frequency domain constraint, IEEE Trans. on Circuits and
High-Order FIR Digital Filters Based on Neural-Network Algorithm”, Systems-II Analog and Digital Signal Processing, vol. 49, No 2,
International Conference on Machine Learning and Cybernetics, 13-16 February 2002, pp 145 - 151.
Aug. 2006, pp 3157 – 3161.
Classification of the Power Transformers using
Dissolved Gas Analysis
T. C. B. N. Assunção, J. T. Assunção and A. D. Neto
Department of Electric Engineering, Federal University of São João del-Rei (UFSJ), Brazil
(e-mail: [email protected], [email protected], [email protected]).
225
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_39, © Springer Science+Business Media B.V. 2010
226 ASSUNÇÃO ET AL.
approximation functions. The main characteristic of LS-SVM AR = Arcing; CT = Sparking; DPCO = Partial Discharge with
is smallest computational cost in relation to the ANN, without Tracking (note CO).
loss in the quality of the solutions. TABLE II
FAULTS DESCRIPTION TABLE – ROGERS METHOD.
In this paper, a comparative study of ANN, LS-SVM and
DGA for the calculation of efficiency for the incipient fault
CH4/ C2H2/ C2H4/ C2H2/
detection is presented. The results of this study are useful in Diagnostic
H2 CH4 C2H6 C2H4
the development of a reliable transformer diagnostic system
0 0 0 0 N
using intelligent systems and DGA.
5 0 0 0 DP
1 or 2 0 0 0 OH1
II. INCIPIENTS FAULTS DIAGNOSTIC
1 or 2 1 0 0 OH2
0 1 0 0 OH3
The power transformers in operation are subject to failures
0 1 0 0 OHC
due to electric and thermal stresses and the stresses due to
1 0 1 0 CCE
environmental influences. The effect of any stress results in
1 0 2 0 CCTC
the ageing of insulation, it leads up to decomposition of the
0 0 0 1 FL
insulating oil and paper, with the generation of dissolved
0 0 1 or 2 1 or 2 ARC
gases in the insulation oil.
0 0 2 2 CT
The oil dissolved gases can signify a possible failure in the
5 0 1 or 2 1 or 2 DPCO
power transformers [1- 2].
The gases more generated inside power transformers during
operation normal or during the degradation of isolation system
are: Hydrogen (H2); Methane (CH4); Ethane (C2H6); Ethylene Doernenburg Method is also based in ratio between gases,
(C2H4); Acetylene (C2H2); Carbon Monoxide (CO) and but, unlikely Roger’s Method, its ratio between gases is not
Carbon Dioxide (CO2); and they can indicate: codified. In this method, the relative value of gases is
associated with the fault conditions, in accordance with the
• Hydrogen (H2): Corona;
Table III. In the Doernenburg Method is not considered the
• Hydrogen (H2); Methane (CH4); Ethane (C2H6);
normal state, but there are restrictive conditions for
Ethylene (C2H4): Oil Thermal Decomposition;
application of this method, as showed in the Table IV. If at
• Carbon Monoxide (CO) and Carbon Dioxide (CO2: least one of the gases in each ratio between gases exceeds the
Paper Thermal Decomposition; limit L1, the ratio key gases procedure is valid, otherwise, if it
• Acetylene (C2H2): Electric Arc. is not significant d the unit should be resample and
investigated by alternative procedures.
In this paper were studied the methods more used for the
dissolved gas analysis, as follows [3-5]. TABLE III
RATIO KEY GASES – DOERNENBURG METHOD.
Rogers Method uses the following relations of gases:
CH4/H2 ; C2H2/CH4; C2H4/C2H6; C2H2/C2H4. In the Tables I Range Gases Ratio Range Gases Ratio
and II are presented the relations between gases and the Fault CH4/ C2H2/ C2H6/ C2H2/
codification related to fault type of the Rogers Method. H2 C2H4 C2H2 CH4
TABLE I
TD >1 < 0,75 > 0,4 < 0,3
RATIO KEY GASES – ROGERS METHOD DPL < 0,1 NS > 0,4 < 0,3
DPH < 1 to
Gases Ratio Range Gases Ratio Code > 0,75 < 0,4 > 0,3
> 0,1
< 0.1 5
0.1 to1.0 0 where: NS = Not Significant; TD = Thermal Decomposition;
CH4/H2
1.0 to 3.0 1
> 3.0 2
DPL = Partial Discharge (Low-Intensity); DPH = Partial
< 1.0 0 Discharge (High-Intensity).
C2H2/CH4
> 1.0 1
< 1.0 0 TABLE IV
C2H4/C2H6 1.0 to 3.0 1 LIMIT CONCENTRATIONS OF DISSOLVED GAS [PPM].
> 3.0 2
< 0.5 0 Gas H2 CH4 CO C2H2 C2H4 C2H6
C2H2/C2H4 0.5 to 3.0 1 L1 100 120 350 35 50 65
> 3.0 2
the likely fault type produced by combination of gas prevention or after the occurrence of these faults, but they will
concentration. have a little sensibility in relation to incipient faults. In spite
In the Duval Triangle (Figure 1), PD = Discharge Partial; of, DGA conventional methods can be used as standards for
T1 = Thermal Fault, Temperature < 300º C; T2 = Thermal the intelligent systems based in the artificial intelligence [2, 8-
Fault, 300 º C < Temperature < 700º C; T3 = Temperature 12]. The intelligent systems applications for the incipient
>700º C; D1 = Discharges of Low Energy; D2 = Discharges of fault diagnostic in transformers are useful, because the
High Energy; DT = Mixture of Electrical and Thermal Faults. intelligent systems are adaptative, capable of handling highly
non-liner relationship and also generalize solutions for a new
solution.
III. INTELLIGENT SYSTEMS FOR FAULTS In these paper is proposed a tool for power transformers
DIAGNOSING classification in relation to their failure condition, using
dissolved gas analysis.
The conventional methods of interpretation of the DGA are DGA methods, ANN and LS-SVM approaches were
easy implemented and present good results for the hard faults studied and implemented for verification of operation
228 ASSUNÇÃO ET AL.
condition of power transformer. The data of this study were Normal State; OH = Overheating; FE = Electric Failure; LS =
obtained from electricity utility, and each method was tested LS-SVM; NI = Fault is Unidentifiable; R = Rogers Method;
with all the data set. DV = Duval Method; DB = Doernenburg; Method; NBR =
NBR 7274 Method; IEC = IEC Standard 6059 Method.
IV. 1. ANN IMPLEMENTATION
TABLE VII
ANN used is a multi-layer perceptron (MLP) with 5 CONCENTRATION OF GASES IN PPM.
neurons in the input layer, 12 neurons in the intermediate
layer and 3 neurons in the output layer. The activation Concentration of Gases [ppm]
TR
functions used in each of the layers were, respectively, tangent H2 CH6 C2H4 C2H2 C2H6
sigmoid function, tangent sigmoid function and logarithmic 1 40 62 11 1 51
sigmoid function. MLP ANN doesn’t presents binary output, 2 600 1800 3800 130 520
so it was used the winner neuron strategy to make the 3 6 2 7 0,4 5
classification. As the MLP used is composed of 3 output 4 1230 163 233 692 27
neurons, then it was determined that: 1. the output pattern of 5 630 670 1100 1700 81
the neuron that to present higher value is defined as 1; 2. the 6 20 3 5 0,4 2
output pattern of the remaining neurons is defined as 0. 7 24 3 8 0,4 4
Because designing an innovative ANN learning procedure is 8 30 520 1000 0,4 310
beyond the scope of this paper, the ANN used in this paper 9 645 86 110 317 13
was implemented using Neural Toolbox MATLAB routines 10 110 860 1500 1800 300
[13]. 11 19 9 7 0,4 20
12 16 1 1 0,4 1
IV. 2. LS-SVM IMPLEMENTATION 13 1600 55000 74000 9300 42000
14 95 10 11 39 0,4
Classification with the LS-SVM approach was made with 15 200 110 150 230 12
two networks. The first network was used to determine the 16 13 10 55 0,4 3
transformer actual operation condition (normality or fault) and 17 48 12 18 0,4 14
the second network was used to determine the fault type 18 44 95 13 0,4 33
(heating or electric failure). Because designing an innovative 19 595 80 89 244 9
LS-SVM learning procedure is beyond the scope of this paper, 20 94 5 25 20 4
the LS-SVM network used in this paper was implemented 21 290 16 12 0,4 24
using the LS-SVMlab Toolbox MATLAB routines [14]. 22 56 6 7 10 5
23 420 1400 1500 7 640
V. PRESENTATION AND ANALYSIS OF RESULTS 24 1790 580 336 919 321
25 0 1200 1900 4400 130
In order to training of the ANN and LS-SVM were used in 26 14 3 12 0,4 3
this study 246 samples of DGA and were selected and used 27 8800 64064 9565 0,4 72128
data of 30 transformers for ANN and LS-SVM testing. The 28 34 41 5 0,4 69
same data were used for the comparison between the results of 29 1330 10 66 182 20
the traditional methods, ANN and LS-SVM. In this study 30 230 120 220 740 14
were used five key gases, all of them combustible, as inputs
features of networks: H2, CH4, C2H2, C2H4 e C2H6.
Overheating, electric failure or normal are the output features. TABLE VIII
DIAGNOSIS ACCURACY RESULTS.
Then, input and output patterns for each diagnosis criterion
are defined according each method (Doernenburg Ratio
Method Diagnosis Accuracy [%]
Method, Rogers Ratio Method, IEC 60599, IEEE C57.104-
Rogers 33.33
1991 and Duval Triangle. The concentration of gases data in
Duval 40.00
ppm are shown in the Table VII.
Dornenbur 50.00
Table VIII shows the accuracy results for the incipient fault
g
detection of the DGA conventional methods, ANN and LS-
NBR 7274 50.00
SVM. It is demonstrated in the Table VIII the superior
IEC 60599 50.00
efficiency of the ANN methods over the other methods.
LS-SVM 70.00
Rogers’s method presented 33.33% efficiency, Dorneburg’s
MLP 83.33
method presented 50%, NBR7274 presented 50%, ANN
presented 83.33% and LS-SVM presented 70%. These results
VI. CONCLUSION
demonstrate that intelligent systems (ANN and LS-SVM) can
be a promising tool in the detection of incipient faults in
power transformers. A comparative study was made employing intelligent
The resulting tests employing the DGA conventional systems (ANN and LS-SVM) and DGA conventional methods
methods, ANN and LS-SVM approaches, are shown in the (Doernenburg Ratio Method, Rogers Ratio Method, IEC
Table IX, where: TR = Transformer; D = Diagnostic; N = 60599, IEEE C57.104-1991 and Duval Triangle) for the
CLASSIFICATION OF THE POWER TRANSFORMERS USING DISSOLVED GAS ANALYSIS 229
efficiency calculation for incipient fault detection on power ANN due to its generalization characteristic is an adequate
transformers. The MLP and LS-SVM were trained with the tool for to solve the conflicts between methods. The results
gases relation of the DGA conventional methods cited presented show the actual capacity of the ANN for the
previously. identification of incipient fault in power transformers and also
to solve the standardized inconsistencies of the methods. In
TABLE IX spite of the superior results obtained with MPL, it was verified
RESULTS SET DISPOSED BY DGA CONVENTIONAL METHODS,
that with the default parameters of the respective algorithms
ANN AND LS-SVM.
the implementation of the LS-VM model is easier than the
MLP model. MLP network involves more experience for its
METHODS modeling and training, mainly for the definition of the number
TR D
R DV DB NBR IEC MLP LS of hidden layers. Therefore, LS-SVM would be studied as an
1 N SA NI N OH OH OH OH important alternative to ANN and DGA conventional methods
2 N NI NI OH OH OH OH FE for the identification of incipient fault in power transformers.
3 OH NI FE N OH OH N N We intend to continue the studies on the application of LS-
4 FE FE FE NI FE FE FE FE SVM in modeling and simulation for the identification of
5 FE NI FE NI NI NI FE FE incipient fault in power transformers, because it showed to be
6 N OH FE N OH OH N N a useful tool in this study.
7 OH NI FE N OH OH N N
8 FE NI FE OH OH OH OH FE ACKNOWLEDGMENT
9 FE FE FE NI FE FE FE FE
10 N NI FE NI NI NI FE FE The authors acknowledge CNPQ – Brazil, for the financial
11 N OH FE N N N N N support.
12 OH NI FE N NI NI N N
13 FE NI FE OH NI NI OH FE REFERENCES
14 FE FE FE N FE FE FE FE [1] Assunção, T. C. B. N. Contribuição à Modelagem e Análise do
15 N FE FE NI FE FE FE FE Envelhecimento de Transformadores de Potência, Tese de Doutorado,
Universidade Federal de Minas Gerais, UFMG, 2007.
16 N NI FE N NI NI N N [2] Singh, J., Sood, Y. R., Jarial, R. K., Verma, P., Condition Monitoring of
17 OH NI FE N OH OH N N Power Transformers - Bibliography Survey, IEEE Electrical Insulation
18 FE OH FE N OH OH N OH Magazine, vol. 24, n° 3, May/June 2008, pag. 11-25.
19 FE FE FE NI FE FE FE FE [3] IEEE Guide for the Interpretation of Gases Generated in Oil-Immersed
Transformers, C57.104-1991, 1992.
20 N NI FE N NI NI FE FE [4] Mineral Oil-Impregnated Electrical Equipment in Service, Guide to the
21 N NI FE N FE FE N FE Interpretation of Dissolved and Free Gases Analysis, IEC Standard, 1999.
22 OH FE FE N FE FE FE FE [5] Interpretação da Análise dos Gases de Transformadores em Serviço, NBR
23 FE NI FE OH OH OH OH FE 7274, ABNT, 1982.
[6] Haykin, S. ; Redes Neurais Princípios e prática, 2ª edição, Prentice Hall
24 FE FE FE NI FE FE FE FE INC., ARTMED Editora S. A., Bookman, 1999.
25 N FE FE NI FE FE FE FE [7] Suykens, J. A. K., L., L., Doren, P. V. e Moorand, B., V. J., Least squares
26 OH NI FE N NI NI N N support vector machines classifiers: A large scale algorithm·, European
27 OH NI FE OH OH OH OH FE Conference on Circuits Theory and Design, pp. 839-842.
[8] Moreira, M. P., Santos, L. T. B., Vellasco, M. M. B. R., Power
28 OH OH FE N OH OH OH OH Transformers Diagnosis Using Neural Networks, Proceedings of International
29 FE NI FE FE NI NI FE FE Joint Conference on Neural Networks, Orlando, Florida, August 2007..
30 FE FE FE NI FE FE FE FE [9] Guardado, J. L., Naredo, J. L., Moreno, P., Fuerte, C. R., A Comparative
Study of Neural Network Effciency in Power Transformers Diagnosis Using
Dissolved Gas Analysis, IEEE Transactions on Power Delivery, vol. 16, nº 4,
The main problem of the DGA conventional methods is that October 2001, pag. 643-647.
different methods applied to the same sample result in [10] Yanming, T., Zheng, Q., DGA Based Insulation Diagnosis of Power
different and often in contrary diagnostic decisions. An Transformers Via ANN, Proceedings of The 6 th International Conference on
adequate utilization of the DGA conventional methods is Properties and Applications of Dielectric Materials, June 2000, Xi'an Jiatong
University, China.
important; because most of expertise’s in incipient fault [11] Farag, A. S., Mohandes, M., Al-Shaikh, A., Diagnosing Failed
diagnostics use these methods for elaboration of expertness of Distribution Transformers Using Neural Networks, IEEE Transactions on
power transformers operation condition. The diagnostic made Power Delivery, vol. 16, n° 4, October 2001, pag. 631-636.
using DGA conventional methods often are based in the [12] Salami, A., Pahlevani, P., Neural Network Approach for Fault Diagnosis
Transformers, 2008 International Conference on Condition Monitoring and
empiric knowledge of the expertises, and then it is necessary Diagnosis, Beijing, China, April 2008.
to mind mainly with the inconsistency (non-decision [13] Demuth, M. B. H. e Hagan, M., Neural network toolbox, for use with
problems) and the conflicts between the methods. MATLAB, The Math Works;
[14] Pelckmans, K.; Suykens, J. A. K, LS-SVMlab toolbox, 2003, Version 1.5.
Modelling Control of Pore Number and Radii
Distribution in Single-Cell Electroporation
Sadhana Talele Department of Engineering
University of Waikato
Hamilton, New Zealand
Email: [email protected]
Paul Gaynor Department of Electrical and Computer Engineering
University of Canterbury
Christchurch, New Zealand
Abstract—Electroporation EP, in which external electric field on the type of application of EP it would be good to know the
pulses create transient pores in a cell membrane, is an important starting parameters of electric field to be applied in order to
technique for delivery of genes and drugs into the cell. To enable achieve a good number of pores in a particular range of size,
a useful level of entry of genes into cells, the pores should have
sufficiently large radii, and remain open long enough without causing as pores that are too small are not of any therapeutic use and
membrane rupture. A numerical model for a single spherical cell reduce the strength of the cell membrane, while pores that are
electroporated by application of direct and/or alternating external too big create a possibility of them becoming permanent leading
electric field pulses has been developed. The model is used to indicate to cell lysis. To date, such controllability of pore number and size
the actual number of pores and their radii distribution developed has not been reported in the literature.
in response to various electric field pulses, as function of time and
position on the cell surface. This study briefly describes the model Previous reported studies have modelled EP to calculate Vm
briefly which is then used to investigate the ability to control the and the pore density N [10], [11]. These models consider the non-
number and distribution of pore radii by choice of electric field linear behavior of EP, which is due to the dynamics of membrane
parameters. We believe this would be one of the first papers to pore formation and its effect on the electrical properties of the
investigate the ability to CONTROL the number and distribution membrane in turn. However, these models do not include the
(range) of pore radii (as opposed to other papers that merely report
the pore number and range with varying pulse parameters. evolution of pore sizes. They assume a constant pore radius.
Recent models [12], [13] include spatial and temporal aspects of
Index Terms—electroporation, numerical model, fractional pore
pore radius evolution. However, the form of electric fields used
area, pore radii
in these models were single unipolar pulses. The study presented
here develops a model of single spherical cell EP and provides
I. I NTRODUCTION simulation of spatial and temporal aspects of pore radius as an
Exposure of biological cells to electric fields can lead to effect of multiple pulses of varied magnitudes. As such it is
a variety of biophysical and biochemical responses [1]–[3]. possible to investigate potential control of both the number of
Electroporation (EP) is a feature common to all lipid bilayers pores created and their radii distribution.
in which electrically conductive pores are induced to form in
the membrane, when the transmembrane potential Vm exceeds a II. M ODEL OF A SINGLE CELL
semi-critical value creating a state of high permeability [4], [5]. Consider a spherical cell of radius a with intracellular conduc-
These pores shunt excess stimulus current across the membrane, tivity σin . The cell is immersed in a medium with conductivity
limiting the growth of Vm . The pores are long lived, sometimes σex . This system is exposed to a time-varying electric field E(t).
surviving in the membrane for up to several minutes, and provid- Azimuthal symmetry about the axis of applied electric field is
ing pathways for the movement of large ions and high molecular assumed. The model discussed in [14] is described below briefly
weight molecules such as drugs and DN A into the cell [6]–[8]. for completeness.
These properties have resulted the application called EP which is
now a common tool in biotechnology and is used in a number of
medical treatments [1]–[3], [6], [9]. However, the process of EP is A. Transmembrane potential
still not completely understood. There is a need for supplementing It is assumed that the intracellular and extracellular regions are
experimental knowledge with theoretical models [10]. Depending charge free. The homogeneous external electric field of strength
231
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_40, © Springer Science+Business Media B.V. 2010
232 TALELE AND GAYNOR
E(t) is used as a boundary condition, the potential being fixed pore, n is the relative entrance length of the pore and vm is the
at nondimensional transmembrane potential [10] given by
Φ(3a, θ) = −3aE(t) cos θ, (1)
F
on a sphere of radius 3a surrounding the cell, where θ is the polar vm = Vm . (10)
RT
angle. Since there are no sources inside the region, the potential
obeys Laplace’s equation This equation for pore current isml (equation 8) accounts for the
∇2 Φ = 0, (2) electrical interactions between the ions and the pore wall [10].
Now, assume Q larger pores exist, and ilge is the current through
except at the cell membrane at a radial distance of (r = a) where the electropores of radius larger than 1 nm. Then, Jlge is the total
the potential is discontinuous because of the abrupt change in current density through Q larger pores; rq being the radius of the
conductivity and the approximated infinitely thin membrane. Thus q th pore. Hence,
internal and external potentials can be defined as,
1
Q
Φin (r, θ) r < a, Jlge = ilge (rq ) (11)
Φ(r, θ) = (3) A q=1
Φex (r, θ) a < r < 3a.
The current across the membrane is used to relate the internal where A is the corresponding cell surface area. For these larger
and external potentials [10], that is, pores, the current–voltage relationship assumes that the trans-
∂Vm membrane potential Vm occurs across the sum of pore resistance
−r̂ · (σin ∇Φin (a, θ)) = −r̂ · (σex ∇Φex (a, θ)) = Cm + Jm
∂t Rp and the series input resistance Rin [13], [15], as follows:
(4)
where Vm
Vm (θ) = Φin (a, θ) − Φex (a, θ). (5) ilge (r) = , (12)
Rp + Rin
Here r̂ is the unit outward radial vector, Cm is the specific
membrane capacitance, Vm is the transmembrane potential and where
Jm is the current density at the cell membrane due to existing h
Rp = , (13)
pores. The current density is made up of three terms as given in πσps rq2
Equation 6
Jm = Jion + Jsml + Jlge , (6) and
1
where Jion = gm (Vm − Vrest ) is the ionic current density [10] (gm Rin = . (14)
2σps rq
is the specific membrane conductance, and Vrest is the membrane
rest potential). The remaining two terms are explained in the
following paragraphs. B. Formation of pores:
Jsml is the current density through small pores and is given by
Initially pores are assumed to be formed with the minimum-
Jsml = N isml (r), (7)
energy radius rm = 0.76 nm, at a rate given by [10]
where N is the pore density of the initial small pores formed and
isml (r) is the diffusion current through a single pore of radius r dN 2 N
= ψe(Vm /Vep ) 1 − , (15)
(true for small pores only). A previously derived expression for dt Neq (Vm )
isml [10], based upon the Nernst-Planck equation models, is used
for pore radius below 1 nm, and is, where N is the pore density of the initial small pores formed,
2
ψ is the creation rate coefficient and Neq is the equilibrium pore
πr σps νm RT (e − 1) νm
isml = · (8) density for a voltage Vm given by
Fh (G− eνm − G+ )
2
with Neq (Vm ) = N0 eb(Vm /Vep ) . (16)
w0 ew0 ±nνm ± nνm
G± = . (9)
w0 ± nνm Here, N0 is the initial pore density with no applied electric field,
Here σps is the conductivity of the aqueous solution that fills the Vep is the characteristic voltage of electroporation and b is the pore
√
pore (approximated by σin σex ), F is Faraday’s constant, R is creation constant equal to (rm /r∗ )2 , where r∗ is the minimum
the universal gas constant, T is the absolute temperature, h is the radius of hydrophilic pores and rm the minimum-energy radius
thickness of the membrane, w0 is the energy barrier inside the [16]. All parameter values are as given in Table 1.
MODELLING CONTROL OF PORE NUMBER AND RADII DISTRIBUTION 233
electric field
As seen in Table 2, the first pulse was chosen to be of amplitude
θ12 = 90˚
r51
135 kV/m with a time duration of 3 μs, as this gave a reasonable
∆r FPA and a maximum pore radius of 10 nm. The aim is to increase
r50
the total FPA, as well as the contribution of FPA by pores in the
range greater than 5 nm, preferably larger than 10 nm. Several
amplitudes for the second pulse were tried (135kV, 175kV, and
rα + 1
195kV) with a break of 3 μs between two pulses.
···
rα θ2 = 15˚
rα – 1 As seen in Table 3, the second pulse of 195 kV/m, increases
θ1 = 7.5˚
the total FPA substantially to 0.05%. The total number of pores
···
r1
∆θ
θ24 = 180˚ θ0 = 0˚ remain nearly the same (a few new pores form as an effect of the
a second pulse), while increasing the number of pores in the range
5-10 nm from 660 to 862, pores in the range 10-20nm from 275
cell membrane
to 376 and pores in the range 20-100 nm from 0 to 14. As seen
2a
in Table 3, the FPA contribution due to large pores in the pore
radius range 5-100 nm has increased from about 18% of total to
about 27 % of total FPA.
In an attempt to increase this ratio, a higher amplitude second
pulse was tried (results not shown). This, instead of increasing
the size of pores, formed new smaller pores. Using a higher
magnitude first pulse hasve been seen to give a higher total FPA
Fig. 1. Spherical cell model geometry showing the angle and radius discretization but with more small pores (under 5 nm).
used in the numerical model
Another way of reaching the target is to use the pulses for a
longer time. The time duration of the first pulse (135kV/m) and
the second pulse (195kV/m) was increased to 4 μs, with different
electric field on pore size distribution and the fraction of cell
break times of 3 μs, 8 μs and 26 μs. The results are shown in
surface area occupied by pores, defined as the fractional pore
Table 4 to see the effects on FPA and pore sizes. The total FPA
area (FPA). Several simulations were carried out and important
increases in both cases, so does the contribution of FPA by pores
results tabulated as shown in Tables 2-5. These tables give
larger than 5 nm. A shorter break ensures that the initial pores
information about the final pore numbers, pore density, %FPA,
do not shrink too much and thus a few pores are able to expand
maximum mean radius, maximum overall radius, and the pore
to a size larger than 20 nm. It is not possible for these pores to
size distribution. The numbers on right hand side of the pore
exist with a longer break. However, they could be possible with
numbers is the %FPA contribution due to corresponding pores
longer first and/or second pulse duration.
in that row. Electric field applied is represented as [135 0 135 ]
To determine if use of a much lower amplitude first pulse or
for a time duration [0 3 6 9] . This represents an electric field
reversing the sequence of the higher and the lower amplitude
pulse of 135 kV/m applied from 0 to 3 μs, a break from 3 to
pulses would have any favorable result, more simulations were
6 μs, followed by an electric field of 135kV/m from 6 to 9 μs
done and results are shown in Table 5. As seen in the first column
and so on. The magnitude of first pulse was chosen such that it
of Table 5, the number of large pores and their FPA contribution
is just enough to create pores resulting in an FPA of more than
(radius larger than 5 nm) drastically reduces although the total
0.03%, as the appropriate level of electroporation is in the range
FPA is similar. Column 4 of Table 5 indicates that while a very
of 0.01% to 0.1% [26]. The magnitude and the time duration of
small (85 kV/m) first pulse followed by a high amplitude second
second pulse were varied to investigate its effect on the number of
pulse (195 kV/m) gives a reasonable value of total FPA, and
pores in the range 10-20 nm. The model was used to investigate
a good number of large pores (5-10 nm), it does not support
the mechanism /usefulness of the two-pulse protocol for control
formation of many pores in the range (greater than 10 nm).
of pore size and number distribution. Different electric field
amplitudes for different time durations were used as the second
pulse for simulations, keeping other parameters like cell radius, V. C ONCLUSION
internal and external conductivity, etc., constant. The final extent There seems to be an interdependent relationship between the
of electroporation (measured by FPA) was targeted to be about pulse amplitude, pulse duration, number of pulses and the time-
0.05 % , with at least over one tenth of the FPA contributed by gap between pulses in order to control the total FPA and the
pores with radii larger than 5 nm, making sure that the number desired pore radii distribution. Generally, the first pulse seems
of these pores larger than 5 nm in radius is also at least 2% of to control the number of pores, and the second pulse helps to
the total number of pores on a spherical cell of radius 15 μm. grow these pores to a larger size. This is in agreement with the
MODELLING CONTROL OF PORE NUMBER AND RADII DISTRIBUTION 235
Table 2
Table 3
Table 4
Table 5
theoretical research by citeneu2003 and experimental findings by [13] K. C. Smith, J. C. Neu, and W. Krassowska, “Model of creation and
[27], who reported better uptake of DNA for a two pulse protocol evolution of stable electropores for DNA delivery,” Biophysical Journal,
vol. 86, pp. 2813–2826, 2004.
compared with a single pulse protocol. The break between the two [14] Talele, S., P. Gaynor, J. van Ekeran, and M. J. Cree. Modelling single cell
pulses will let the initial pores reduce in size. Thus if the break electroporation with bipolar pulse: Simulating dependence of electroporated
is too long, the second pulse tends to form new pores and has to fractional pore area on the bipolar field frequency. In: Proceedings
International Joint Conferences on Computer, Information, and Systems
expand the initial pores from a smaller size, thus ending up with Sciences, and Engineering. USA (2008).
a comparatively smaller size of pores than if a shorter break was [15] J. Newman, “Resistance for flow of current to a disk,” Journal of The
used. Electrochemical Society, vol. 113, pp. 501–502, 1966.
[16] J. C. Neu and W. Krassowska, “Modeling postshock evolution of large
electropores,” Physical Review E, vol. 67, pp. 1–12, 2003.
R EFERENCES [17] J. C. Weaver and Y. Chizmadzhev, “Theory of electroporation: a review,”
Bioelectrochemistry and Bioenergetics, vol. 41, pp. 135–160, 1996.
[1] M. Cemaẑar, D. Miklavĉiĉ, and G. Serŝa, “Intrinsic sensitivity of tumor cells [18] J. Israelachvili, Intermolecular and Surface Forces. London, UK: Academic
to bleomycin as an indicator of tumor response to electrochemotherapy,” Press, 1992.
Japanese Journal of Cancer Research, vol. 89, pp. 328–333, 1998. [19] J. C. Neu and W. Krassowska, “Asymptotic model of electroporation,”
[2] G. Serŝa, “Electrochemotherapy,” in Electrochemotherapy, Electrogenether- Physical Review E, vol. 59, pp. 3471–3482, 1999.
apy, and Transdermal Drug Delivery : Electrically Mediated Delivery of [20] J. C. Weaver and K. T. Powell, “Theory of electroporation,” in Electropo-
Molecules to Cells (Methods in Molecular medicine), M. J. Jaroszeski, ration and Electrofusion in Cell Biology, E. Neumann, A. E. Sowers, and
R. Heller, and R. Gilbert, Eds. Totowa, New Jersey: Humana Press, 2000, C. A. Jordan, Eds. New York: Plenum Press, 1989, ch. 7, pp. 111–126.
ch. 6, pp. 119–133. [21] I. G. Abidor, V. B. Arakelyan, L. V. Chernomordik, Y. A. Chizmadzhev,
[3] G. Serŝa, S. Novaković, and D. Miklavĉiĉ, “Potentiation of bleomycin V. F. Pastushenko, and M. R. Tarasevich, “Electric breakdown of bilayer
antitumor effectiveness by electrotherapy,” Cancer Letters, vol. 69, pp. 81– membranes: I. The main experimental facts and their qualitative discussion,”
84, 1993. Bioelectrochemistry and Bioenergetics, vol. 6, pp. 37–52, 1979.
[4] R. W. Glaser, S. L. Leikin, L. V. Chernomordik, V. F. Pastushenko, , and [22] M. Kanduŝer, M. Foŝnaria, M. Ŝentjurc, V. Kralj-Igliĉ, H. Hägerstrand,
A. I. Sokirko, “Reversible electrical breakdown of lipid bilayers: Formation A. Igliĉ, and D. Miklavĉiĉ, “Effect of surfactant polyoxyethylene glycol
and evolution of pores,” Biochim Biophys Acta, vol. 940, p. 275287, 1988. C12 E8 on electroporation of cell line DC3F,” Colloids and Surfaces A, vol.
[5] T. Kotnik, D. Miklavĉiĉ, and T. Slivnik, “Time course of transmembrane 214, pp. 205–217, 2003.
voltage induced by time-varying electric fields: a method for theoretical [23] A. Barnett and J. C. Weaver, “Electroporation: A unified, quantitative theory
analysis and its application,” Bioelectrochemistry and Bioenergetics, vol. 45, of reversible electrical breakdown and mechanical rupture in artificial planar
pp. 3–16, 1998. bilayer membranes,” Bioelectrochemistry and Bioenergetics, vol. 25, pp.
[6] L. M. Mir, “Therapeutic perspectives of in vivo cell electropermeabilization,” 163–182, 1991.
Bioelectrochemistry, vol. 53, pp. 1–10, 2000. [24] R. P. Joshi, Q. Hu, K. H. Schoenbach, and H. P. Hjalmarson, “Improved
[7] T. Budak-Alpdogan, D. Banerjee, and J. R. Bertino, “Hematopoietic stem energy model for membrane electroporation in biological cells subjected to
cell gene therapy with drug resistance genes: An update,” Cancer Gene electrical pulses,” Physical Review E, vol. 65, no. 4, pp. 041 920–041 928,
Therapy, vol. 12, pp. 849–863, 2005. Apr 2002.
[8] J. R. Bertino, “Transfer of drug resistance genes into hematopoietic stem [25] V. F. Pastushenko and Y. A. Chizmadzhev, “Stabilization of conducting pores
cells for marrow protection,” The Oncologist, vol. 13, pp. 1036–1042, 2008. in BLM by electric current,” General Physiology and Biophysics, vol. 1, pp.
[9] S. B. Dev, D. P. Widera, and G. A. Hofmann, “Medical applications of 43–52, 1982.
electroporation,” IEEE Transactions on Plasma Science, vol. 28, pp. 206– [26] M. Hibino, M. Shigemori, H. Itoh, K. Nagayama, and K. Kinosita, Jr.,
223, 2000. “Membrane conductance of an electroporated cell analyzed by submicrosec-
[10] K. A. DeBruin and W. Krassowska, “Modeling electroporation in a single ond imaging of transmembrane potential,” Biophysical Journal, vol. 59, pp.
cell. I. Effects of field strength and rest potential.” Biophysical Journal, 209–220, 1991.
vol. 77, pp. 1213–1224, 1999. [27] S. Ŝatkauskas, M. F. Bureau, M. Puc, A. Mahfoudi, D. Scherman,
[11] T. R. Gowrishankar and J. C. Weaver, “An approach to electrical modeling of D. Miklavĉiĉ, and L. M. Mir, “Mechanisms of in vivo DNA electrotransfer:
single and multiple cells,” Proceedings of the National Academy of Sciences, Respective contributions of cell electropermeabilization and DNA elec-
vol. 100, pp. 3203–3208, 2003. trophoresis,” Molecular Therapy, vol. 5, pp. 133–140, 2002.
[12] W. Krassowska and P. D. Filev, “Modeling electroporation in a single cell,”
Biophysical Journal, vol. 92, pp. 404–417, 2007.
Single Allocation P-Hub Median Problem to
Monitor Land Borders by Using Unmanned
Aircraft
Yasemin Akku, nci Sarçiçek
Eski ehir Osmangazi University,
Department of Industrial Engineering
Abstract - During the last two decades, hub location problems (single allocation) or to more hubs (multiple allocations).
have become popular with successful applications in Capacitated versions of hub problems may involve different
transportation, telecommunication, networks, retailing and kinds of capacity constraints [3].
other logistic systems. In this paper we consider Turkey’s land In general, problems are handled in 4 types: p-Hub
borders security. Because security is the one of the most
important deal nowadays, countries are spending so much to
Median Problem; p-Hub Center Problem; Hub Covering
prevent threats that may come from neighbor countries’ Problem; Hub Location Problem with Fixed Costs. The
borders. Also several geographical restrictions at our land objective of the p-hub median problem is to minimize the
borders cause insufficiency to monitor and to take the requiring total transportation cost (time, distance, etc.) needed to serve
data. We focus on the problem of the single allocation p-hub the given set of flows, given n demand nodes, flow between
median that is addressed to monitor the movement at Turkey’s origin–destination pairs and the number of hubs to locate (p)
land borders by using unmanned aircraft. In addition, aiming [4]. p-hub center problem is a minimax type problem. The p-
assignment of demand points that are defined as cities at land hub center problem is to locate p hubs and to allocate non-hub
borders to hubs mathematical model is built and the results are nodes to hub nodes such that the maximum travel time (or
evaluated. Model is encoded by using GAMS software and
solved by CPLEX solver.
distance) between any origin–destination pair is minimized
[5]. Hub covering problem’s aim is to maximize the covered
area by the hubs obeying the maximum time bound on travel
Key Words- P-Hub Median Problem, Unmanned Aircraft time [4]. In the p-hub location problem the fixed cost of
opening facilities is disregarded. On the other hand, the
I. INTRODUCTION simple plant location problem includes fixed facility costs. In
1992, O’Kelly introduces the fixed facility costs into a hub
Hubs form critical elements in many airline, location problem and thereby making the number of hubs a
transportation, postal and telecommunications networks. They decision variable [6].
are centralized facilities in these networks whose functions In this study, we addressed to monitor the movement at
are to consolidate, switch and sort flows. Typical applications Turkey’s land borders by using unmanned aircraft. Within the
of hub location include airline passenger travel, scope of the problem, we consider single allocation p-hub
telecommunication systems and postal networks [1]. The hub median problem.
location problem is concerned with locating hub facilities and The research on the p-hub median problem with single
allocating demand nodes to hubs in order to route the traffic assignment started with the works of O’Kelly (1987). O’Kelly
between origin–destination pairs. (1987) presented the first mathematical formulation for the
Studies on the hub location problem often assume three single allocation p-hub median problem as a quadratic integer
things [2]: program which minimizes the total network cost. This
1. The hub network is complete with a link between every quadratic integer program is considered as the basic model for
hub pair; hub location problem. Campbell (1994) produced the linear
2. There are economies of scale incorporated by a discount integer programming formulation for the single allocation p-
factor for using the inter-hub connections; hub median problem. His formulation has (n4+n2+n) variables
3. No direct service (between two non-hub nodes) is allowed of which (n2+n) are binary and it has (n4+2n2+n+1) linear
There are several kinds of hub location problems based on constraints [2].
characteristics of a particular hub network. If the number of Ernst and Krishnamoorthy (1996) proposed a different
hub nodes is fixed to p, we are dealing with p-hub location linear integer programming formulation which requires fewer
problems. Each non-hub node can be allocated either to one variables and constraints in an attempt to solve larger
237
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_41, © Springer Science+Business Media B.V. 2010
238 AKKU AND SARIÇIÇEK
When compared to manned RECCE aircraft they are cost Table 2. Potential hubs
effective and versatile systems. While reconnaissance,
intelligence, surveillance, and target acquisition (RISTA) are NUMBER OF HUBS POTENTIAL HUBS
17 KARS
18 KAYSER
19 KONYA
20 MALATYA
21 MARDN
22 KAHRAMANMARA
23 MULA
24 MU
25 NEVEHR
26 SAMSUN
27 SRT
28 SNOP
29 SVAS
Fig. 2. Airports
30 TEKRDA
Table 1. Demand nodes 31 TRABZON
32 ANLIURFA
NUMBER OF
DEMAND NODES 33 UAK
DEMAND NODES
34 VAN
1 HATAY
2 KLS
7 HAKKAR – IRAN
B. Assumptions
8 HAKKAR - IRAQ
• The aircraft altitude is constant.
9 VAN
10 ARI
• There is always a sight available.
11 IDIR • The model of all the aircrafts is same.
12 KARS • The speed of the aircrafts is constant (220km/h).
13 ARDAHAN • The speed of the aircrafts during the observation
14 ARTVN operations is 110 km/h.
15 EDRNE • Observation diameter is 15 km.
16 KIRIKKALE
• The communication range of the aircrafts is 300 km.
240 AKKU AND SARIÇIÇEK
• Aircrafts turn back to their hubs. Equation (1) defines the aim function. It decides about
• Duties are performed in good weather conditions. which hub will be activated for maximum benefit using
weighting parameter. Equation (2) assigns a demand node for
• There is no threat for the aircrafts.
each hub using (3). The maximum number of the active hubs
• Considered hubs are controled by General Directorate is limited by p with (4). Equation (5) does the hub selection
of State Airports Authority of Turkey. and assignment using the maximum range of the aircraft.
C. Parameters F. Results
H ; {1,..., m} is the sequence of the main possible hubs The GAMS/CPLEX solver is used to select the hubs to be
N ; {1,..., n} is the sequence of the demand nodes. activated among the possible hubs.
Model is solved for each p value individually
i ; demand node i = 1,2,...,16 while p = 1,2,...,16 . Current airports are defined as possible
j ; possible hub j = 1,2,...,34 hubs. Therefore, the cost of adding a new hub is not
considered. According to the results of our model, Tekirdag,
β j ; the weight of the j th possible hub Trabzon, Gaziantep, and Van airports are decided to be used
d ij ; distance between i th demand node and j th hub for UAV system (Table 3).
¦Y
j
j =p (4)
X ij ≤ a ij ∀i ∈ N ∀j ∈ H (5)
X ij ∈ {0,1} ∀i ∈ N ∀j ∈ H (6)
SINGLE ALLOCATION P-HUB MEDIAN PROBLEM 241
REFERENCES [7] J. Sohn and S. Park, “A linear program for the two-hub
location problem”, European Journal of Operational
[1] A.T. Ernst, H. Hamacher, H. Jiang, M. Krishnamoorthy, Research, 100, pp.617-622, 1997.
and G. Woeginger, “Uncapacitated single and multiple [8] J. Ebery, “Solving large single allocation p-hub problems
allocation p-hub center problems”, Computers and with two or three hubs”, European Journal of Operational
Operations Research, 36, pp. 2230-2241, 2009. Research”, 128, pp.447-458, 2001.
[2] S. Alamur and B.Y. Kara, “Network hub location [9] H. Yaman, B.Y. Kara and T. Barbaros, “The latest arrival
problems”, European Journal of Operational Research, 190, hub location problem for cargo delivery systems with
pp.1-21, 2007. stopovers”, Transportation Research, Part B(43), pp. 643–
[3] J. Kratica, Z. Stanimirovic, D. Tosic, and V. Filipovic, 658, 2009.
“Two genetic algorithms for solving the uncapacitated single
[10] A. Özger and H. Oktal, “Integer model approach for
allocation p-hub median problem”, European Journal of
uncapacitated multiple allocation p-hub median problem in
Operational Research, 182, pp.15-28, 2007.
air cargo transportation”, Havac l k Ve Uzay Teknolojileri
[4] J.F. Campbell, “Integer programming formulations of
Dergisi, 4, pp.47-60, 2009.
discrete hub location problem”, European Journal of
[11] www.stratejikboyut.com.article_detail.php (Cafer
Operational Research, 72, pp. 387-405, 1994.
Serinkan)
[5] J.F. Campbell, “Hub location for time definite
[12] http://www.fas.org/irp/program/collect/uav.htm
transportation”, Computers & Operations Research, vol.36,
[13] http://www.globalsecurity.org/intell/ systems/ uav-
pp.3107-3116, 2009.
intro.htm
[6] S. Elmasta, “Hub location problem for air ground
transportation systems with time restrictions”, PhD Thesis,
Bilkent University, 2006.
LPT to RS232 communication converter
Grofu Florin “Constantin Brancusi” University – Tg-Jiu
Cercel Constantin “Constantin Brancusi” University – Tg-Jiu
Table 1
Octet 1 Address byte address byte
Octet 2 Nr. bytes data bytes
Octet 3 Actions data bytes
Octet 4 Data data bytes
............ .............. ..............
243
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_42, © Springer Science+Business Media B.V. 2010
244 FLORIN AND CONSTANTIN
LabWindows/CVI is a programming environment that has An alternative solution to serial communication is the
been widely adopted throughout industry, academia, and parallel communication that provides high speed
research labs as the standard for data acquisition and communication but on short distance. If there is already exist
instrument control software. It is a powerful and flexible serial communication equipment then must be implemented a
instrumentation software system that embraces the traditional parallel to serial communication converter. This converter
programming methodologies and enables you to create simply takes parallel data package and then sends it serial to
structured code, and it features an easy-to-use programming acquisition system. It also takes serial data package from the
environment.[8] acquisition system and forwards the data parallel to the
LabWindows/CVI includes all the tools necessary for data application.[9]
analysis and presentation of the results on the Graphical User Thus, all changes of state of serial communication port is
Interface (GUI) on your computer screen. It comes with a transferred to the parallel to serial conversion module that
complete set of integrated input/output (I/O) libraries, carried out in real time. The structure of such a
analysis routines, and user interface tools that aid you in rapid communication system is shown in figure 2.
application development, reducing the time that would be The converter is made bidirectional, meaning it can
required by other conventional programming transmit and also receive data, so the information will flow
environments.[6] not only from computer to peripheral, but also from the
LabWindows/CVIs forte lies in building virtual computer peripheral.
instrumentation systems with General Purpose Interface Bus
(GPIB), VME (Versa-Modular Eurocard) extensions for
Instrumentation (VXI), PCI (Peripheral Component
Interconnect) extensions for instrumentation (PXI), serial
interface communication (RS-232), Transmission Control
Protocol/Internet Protocol (TCP/IP) based devices with plug-
in data acquisition (DAQ) boards without spending too much
effort to create the applications. It combines an interactive,
easy-to-use development approach with the programming
power and flexibility of compiled ANSI C code.
The full potential of LabWindows/CVI is used to automate
test systems, bench top experiments, DAQ monitoring
projects, verification tests and measurements, process
monitoring, and controlling systems.
Parallel Interface
III. LPT-RS232 CONVERTER
Microcontroller system for
In case we use the RS232 communication between PC and LPT/RS232 converter
data acquisition system up to 10 times per second the RS232 Interface
application is easy to use.
If we need to increase the speed of communication the
problems begin to appear and only a very good LabWindows
programmer can resolve them. This problem appears as effect
of serial port accessing mode through the existing functions.
The communication is delayed because the ninth bit (parity
bit) must be different set for a command byte or a data byte. DAS
From experiments we observed that delays occur when
accessing the change configuration functions of the serial port
or on change direction of transfer. Thus if the complete Fig. 2. Using LPT/RS232 converter
communication means transmission and reception of a Knowing the communication protocol for serial transfer,
package of 6 bytes of data at a transfer rate of 9600 bits / sec, how to transfer data using PC parallel port and the structure
theoretically should last up to 15 ms, practically takes about and functioning of a 89S52 Intel microcontroller we designed
100 to 150 ms depending on the architecture and performance and achieve a parallel to RS232 communication converter.
of computer system used. Functional diagram of the converter is shown in figure 3:
This leads to a decrease of almost 10 times of the effective
speed of communication resulting the possibility of a
bidirectional data transfer up to 10 times per second.
LPT TO RS232 COMMUNICATION CONVERTER 245
R/W
P1
1 Data Strobe VCC_+5V VCC_+5V
14 3 1
2 D0 S1
15 RESET 4 2 R12
3 D1 360 R12 U7
Sens 16 R1 6K8 360
4 D2 5 4
+
VCC_+5V SDA Vss
17 C1 10uF
5 D3 6 3
18 SCL A2
6 D4 7 2
19 WP A1
20
9
7 D5 U1 8 1
20 D0 1 39 Vcc A0
GND
RESET
8 D6 D1 2 P1.0/T2 P0.0 38 VCC_+5V 24LC256
21 D2 3 P1.1/T2X P0.1 37
9 D7 D3 4 P1.2 P0.2 36
22 D4 5 P1.3 P0.3 35
10 D5 6 P1.4 P0.4 34
23 D6 7 P1.5 P0.5 33 VCC_+5V
11 ACK D7 8 P1.6 P0.6 32
24 P1.7 P0.7
12 21 Led_1 Led_1 R2 360 D1 LED
25 R/W 12 80C52 P2.0 22 Led_2
13 Busy Data Strobe 13 INT0 P2.1 23 Led_3
DB25 ACK 14 INT1 P2.2 24 Led_4 Led_2 R3 360 D2 LED
Sens 15 T0 P2.3 25 Led_5
Busy 16 T1 P2.4 26 Led_6
WR P2.5 27 Led_7 Led_3 R4 360 D3 LED
U2 RS485 17 P2.6 28 Led_8
2 RS485 RD P2.7
6 RE 3 30 Led_4 R5 360 D4 LED
ARK350/2
2 A DE RX_date 10 ALE/P 29
7 1
<----
RX_date_2 RXD PSEN
1 B R 4 TX_date TX_date 11 Led_5 R6 360 D5 LED
D TXD
EA/VP
---->
VCC
X1
X2
SN75176 Led_6 R7 360 D6 LED
RX_date
19
18
31
40
2 RX_date_2 Led_7 R8 360 D7 LED
3 74HCT08
P2 RX_date_1 Y1 VCC_+5V
1
5 Led_8 R9 360 D8 LED
24MHz Fund
4
COMUNICATIE C2 C3
7
U5 RS232 33pF 33pF
2 8 9 RX_date_1
<---- <----
R2in R2out
3 ----> 7 10 TX_date ---->
T2out T2in
13 12
DB9_T R1in R1out
14 11
T1out T1in
5 15
C4 C2- GND C5
100n 4 6 100n
C2+ Vs-
3 2
C9 C1- Vs+ C7
100n 1 16 100n
C1+ VCC VCC_+5V
Fig. 3. Structure of LPT-RS232 converter
MAX232
communication is increased with at least ten times. We The LPT to RS232 converter is very useful in situations
consider this time the time necessary to send and receive a that the data packages sent and received is not so big (under
data package. 128 bytes), case occurred in monitoring and control
applications for industrial process.
REFERENCES
[1]. Cercel Constantin, Grofu Florin, USB to Parallel communication [6]. www.ni.com/pdf/manuals/373551c.pdf
interface, Annals of „Constantin Brancusi” University of Tirgu Jiu, [7]. Vektrex Application Note: 11082002, Using IVI-COM Drivers From
The XIII-th edition of the international scientific conference of the LabWindows/CVI
Engineering Faculty, November 07-08, 2008 , ISSN 1844-4856 [8]. Advanced Topics in LabWindows/CVI by Khalid, Shahid F
[2]. Porat B., A Course in Digital Signal Processings, John Wiley & Sons, [9]. Grofu Florin, Popescu Luminita Georgeta, Cercel Constantin
Inc., New York, 1997. Acquisition system for monitoring vibrations, CISSE2007 – Novel
[3]. Thomas B. Cho, Circuit Techniques for Low Power, High Speed Algorithms and Techniques in Telecommunications, Automation and
Pipelined A/D chapter 2, University of California, Berkeley 2000 Industrial Electronics, Springer, 2008, ISBN: 978-1-4020-8736-3
[4]. Dallas Semiconductor, Ultra-High-Speed Flash Microcontroller [10]. Runceanu Adrian, Fragmentation in distributed databases,
User’s Guide , http://www.maxim-ic.com/ 2004 CISSE2007 - Innovations and Advanced Techniques in Systems,
[5]. Maxim-IC, App Note 604: Fast Memory Transfers with the Ultra Computing Sciences and Software Engineering, Springer, 2008,
High-Speed Microcontroller - DS89C420, ISBN: 978-1-4020-8734-9
Simulation Model of Coremaking Plant
Stefan Beleva, Miroslav Penchevb,
a,b
University of Rousse, 8 Studentska str., 7017 Rousse, Bulgaria
Abstract
The paper presents simulation model of coremaking plant which gives a possibility for optimal utilization of working machines. The
aim of modeling is investigation of coremaking equipment idle time in dependence of a section production plan within a working day.
The model is realized like simulation program which is built in the system for operative management (SCADA) at one of coremaking
plants in Vidima Ltd. Sevlievo, Bulgaria.
Keywords: coremaking plant, simulation model
247
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_43, © Springer Science+Business Media B.V. 2010
248 BELEV AND PENCHEV
core mixture (the mass of the respective product m J and the (FIFO) discipline of servicing. The equipment parameters are
determined and can be presented by the set Mix:
number of products, made within one technological cycle n J ).
The index “J” designates the type of the produced item: Mix = {d, M, , TtrI , Tul} , (6)
h = {mh 1 , mh 2 , , mh I }; I = 1 ÷ 10 (2)
Count of requests -
Car Backward (t)
Active requests:
Car Forward (t)
Active (t)
Car Ready (t)
MPE = Z Mh (3)
NR(t)
Each of the core-blowing machines in the plant is equipped
with a hopper for sand core mixture with adjustable minimum
Core sand mix installation
level rI , (in % of the hopper’s content). When the mixture level Flow of requests
M,
falls below the preset minimum, the machine transmits request TUL, TTR TMX 10
Supply Cain:
where: Request departure: Departure (I, t)
I is the location of the core-blowing machine which has Machine load: Load (I, dI*M, t)
transmitted a request;
Requests chain:
d - the number of the requested (by the respective machine)
Load (dI*M,t)
Load (dI*M,t)
Load (dI*M,t)
RI = R (t, d)
RI = R (t, d)
10
requests may be considered as a queuing system with standard Fig. 3. Conceptual model of coremaking plant.
1
The mixture consumption model in the hopper is reviewed in following In our case the subject of investigation is a casual flow of
section. events: moments of turning off & on and the duration of intervals
SIMULATION MODEL OF COREMAKING PLANT 249
TIT during which the separate machines do not work because of where: t is the current time of the simulation experiment, and kJ –
lack of sand core mixture (Fig. 3 and Fig. 4). the current number of the machine cycle from the moment of
The aim of the investigation (in order to enhance the loading the machine.
effectivity of loading of the core-blowing machines) may be The individual technological cycle may be described by
defined as: minimizing the idle time caused by lack of servicing of equation (11), Fig. 5a:
the machines in the plant (7), by changing the location of the (11)
TC = TRW + TFIL + TFIX + TREM + TSERV = hC + TSERV
product of the current daily job in the technological scheme of the
plant. where: TRW is the time for moving the operating organs from
10
(7) initial to operating position; FIL – the time for filling and
TIT = ¦ TIT = min .
I congesting the mould; TFIX – the time for curing the core into the
I =1
mould; TREM – the time for opening the mould, removing the
2.2. Core Blowing Machine model core and returning the operating organs to initial position; TSERV
– time for cleaning and lubricating the corebox.
A model of consumption of the mixture may represent the TFIX + TREM
moment of sending the request and an idle time interval that is
caused by lack of sand mixture, Fig. 4. TFIL
TSERV
The following denotations are used in the Figure:
L – current hopper level, %;
r – request level, %: TRW
0 L >r (8)
Re quest =
1 0 ≤L ≤r TC
C – core blowing machine cycle, sec;
TAT – arrival time interval, sec; ) Cycle Time b) Service Time Model
WT – waiting time interval, sec; Fig. 5. Core blowing machine – technological cycle.
IT, – idle time interval, machine does not work, sec.
The time intervals TRW, FIL, TFIX and TREM are functions
of the machine settings for the particular product J. And the
Machine Run Machine Run Machine stopped. operation of cleaning and lubricating the corebox is performed
No Request Request Request manually by the operator and is of probability nature.
Observation of the work of machine 6 with product: faucet “bath
L, % L ≥ r 0≤L≤r L=0
body SL – A” is shown in Fig. 5b. The realized observations and
100% statistic investigations show that the time for service may be
described by an exponential distribution which parameter is
defined by the average time.
TC
r Request
3. Model implementation
Level
0% t
TAT TWT TIT On a system level the simulation model assumes description
of the functioning of the simulated system using the concepts
activities, events and processes by means of a definite
Work cycle of Core Blowing Machine programming language. The methods known in practice for
organizing the computing process in the simulation programs
Fig. 4. Core blowing machine – hopper level.
(activity scanning, event schedule and processes interaction) are
Taking into consideration the discrete character of the striving to formalize and solve the problem with the concurrent
machine cycle and the casual character of the interval TWT for the processes, common to all computer simulators. In this particular
mixture level in the hopper of machine I, producing item J at the case due to the elementary organization of the simulation
moment of supplying, it could be written down: machine and the relatively easy realization by means of a
universal programming language, for realization of the simulator
LI, J (t, k JTC J ) = LI, J (t, k JTC J ) + 100 * dI ,% , (9) the “activity scanning” method is chosen.
kJ = 0
3.1. Model formalization
and for the level after each cycle:
mJnJ
LI, J (t, k JTC J ) = LI, J (t, (k J − 1)TC J − 100 ,% Definition 1: The dynamic model of a simple object may be
M , (10) represented as an arranged five, [1]:
LI, J (t, k JTC J ) ≥ 0
DM = (S; EINP; EOUT; SC; g), (12)
kJ = kJ + 1
where S is the endmost set of states of the object;
EINP – the set of the input events of the object;
EOUT – the set of the output events of the object;
250 BELEV AND PENCHEV
g : S × E INP → S - transition function. preset and/or calculated time of life in state SI. (Computation
SC – the set of conditions, which may be defined as a of the times is performed beforehand).
subset of the events, admissible for the separate states Q(1)
MhLoad
SC : ∀s ∈ S; EINP
S
⊂ EINP .
c : (Q(1) = I) ∧ (MhLoad = 1)
This model of the states and transitions is interpreted in the a : Re set MhLoad
following manner. If the object is in state SI and event ∈ INP Machine
Load S0I
occurs, which is admissible for that state, then the object transits
in a new state defined by the function g : S × E INP → S . c : TRUE
L I = L I + 100d Q(1) ,%
The visual formalism of that interpretation is “State
Re set R(I, d) Queue Re duce
Transition Diagram”, well known from the theory for the finite a:
Set TS1I = t + Tc I
state machine and widely used in automation. Store Depart(I, t )
In this paper the diagram specification defined by IEC 61131
[2] is used. It is the base of many programming languages for Machine
controlling devices. The diagram is a directed graph whose Cycle S1I
points (denoted by a circle) represent the object’s states and the
c : (t ≥ TS1I )
directed line denotes the transition to the next state (Fig. 6). In a m In I
a : L I = L I − 100
rectangle to the right of the state the actions performed by the M
,%
State
object in that state are denoted. In a random moment of time the
Evaluate S2I
object may be only in one state – active state.
SK -1
c : L I ≥ rI ,% c : L I ≤ 0,%
a : Set TS 3I c : 0 < L I ≤ rI ,% a : Store Empty(I, t )
TS 3I = t + RND(λ I ) Set TS 3I = t + RND(λ I )
Logical Segment
a : Set R(I, d)
Store Arive(I, t )
State
SK : - State Actions Corebox Stop
service S3I
Abstract—Robotics has been developed in a high overpass obstacles and others can be made. One of the
level in the past few years, especially in mobile ways to perform distance measurement is through digital
Robotics. Its main contribution is shown in human images. Image stereo systems are made with two images
entertainment and performance in dangerous works.
However, those are not just the only applications they taken simultaneous, or not, of the same scene, similar to
perform. There is a wide field to be developed. Distance animals that have two eyes. With this system you can
calculation is something really important for localiza- recover the depth information in the images acquired [1].
tion and navigation of mobile robots. The classic way This work aims to develop a system of distance mea-
to calculate distances are using sonar and laser radar.
surement using computer vision to use in modules of
But, the use of computer vision in stereo, based on
two cameras, is widespread in Robotics. Therefore, perception in mobile robots.
algorithms, to detect new points, allows the creation The development of a computer vision system is of
of new methods to execute stereo vision. In those great importance to complete tasks by robots, with special
algorithms there is one that has been used, the SIFT regard to navigation of autonomous robots in unfamiliar
(scale-invariant feature transform). It may be used
with classic algorithms of stereo vision in the Com-
environments. This statement is true, since vision is the
puter Vision 3D system development. This project’s primary sense used by humans and that guides their
objective is to develop a distance acquisition system daily lives, in their work and their leisure time. There
using computer vision in stereo with the help of SIFT is a tendency of current mobile robots incorporate vision
method. The results are very close with what we were systems. As the use of stereo systems allows the mea-
expecting. The next steps to be done are the execution
of quantitative tests to determinate the focal distance
surement of depth, it is going to be useful in the task of
of the utilized cameras and to calibrate the system. locating and overpass obstacles. It should be mentioned
that several other algorithms can worked together with
I. Introduction the vision system in stereo, using the same cameras and
A typical mobile robot is constituted by subsystems images acquisition system [4]. This work is organized as
such as: perception, localization, navigation and movement bibliographic review from Transformed SIFT and Depth
control [1]. The perception system is responsible to acquire calculation using Digital Image Processing in the second
data from the space where the robot is placed. This data, part, there is the methodology used to achieve the results
after being stored, is processed to be used to generate in the third part and in the forth there are the results of
information on the location where the robot is placed, the research.
through its location system. In possession of its location,
the robot is able to generate the environment map where II. Bibliographic Review
it is situated. After it may compare its local position with
a more comprehensive map generated or delivered by its The main methods to execute the implementation and
developer or user [2]. The ability to navigate using the to understand this system are the transformed SIFT and
map and the mission that the robot should perform are the use of stereo images to calculate the coordinate z
the main tasks that should be performed by its navigation distance.
system. This last task informs which path the system of
A. SIFT Method
motion control should perform.
One of the most important information for a mobile SIFT is a robust method to extract and describe key-
robot is the distance measurement to barriers. This infor- points. The features extracted by this technique are in-
mation can be obtained through various sensors, among variant to translation, rotation, and scale and partially
them images from the environment might be obtained. invariant to changes in lighting and perspective [2] [3] [5].
In possession of this distance, decisions such as location, Basically, the algorithm is performed in four steps:
253
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_44, © Springer Science+Business Media B.V. 2010
254 NETO ET AL.
Abstract: This paper presents the results from considered a hybrid between a bipolar transistor and
a study on the dynamic new technologies that a field effect transistor (MOSFET) obtaining the best
present under various conditions when the features of both. The IGBT copied the best features
current varies in hard switching, these data were of MOSFET: simple control circuit, safe working
obtained in response to use as a case study of the area and high peak currents; and the bipolar
isolation gate bipolar transistor, which is a power transistor [1].
semiconductor device.
II. MANUFACTURING TECHNOLOGIES
I. INTRODUCTION
Currently, in the market there are three IGBTs
The proper selection of a power electronic device, from manufacturing technologies: Punch-Through,
one must consider aspects such as cost, the parasitic Non-Punch-Through and Field-Stop [9] - [11].
elements and the power dissipated. This latter aspect
directly influences the efficiency of the converter, A. Technology “Punch-Through (PT)”
especially topologies that used hard switching on the This technology IGBT has a non-homogeneous
switches. Therefore, correct evaluation of the losses structure. In this device fabrication technology is
generated in the type of converter and its operating built on a (p+) type substrate thickness (≈300m).
conditions are very important to optimize the circuit The (n-) layer is obtained by epitaxial growth and is
design. The information collected for the assessment relatively thin. “Fig. 1” shows a very thin (n+) layer
of the manufacturer's data sheet, is often very limited between the base and emitter of the internal pnp
or incomplete and represents approximate values, named buffer, which limits the expansion of the
taking into account the tolerances of the devices space charge region in a state of Punch-Through
commonly present. Another drawback stems from blocking effect and reduces the emitter injection in
the fact that these data do not lend themselves to conducting.
comparison between devices from different Gate Gate Gate
manufacturers, since they were obtained under Emitter
Metal
Emitter
Metal
Emitter
Metal
257
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_45, © Springer Science+Business Media B.V. 2010
258 ESQUIVEL ET AL.
IV. SWITCHING MODES Depending on the time in which it occurs, there are
two types of short circuits:
The following describes the different switching Short circuit type I: the short circuit occurs during
modes that can be put a switch-controlled power the locking phase, before the device turn on "Fig. 3".
converter in an application [1]. Short circuit type II: the short circuit occurs
during the driving phase, before the device off in the
"Fig. 3".
V. TEST CIRCUIT
The test circuit used for the characterization in
hard switching converter is a “DC/DC” type gearbox
or also called chopper. The circuit diagram is shown
in "Fig. 4", including the device under test “DUT”
and the auxiliaries.
LC Rind
AUX1
Fig 3: Different switching converter. Dfw
Lcarga
1) Hard Switching: The switching of a circuit
breaker is defined as hard when there is VO
simultaneous presence of current and voltage in the
time period that lasts the switching phase. It can
DUT
occur in the ignition and off the device and is AUX2
D1
characterized by high losses "Fig. 3". IMP1 IMP2
2) Soft Switching: The switching of a circuit
breaker is defined as mild when the voltage or
current is zero at the start of phase switching. It can
occur in the ignition and shut off and is Fig 4: Test circuit for hard switching.
characterized by low losses "Fig. 3". The soft
The “AUX1” auxiliary device works on one side a
switching is classified into:
free movement diode in the normal operation tests
Zero voltage switching or ZVS: This type of
and on the other side as for short-circuit tests. The
switching occurs only during power when the
“AUX2” auxiliary device is placed parallel with the
voltage is zero before it starts to flow a current
device being tested and it circulates the initial
through the device.
current of the inductive load to the level of current
Zero current switching or ZCS: This type of
necessary for testing. Thus, the “DUT” leads the
switching occurs only during shutdown, when the
minimum time required and avoids the effect of
current is zero before the device starts to block a
internal heat losses by conduction (conduction
voltage. [3], [4].
current by saturation voltage), obtaining a better
3). Short circuit: The short is a type of switching
control junction temperature of the device. “LC” is
is not desirable in a converter, which in principle
the parasitic inductance wiring. “Rind” is inductance
corresponds to a hard switching on the ignition
resistance load. “Lcarga” is load inductance where the
under certain conditions the external circuit, being
energy is stored for the current test. “IMP1” and
the most frequent short-circuiting the load. This
“IMP2” are drivers that perform the action to control
switch is characterized by very high losses that lead
the activation of auxiliary devices under test.
to the device to the limits of its safe operation.
260 ESQUIVEL ET AL.
The nominal test circuit: collector current caused by the same reverse recovery diode containg
IC=25A, gate resistance RG=22Ω, voltage VO=600V, in the testbank causing the same effect. This displays
wiring inductance LC=116nH, gate voltage an increase in power and the energy dissipated as it
VGG=15V and junction temperature Ti=30°C. switches to the on position. Infineon Technologies in
hard switching consumed to the maximum power in
VI. EXPERIMENTAL RESULTS testing (6mJ) as well as International Rectifier.
While Fairchild only consumed (5mJ).
In this section we present the results in three 800 80
Vce [V]
400 40
Ic [A]
are shown only in the ignition and shutdown, which 200 20
Pon [KW]
Eon [mJ]
20
800 80 -10 0
0 0.5 1 1.5 2 0 0.5 1 1.5 2
(3) t:0.5us/div (4) t:0.5us/div
600 60
Fig. 6: Behavior of the voltage VCE, current IC, power dissipated energy PON
Vce [V]
400 40
Ic [A]
-200 -20
Lastly "Fig. 7" shows the switching on the
0 0.5 1
(1) t:0.5us/div
1.5 2 0 0.5 1
(2) t:0.5us/div
1.5 2
International Rectifier were, where the collector-
40 8 emitter voltage VCE generating the same step
30
6
decreases due to parasitic inductance. This displays
an increased potency similar to that of Infineon and
Pon [KW]
Eon [mJ]
20
4
10
dissipated energy at the time of power up the device
from company International Rectifier. Note that the
2
0
-10
0 0.5 1 1.5 2
0
0 0.5 1 1.5 2 time scales are similar and can see very well the best
(3) t:0.5us/div (4) t:0.5us/div
benefits in switching the ignition.
Fig. 5: Behavior of the voltage VCE, current IC, power dissipated energy PON 800 80
and EON in the ignition hard switching, “IKW25T120”.
600 60
400 40
Ic [A]
-200 -20
Eon [mJ]
20
4
Technologies.
-10 0
0 0.5 1 1.5 2 0 0.5 1 1.5 2
(3) t:0.5us/div (4) t:0.5us/div
"Fig. 6" also demonstrates the Fairchild Company
Fig. 7: Behavior of the voltage VCE, current IC, power dissipated energy PON
power switching device, where the collector-emitter and EON in the ignition hard switching, “IGTG11N120CN”.
voltage VCE generating the same step decreases due
to the parasitic capacitance that is similar to above,
with a much greater slope. The current spike is
DYNAMIC OF NEW TECHNOLOGIES IN SWITCHING POWER DEVICES IGBTS 261
The case of commutation when you want to turn a good behavior by having a relatively small tail off
off a device is different, see “Fig. 8, Fig. 9 and Fig. and has an energy of (2.3mJ) which is about half the
10”. energy the company of Infineon Technologies.
"Fig. 8" again this switching device Infineon, is 800 60
Vce [V]
400
Ic [A]
20
-200 -20
800 60
25 2.5
600 20
40 2
15
Vce [V]
Poff [kW]
Eoff [mJ]
400
Ic [A]
1.5
20
10
200
1
5
0
0 0.5
0
-200 -20 -5 0
0 0.5 1 1.5 2 0 0.5 1 1.5 2 0 0.5 1 1.5 2 0 0.5 1 1.5 2
(1) t:0.5us/div (2) t:0.5us/div (3) t:0.5us/div (4) t:0.5us/div
30 5
Fig. 10: Behavior of the voltage VCE, current IC, power dissipated energy PON
20
4
and EON in the ignition hard switching shutdown, “IGTG11N120CN”.
Eoff [mJ]
Poff [kW]
3
10
0
2
VII. CONCLUSION
1
-10
0 0.5 1 1.5 2
0
0 0.5 1 1.5 2 The experimental results on tests (hard-
(3) t:0.5us/div (4) t:0.5us/div
switching), the three technologies analyzed show
Fig. 8: Behavior of the voltage VCE, current IC, power dissipated energy PON that the ignition energies during on and off
and EON in the ignition hard switching shutdown, “IKW25T120”.
switching is different as the parameters vary. There
In "Fig. 9" shows the switching behavior in are differences in the level of impact it can have a
shutdown where the device is the most deficient, parameter variation in each of the devices, those
approximately consumes the energy (23mJ) of belonging to different technologies. Considering the
Fairchild Company, showing that this device can’t results, is necessary extend the experimental
be used at high frequencies have a very slow characterization from hard switching, changes in
shutdown. current, voltage, temperatures of union, gate
resistance and parasitic inductance. The optimization
800 60 of devices, increases efficiency and therefore energy
600
40
saving.
Vce [V]
400
Ic [A]
20
200
VIII. FUTURE WORK
0
0
-200
0 1 2 3 4
-20
0 1 2 3 4 Obtain the complete mathematical model of hard
(1) t:1us/div (2) t:1us/div
switching behavior and maximize the efficiency of
40 25
the tested power devices.
30 20
Poff [kW]
Eoff [mJ]
20 15
10 10
REFERENCES
[1] J.Aguayo., “Análisis Comparativo de Transistores IGBT
0 5
-10
0 1 2 3 4
0
0 1 2 3 4 tipo PT y NPT en Diferentes Modos de Conmutación”,
(3) t:1us/div (4) t:1us/div
Tesis de Maestría, CENIDET, Cuernavaca, México, 2000.
Fig. 9: Behavior of the voltage VCE, current IC, power dissipated energy PON [2] H. Ruething, F. Umbach, O. Hellmund, P. Kanschat,
and EON in the ignition hard switching shutdown, “IRG4PH50S”. G.Schmidt, “600V-IGBT3: Trench Field Stop Technology
in 70μm. Ultra Thin Wafer Technology”, Memorias de la
Unlike earlier in the "Fig. 10", is shown device conferencia ISPSD, 2003, p.
off from International Rectifier, this device presents
262 ESQUIVEL ET AL.
[3] T. Laska, G. Miller, M. Pfaffenlehner, P. Türkes, D. Stop IGBTs”, IEEE Trans. Industry Applications, vol.
Berger, B. Gutsmann, P. Kanschat, M. Münzer, “Short IA-39/4, p. 922-928, 2003.
Circuit Properties of Trench-/Field-Stop-IGBTs – Design [7] R. Kraus, M. Redding, K. Hoffmann, "The Short-Circuit
Aspects for a Superior Robustness” Memorias de la Behaviour of IGBTs Based on Different Technologies",
conferencia ISPSD, 2003, p. 1-4 Memorias de la conferencia, EPE, 1995, p. 1_157-1_160.
[4] H. Iwamoto, H. Kondo,S. Mori, J. Donlon, A. Kawakami, [8] S.Azzopardi,C. Jamet, J.M. Vinassa, C. Zardini,
“An Investigation of Turn-off Performance of Planar and "Switching Performances Comparison of 1200V Punch-
Trench-Gate IGBTs under Soft and Hard Switching”, Through and Punch-Through IGBTs under Hard-
Memorias de la conferencia IAS, 2000, p. 1-6. Switching at High Temperature", Memorias de la
[5] R. Mallwitz,R. Tschirbs,M. Pfaffenlehner, A. Mauder, C. conferencia IEEE PESC, 1998, p. 1201-1207.
Schaeffer, “1700V Trench IGBT Modules”, Memorias de [9] www.infineon.com
la conferencia PCIM, 2001, p. 1-6. [10] www.fairchildsemi.com
[6] X.Kang, A. Caiafa, E. Santi, J. Hudgins, R. Palmer, [11] www.irf.com
“Characterization and Modeling of High-Voltage Field-
Anisotropic Percolating Pathways in the Thin Films
of Polymeric PEDT/PSS Complex and their Relation
to the Electrical Conductivity as Revealed by
the Mesoscale Simulation
Toomas Kaevanda, Jaan Kaldab, Andres Öpika, Ülo Lillea1
a
Department of Material Sciences, Tallinn University of Technology, Ehitajate tee 5, 19086 Tallinn, Estonia
b
Institute of Cybernetics, Tallinn University of Technology, Akadeemia tee 21, 12618 Tallinn, Estonia
1
E-mail addres [email protected].
Ü. Lille thanks Estonian Science Foundation for financial support (Grant
No 6633).
263
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_46, © Springer Science+Business Media B.V. 2010
264 KAEVAND ET AL.
the bottleneck site, is completely neglected. In reality, Pay attention to the fact that these correlation analysis
however, this is not correct. Another drawback of the results are relatively insensitive with respect to the change of
bottleneck model is that for small samples (like our 323), the the fitting parameter (dc vs 〈Δd〉). Further, the values of the
statistical fluctuations of the bottleneck value are rather large. exponential factors for different directions are rather close and
Therefore, it can be argued that the average hopping length, as therefore, the expected high anisotropy, i.e. high ratio of the
reflected by the average field density within the conduction conductivity in parallel to the substrate surface plane (║, in-
window 〈Δd〉, can serve in our case as an alternative measure plane) and perpendicular to it (┴, out-of-plane) cannot be
with a lesser statistical variance. To summarize, dc is the lower identified. An exceptional case is the item 4 (S/T = 6), for
bound of the conducting window, and 〈Δd〉 is the median value which this ratio achieves a value of about 25. Thus, these field
of it. density values (dc and 〈Δd〉) are unable to reflect the electrical
Hence, it is expected that the global conductivity is anisotropy in the set of studied systems. In what follows we
proportional to exp(bdc-1/3), or to exp(b〈Δd〉-1/3), respectively, argue that this fact can be explained by the imperfections of
where b is an unknown constant; recall that dc and Δd are the the bottleneck model and that the strong anisotropy of the
bottleneck field density and the average density of the global conductivity can be ascribed to the strong anisotropy of
conducting window of the PEDT-beads, respectively. the topological structure of the percolation clusters at
criticality (which is, indeed, observed in our simulations).
TABLE II .
Maximal (dmax) and critical (dc) field densities in the perpendicular III. C. ANISOTROPY OF THE PERCOLATION PARAMETERS AND
directions 1, 2, 3. THE MACROSCOPIC ELECTRICAL CONDUCTIVITY.
It dmax dc Below are compared the percolation parameters
em 1 2 3 characterizing the subcritical networks generated for the
1 0.9 0.8 0.7 0.7 samples with different S/T ratios (see Tab III).
542 09 960 91 The relatively high p*c values give an evidence that a lot of
2 0.8 0.6 0.6 0.6
452 38 012 01 cells are located in the finite clusters. A rather modest
3 0.6 0.4 0.3 0.3 anisotropy is observed.
845 07 971 97 One can see that there is a considerable anisotropy for P(p),
4 0.5 0.3 0.1 0.0 as evidenced by the fact that the values for the first two
347 97 930 812
directions are noticeably lower than in the case of the third
5 0.4 0.0 0.0 0.0
301 794 778 763 direction. This means that more sites are needed to circumvent
6 0.3 0.0 0.0 0.0 the insulating PSS beads, i.e. in order to percolate in the third
536 694 694 691 direction.
7 0.1 0.0 0.0 0.0 Remarkable is that P(p) values are rather close to the P(p)b
191 495 495 494
8 0.5 0.4 0.4 0.4
values. Hence, the subcritical network has very few dead ends.
73 67 66 43 Note also that according to our previous report [34], P(p)
9 0.5 0.4 0.4 0.4 behaves according to a power law, P(p) ∝ |p−pc|β, where the
55 56 54 33 critical exponent for the infinite cluster density β >1. We also
1 0.5 0.4 0.4 0.4
0 41 47 21 21
showed there that our systems correspond to the correlated
percolation, and the power law of P(p) is observed up to p = 1.
The simulation data (see Tab. II), indeed, confirm the This is important, because in the real systems p − pc. can take
expected correlation, for all the three directions (items 1-10): non-small values.
ln σ = -4.8767 dc1-1/3 +4.3383, R2 = 0.8215 (3), The most anisotropic quantity is P(p)b : over the full range
ln σ = -5.1979 dc2-1/3 + 5.0938, R2 = 0.90 (4) of the mass ratios and simulation conditions one can observe
ln σ = -4.138 dc3-1/3 + 4.0129, R2 = 0.65 (5) the dependence of P(p)b on the direction. The average value of
Note that excluding the special cases 8,9, and 10 improves the the charge-carrying backbone strength in the third direction
correlation for the 3rd direction, resulting in σ ∝ exp(-4.5045 exceeds that of in the first direction ca four-fold.
dc3-1/3) and R2 = 0.89. Being guided by this finding, we also studied the correlation
Pay attention that b < 0, i.e. ln σ decreases when dc-1/3 between P(p)b1 and the macroscopic conductivity; a
increases (which corresponds to decreasing dc). moderately correlated linear relationship was found (items 1-
Similarly good (even slightly better) correlation is observed 10):
when the average conducting window density 〈Δd〉 is used ln σ = -84.79 P(p)b1 - 0.707, R2 = 0.51 (9).
instead of dc (items 1-7): This correlation improves somewhat, if the discordant item
ln σ = -7.107 〈 Δd 〉c1-1/3 +5.5708, R2 = 0.87 (6) 5 is discarded, resulting in
ln σ = -132.4 P(p)b1 +0.3641 R2 = 0.62. (10)
ln σ = -7.358 〈 Δd 〉c2-1/3 +6.071, R2 = 0.90 (7)
Pay attention that b < 0, thus, the conductivity decreases very
ln σ = -7.4394 〈 Δd 〉c3-1/3 +6.2799, R2 = 0.91 (8)
rapidly with the increase of P(p)b1 values (e.g. 3-6 orders of
Qualitatively analogous relationship (however, just for a
magnitude at 4-fold increase of the backbone density).
single direction) was observed in our recent report [39].
266 KAEVAND ET AL.
Although Eqs. (9) and (10) reflect just a modest correlation, respectively. From the point of view of the experimental data,
let us have a look on the consequences of them. Particularly these are quite reasonable results. A special case is the item
remarkable in Tab. III is the item 4 (S/T = 6, as was used in 10: a shear applied in the X direction, resulting in the equality
[29]), where the ratio 〈P(p)b1,2〉/P(p)b3 = 7.5 corresponds to of backbone densities in 2nd and 3rd directions, clearly
〈σ1,2〉/σ3 = 1018, i.e. the out-of-plane conductivity is differing from that of the first one, similarly to the item 9 (in
vanishingly small. Items 2 and 9 also contrast with the the latter, the unassociated PSS chain is ca 3-fold longer as
remaining items: the backbone strength in the directions 1 and compared to the conventional one). Note that shear (the
2 is very similar and differ strongly from that of the third processing tool to introduce an anisotropy) and subtle changes
direction, leading to the 〈σ1,2〉/σ3 values of 102 and 107, in topology increase the anisotropy.
TABLE III
Percolation functions in perpendicular directions (1, 2, 3) at various PSS/PEDT mass ratios (S/T) at the doping level 0.25 (if not shown
otherwise) and temperature 298K.
1) PSS is modeled as an alternating copolymer; 2) as in 1), but unassociated PSS chain is ca 3-fold longer; 3)under shear 0.001 ns-1, repulsive interaction 2.6 RT,; 4) doping level 0.03; 5) average
from 3 parallel experiments.
Item, 10-2 x p*c 10-2 x P(p) 10-2 x P(p)b
(S/T) 1 2 3 1 2 3 1 2 3
1,(1.4) 6.26 8.24 9.19 2.47 3.69 5.17 2.23 3.45 4.87
2, (2.5) 10.2 12.5 14.8 2.25 2.95 13.4 2.15 2.76 12.8
3, (4.2) 12.6 14.0 14.0 2.76 10.5 10.5 2.56 9.95 9.98
4, (6.0) 6.67 23.8 40.4 1.63 8.12 37.9 1.44 10.4 37.75)
5, (8.0) 35.5 37.4 39.5 9.27 10.7 20.7 9.24 10.7 20.3
6, (10) 36.8 35.8 37.7 6.15 6.53 15.9 5.97 6.14 15.4
7, (20) 25.5 25.5 27.0 7.36 7.36 11.6 6.10 6.13 9.95
81),(2.5) 6.97 7.14 16.6 3.69 4.72 12.3 3.05 4.02 11.4
92),(2.5) 6.10 6.76 12.6 0.82 1.71 10.5 0.65 1.44 9.86
103),(2. 7.86 19.0 19.0 1.22 10.2 10.2 1.0 9.42 9.48
5)
114),(2. 12.9 15.2 16.7 5.1 6.0 10.5 5.02 5.92 10.4
5)
Aver. 15.4 19.0 23.1 3.76 6.65 14.8 3.44 6.45 14.2
Stdev. 10.3 9.29 10.5 2.30 2.73 6.01 2.63 3.31 8.74
The previous paragraph 1) demonstrates an approximate densities (Tab. IV). Hence, not all red sites act as connectors
character of the correlations ln σ vs P(p)b [given by Eqs. of finite clusters into the infinite one, i.e. part of them do not
(9),(10)], and 2) clearly reflect the conductivity anisotropy. take part directly in the formation of the subcritical network.
Let us now study the structure of the charge carrying Therefore the role of the red sites in ambiguous.
backbone (some of them are shown in Fig.). The P(p)b values
are ca one order of magnitude higher than those of fr. (see Tab. TABLE IV
III and IV). Hence, the backbone contains a number of Fractions of the singly connected sites (fr) in the perpendicular directons
-3
multiconnected pieces which provide several routes in parallel Item 10 x fr
1 2 3
(loops). This is because the red sites connect the subsets of the 1 5.19 7.51 10.3
finite clusters into the infinite one. Note that the ratio of the 2 3.75 3.84 18.1
masses of the percolation path to the overall mass of all the 3 4.61 15.6 15.4
cluster pieces is ca ¼ for the direction 1 and 2/3 for the 4 2.01 4.70 7.48
direction 3 (compare the average values of P(p) and p*c in 5 1.95 1.37 7.96
6 4.58 6.32 8.88
Tab. III). These structural features may result in reducing the 7 17.8 17.7 24.4
correlation strength ln σ vs P(p)b. 8 5.19 7.51 10.3
Note that the average value of the fractions of the red bonds 9 2.78 4.94 17.9
in directions 1 and 2 is lower in comparison to that of in the 10 4.00 14.8 14.8
third direction, i.e. the behavior is similar to the backbone Average 5.19 8.43 19.4
ANISOTROPIC PERCOLATING PATHWAYS IN THE THIN FILMS 267
REFERENCSES [21] J. Ouyang; Q. Xu; C. Chu; Y. Yang; G. Li; J. Shinar, "On the
mechanism of conductivity enhancement in poly(3,4-
[1] M. Kemerink; S. Timpanaro; M. M. deKok; E. A. Meulemkamp; ethylenedioxythiophene):poly(styrene sulfonate) film through solvent
M. J. Touwslager, "Three-dimensional inhomogeneities in PEDOT: PSS treament," Polymer vol. 45, pp. 8443-8450, 2004.
films," J.Phys.Chem.B vol. 108, pp. 18820-18825, 2004. [22] M. M. De Kok e. al, "Modification of PEDOT:PSS as Hole
[2] A. M. Nardes; M. Kemerink; R. A. J. Janssen, "Anisotropic Injection Layer in Polymer LEDs," in Physics of Organic Semiconductors,
hopping conduction in spin-coated PEDOT : PSS thin films," Physical Review W.Brütting, Ed.; WILEY-VCH: 2005, p 451-473.
B vol. 76, pp. 1-7, 2007. [23] C. Ionescu-Zanetti, Mechler, A., Carter, S.A., Lal,R.,
[3] D. M. DeLongchamp; B. D. Vogt; C. M. Brooks; K. Kano; J. "Semiconductive Polymer Blends: Correlating Structure with Transport
Obrzut; C. A. Richter; O. A. Kirillov; E. K. Lin, "Influence of a water rinse on Properties at the Nanoscale," Adv. Mater. vol. 16, pp. 385-389, 2004.
the structure and properties of poly(3,4-ethylene dioxythiophene):poly(styrene [24] P. Vacca; M. Petrosino; R. Miscioscia; G. Nenna; C. Minarini; D.
solfonate) films," Langmuir vol. 21, pp. 11480-11483, 2005. Della Sala; A. Rubino, "Poly(3,4-ethylenedioxythiophene): poly(4-
[4] B. E. e. a. Kilbride, "Experimental observation of the scaling laws styrenesulfonate) ratio: Structural, physical and hole injection properties in
for alternating current and direct current conductivity in polymer-carbon organic light emitting diodes," Thin Solid Films vol. 516, pp. 4232-4237,
nanotube composite thin flms.," Journal of Applied Physics vol. 92, pp. 4024- 2008.
4030, 2002. [25] A. M. Nardes; R. A. J. Janssen; M. Kemerink, "A morphological
[5] S. S. Rahatekar; M. S. P. Shaffer; J. A. Elliott, "Electrical. model for the solvent-enhanced conductivity of PEDOT : PSS thin films,"
percolation in carbon nanotube dispersions: A mesoscale modeling and Advanced Functional Materials vol. 18, pp. 865-871, 2008.
experimental study," Smart Nanotextiles vol. 920, pp. 63-69 [26] M. Deepa; S. Bhandari; R. Kant, "A comparison of charge
151, 2006. transport behavior in functionalized and non-functionalized poly 3,4-
[6] S. Ghosh; O. Inganas, "Self-assembly of a conducting polymer (ethylenedioxythiophene) films," Electrochimica Acta vol. 54, pp. 1292-1303,
nanostructure by physical crosslinking: applications to conducting blends and 2009.
modified electrodes," Synthetic Metals vol. 101, pp. 413-416, 1999/5. [27] H. Yan; S. Arima; Y. Mori; T. Kagata; H. Sato; H. Okuzaki,
[7] S. Kirchmeyer; K. Reuter, "Scientific importance, properties and "Poly(3,4-ethylenedioxythiophene)/poly(4-styrenesulfonate): Correlation
growing applications of poly(3,4-ethylenedioxythiophene)," Journal of between colloidal particles and thin films," Thin Solid Films vol. 517,
Materials Chemistry vol. 15, pp. 2077-2088, 2005. pp. 3299-3303, 2009.
[8] L. B. Groenendaal; F. Jonas; D. Freitag; H. Pielartzik; J. R. [28] A. Dkhissi; D. Beljonne; R. Lazzaroni, "Atomic scale modeling of
Reynolds, "Poly(3,4-ethylenedioxythiophene) and its derivatives: past, present interfacial structure of PEDOT/PSS," Synthetic Metals vol. 159, pp. 546-549,
and future " Adv.Mater. vol. 12, pp. 481-494, 2000. 2009.
[9] X. Crispin; e. al, "Conductivity, morphology, interfacial chemistry, [29] A. M. Nardes; M. Kemerink; R. A. J. Janssen; J. A. M.
and stability of poly(3,4-ethylene dioxythiophene)-poly(styrene sulfonate): A Bastiaansen; N. M. M. Kiggen; B. M. W. Langeveld; A. J. J. M. van Breemen;
photoelectron spectroscopy study," J.Polym.Sci.B vol. 41, pp. 2561-2583, M. M. de Kok, "Microscopic understanding of the anisotropic conductivity of
2003. PEDOT : PSS thin films," Advanced Materials vol. 19, pp. 1196-1200, 2007.
[10] A. B. Kaiser, "Systematic conductivity behavior in conducting [30] C. S. S. Sangeeth; M. Jaiswal; R. Menon, "Correlation of
polymers: effects of heterogeneous disorder," Adv.Mater. vol. 13, pp. 927- morphology and charge transport in poly(3,4-ethylenedioxythiophene)-
941, 2001. polystyrenesulfonic acid (PEDOT-PSS) films," Journal of Physics-Condensed
[11] F. Louwet; L. Groenendaal; J. Dhaen; J. Manca; J. Van Luppen; E. Matter vol. 21, 2009.
Verdonck; L. Leenders, "PEDOT/PSS: synthesis, characterization, properties [31] J. G. E. M. Fraaije; B. A. C. van Vlimmeren; N. M. Maurits; M.
and applications," Synthetic Metals vol. 135-136, pp. 115-117, 2003. Postma; O. A. Evers ; C. Hoffmann ; P. Altevogt; G. Goldbeck-Wood, "The
[12] K. Z. Xing; M. Fahlman; X. W. Chen; O. Inganas; W. R. Salaneck, dynamic mean-field density functional method and its application to the
"The electronic structure of poly(3,4-ethylene-dioxythiophene): studied by mesoscopic dynamics of quenched block copolymer melts," Journal of
XPS and UPS," Synthetic Metals vol. 89, pp. 161-165, 1997. Chemical Physics vol 106, pp. 4260 - 4269, 1997.
[13] K. E. Aasmundtveit; E. J. Samuelsen; L. A. A. Pettersson; O. [32] G. H. Fredrickson In International series of monographs on
Inganas; T. Johansson; R. Feidenhans'l, "Structure of thin films of poly(3,4- physics; Clarendon Press ;Oxford University Press: Oxford, New York, 2006,
ethylenedioxythiophene)," Synthetic Metals vol. 101, pp. 561-564, 1999. pp. 19-22, 196-199.
[14] G.Greczynski; Th.Kugler; M.Keil; W.Osikowicz; M.Fahlman; [33] D. Stauffer; A. Aharony, "Introduction to Percolation Theory,"
W.R.Salaneck, "Photoelectron spectroscopy of thin films of PEDOT-PSS Taylor&Francis: 1994; Second Edition., pp. 89-114.
conjugated polymer blend: a mini-review and some new results," J.of Electron [34] T. Kaevand; J. Kalda; A. Öpik; Ü. Lille; Springer: 2009; Vol.
Spectroscopy and Related Phenomena vol. 121, pp. 1-17, 2001. Advances in Computer and Information Science and Engineering.
[15] L. A. A. Pettersson; S. Ghosh; O. Inganas, "Optical anisotropy in [35] T. Kaevand; J. Kalda; A. Öpik; Ü. Lille, "Correlated percolating
thin films of poly(3,4-ethylenedioxythiophene)-poly(4-styrenesulfonate)," networks in the thin film of polymeric PEDT/PSS complex as revealed by the
Organic Electronics vol. 3, pp. 143-148, 2002. mesoscale simulation," Macromolecules vol. 42, pp. 1407-1409, 2009.
[16] J. Y. Kim; J. H. Jung; D. E. Lee; J. Joo, "Enhancement of electrical [36] R. Zallen; John Wiley Sons, Inc.: New York, Chicester, Weinheim,
conductivity of poly(3,4-ethylenedioxythiophene)/poly(4-styrenesulfonate) by Brisbane, Singapore, Toronto, 2004, pp. 135-251.
a change of solvents," Synthetic Metals vol. 126, pp. 311-316, 2002. [37] D. ben-Avraham, Diffusion and Reactions in Fractals and
[17] S. K. M. Jonsson; J. Birgerson; X. Crispin; G. Greczynski; W. Disordered Systems, vol. Cambridge: Cambridge University Press, 2004,
Osikowicz; A. W. Denier van der Gon; W. R. Salaneck; M. Fahlman, "The pp. 19-31..
effects of solvents on the morphology and sheet resistance in poly(3,4- [38] V. G. Gantmakher; Electrons and Disorder in Solids Clarendon
ethylenedioxythiophene)-polystyrenesulfonic acid (PEDOT-PSS) films," Press Oxford, 2005, p 206-211.
Synthetic Metals vol. 139, pp. 1-10, 2003. [39] T. Kaevand; A. Öpik; Ü. Lille; T., S., Ed.; Springer: 2008; Vol.
[18] A. Zykwinska; W. Domagala; A. Czardybon; B. Pilawa; M. Advances in Computer and Information Science and Engineering, p 540-546.
Lapkowski, "In situ EPR spectroelectrochemical studies of paramagnetic [40] S. Kirkpatrick, "Percolation and conduction," Reviews of Modern
centres in poly(3,4-ethylenedioxythiophene) (PEDOT) and poly (3,4- Physics vol. 45, pp. 574-588, 1973.
butylenedioxythiophene) (PBuDOT) films," Chemical Physics vol. 292, [41] V. Ambegaokar; P. I. Halperin; J. S. Langer, "Hopping
pp. 31-45, 2003. conductivity in disordered systems," Physical Rev.B vol. 4, pp. 2612-2620,
[19] R. R. Smith; A. P. Smith; J. T. Stricker; B. E. Taylor; M. F. 1971.
Durstock, "Layer-by-layer assembly of poly(ethylenedioxythiophene): [42] B. I. Shklovskii; A. L. Efros, Electronic properties of doped
poly(styrenesulfonate)," Macromolecules vol. 39, pp. 6071-6074, 2006. semiconductors vol. Berlin Heidelberg New York Tokyo: Springer-Verlag,
[20] S. Timpanaro; M. Kemerink; F. J. Touwslager; M. M. De Kok; S. 1984, pp. 95-136.
Schrader, "Morphology and conductivity of PEDOT/PSS films studied by [43] R. K. P. Zia; Y. Wu; B. Schmittmann, "Percolation of a collection
scanning-tunneling microscopy," Chemical Physics Letters vol. 394, pp. 339- of finite random walks: a model for gas permeation through thin polymeric
343, 2004. membranes," Journal of Mathematical Chemistry vol. 45, pp. 58-64, 2009.
Predicting Trading Signals of Sri Lankan Stock
Market Using Genetic Algorithms and
Neural Networks
M.M.K Dassanayake
Dr. Chandima Tilakarathne
Department of Statistics, University of Colombo,Sri Lanka
[email protected]
Abstract-This study predict the trading signals of Sri Lankan which corresponds to approximately 24% of the Gross
stock market using two sophisticated machine learning techniques Domestic Product of the country. The All Share Price Index is
called Genetic Algorithm (GA) and Neural Networks. These two the principal stock index of the CSE in Sri Lanka. ASPI
techniques in combination predict the direction (going up or not) measures the movement of share prices of all listed companies.
of the close price of tomorrow’s (day t+1) ‘All Share Price Index’ Hence, this study predicts the direction of ASPI on behalf of
(ASPI) of the Colombo Stock Exchange (CSE). The study period the Sri Lankan stock market.
considered was from 1st November 2002 to 31st December 2008. However, published researches related to the predictions of
The influential factors considered in this study represent the the CSE are very limited. Among the published work, the aim
intermarket influence, political and environmental factors, was to predict the value (price level), but not the direction. Past
economic stability and microeconomic factors: such as interest studies [1] showed the importance of directional prediction as
rate and exchange rate. A software called ‘genetic model’ was the predictability and the profitability of such predictions are
developed to find the optimum input variable combination that higher. Also there are no past studies related to predicting
will affect the direction of tomorrow’s ASPI value. Two identical ASPI using these techniques (Genetic algorithm and neural
neural network models called A and B were created for two network in combination) and this study may be the first of that
different time periods, to predict the direction of ASPI of day kind. Section II of this paper describes the methodology and
(t+1). the techniques applied. Section III includes data description
and data preprocessing. Section IV discuses how optimum
I. INTRODUCTION input variable combination will be selected using genetic
algorithm. Section V discusses how the directional prediction
Trading on the stock markets has long been considered as an can be performed using neural networks. At last, section VI
alternative investment which can make fast profits. Stock presents the conclusion, the limitations of the study and the
markets are very dynamic, where the functionalities and the improvements suggested.
value of transactions highly depend on the other local and
global factors as well as financial and non-financial factors. II. METHODOLOGY AND THE TECHNIQUES APPLIED
Hence, to obtain a profit through the stock market, it is
necessary to make intelligent trading decisions. In an attempt There are several variables that may affect the close price of
to gain advancement in decision making, there have been tomorrow’s ASPI [2]. From them several potential input
innovations in stock market trading technologies and tools. variables were selected based on inter-market influence,
The aim of the present study is to employ such technologies macroeconomic factors and news.
(Genetic algorithm and neural network in combination) to A preliminary analysis was carried out in order to identify
predict the direction of Sri Lankan stock market. Initially, the the association between these potential input variables and the
genetic algorithm will be employed to find the best set of input response variable. Graphical description, line charts and other
variable combination that will affect the direction of Sri statistical tests were incorporated, in order to identify their
Lankan stock market. Then neural networks will be used to trends. Then the variables with trends were preprocessed again
check the prediction ability of the best input variable in order to clarify the associations. The variables that have high
combination. impact on the tomorrow’s ASPI were selected as the ‘final
The Colombo Stock Exchange (CSE) is the main stock input variables’. Past data of these variables (six years) was
exchange in Sri Lanka. It is one of the most modern exchanges transformed in to binary status using relative return function
in South Asia, providing a fully automated trading platform. (1).Genetic Model was used to select the optimum input
As of 31 May 2008, 234 companies are listed on the CSE, variable combination out of theses final input variables that
representing twenty business sectors with a market will affect the direction of tomorrow’s ASPI. ‘Gens’ of the
capitalization of 850 billion rupees (over US$7.2 billion), chromosome [3] should be final input variables and the
269
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_47, © Springer Science+Business Media B.V. 2010
270 DASSANAYAKE AND TILAKARATHNE
The variables were used in the study represent the inter NEWS = 1 if there was non-negative news on the day t
market influence, microeconomic factors and news. The US NEWS = 0 if there was negative news on the day t
S&P 500 index, Indian BSE Senex index, the Amex oil index,
Exchange rate, Interest rate and news were considered as the The classification of the NEWS was done by considering the
potential influential variables that affecting the direction of the factors such as political situation, civil war situation,
close prices of the ASPI of day (t+1). The Stock market data environment factors, effective price changes of commodities
was downloaded from www.finance .yahoo.com website and and oil price change in day t.
the corresponding symbols for these stock markets are as
follows IV. GENETIC MODEL
• All Share price Index - ^CSE In this study GA is employed and created a Genetic Model
• US S&P 500 - ^GSPC with the intention of finding the optimum variable combination
• BSE Sensex -^BSE that will affect the direction of the close price of ASPI of day
• AMEX Oil Index -^XOI (t+1). Fig. 1, illustrates the structure of the Genetic model.
Criteria A
ଵ
݂ሺܥሻ ൌ ቂ ቃቀ ቁሺ ሻ (2) It emphasis that the optimum input variable combination that
ା
affects the direction (up or down) of the close price of the
p is the number of positive samples the rule covers, n is the tomorrow’s (day (t+1)) ASPI is ‘NEWS’ and ‘ASPI Lag1’.
number of negative sample the rule covers, P is the total ‘ASPI Lag 1’ stands for the close price index of ASPI of day t
number of positive samples and C is the rule or the condition. and ‘NEWS’ Represents the environment, political and war
The rule C of the study was, whether the close price of the related information of the day t.
direction of tomorrow’s ASPI will go up for a particular input
combination. V. NEURAL NETWORK MODEL
The termination criterion chosen for the Genetic Model is
related with the number of generations formed. Thus when the Neural network (NN) model was employed to predict the
algorithm reaches a specific number of generations it stops and direction of the close price index of the tomorrow’s ASPI value
returns the chromosomes structure of the fittest individuals. using the optimum input variable combination given by
Max allowed evaluations for the GA model [3] can be Genetic Model.
changed, by changing the value of the The data set considered for the study contains data over six
“MAX_ALLOWED_EVOLUTIONS“ variable in the years. But sometimes the earliest time series data did not
“OptimumStockVariables.java” module. After several changes affects the current time series data. Therefore the data set was
of the value of the “MAX_ALLOWED_EVOLUTIONS“ divided in to two time windows. So the first window was, from
variable, it has been decided that the two hundred evaluations 4th November 2002 to 24th February 2005 and the second
are high enough for successful convergence. So by adding window was, from 25th February 2005 to 31st December 2008.
further high value for number of evaluations will only add Hence two identical neural network models were employed for
more computational time. Fig. 3, illustrates the result of the the predictions; namely NN Model A and NN model B. Past
Genetic model. data related to optimum variable combination contains 1572
samples. Those data was divided in to two for the two NN
models. Since there should be training and validation data sets
TABLE I to NN to perform well, two samples from those data sets each
CHROMOSOME OF THE GENETIC MODEL have hundred observations) were separated for validation and
testing.
CSE 1st Lag GSPC BSE EX_R NEWS IN_R XOI There were two layers in the neural network model and the
weight adjustments did in batch mode [4]. The training
272 DASSANAYAKE AND TILAKARATHNE
algorithm was back propagation algorithm and the error By observing Fig. 5 and 6, it is clear that the NN Model B
function was mean square error function [5]. Both neural gives a high performance than NN Model A (since same
network models (A&B) had these same properties. The Fig. 4, criteria was used for the both NN Models, normalization of the
will illustrate NN model structure in a graphical way. performance values was not necessary for model comparison).
Neural Network tool box of MATLAB R2007a software was After Models were created successfully, another unseen data
used to create the NN models. Data related to training, testing of three months were given to the network models to evaluate
and validation were fed into the models and the outputs of the the model. Output of this data were recorded using the
models were shown in the Fig. 5 and 6. stimulate option of the NN tool box of MATLAB.
For evaluation, unseen data of three months will be given to
the NN model and stimulate the output. Since the transfer
function called tansigmoid [4] was used output values will be
within the interval [-1 1]. So following criteria was employed
to normalize those values.
Criteria C
TABLE II
HIT RATES FOR THE TWO MODULES
Abstract- We propose an automatic measurement algorithm for irradiation. Therefore, this method cannot provide the
the diameters of carbon nanotubes (CNTs) electron microscopy variations of numerous diameters, and also cannot estimate
images by using image processing. In the algorithm, Otsu’s the uniformity of diameters in a wide measurement area.
method (discriminant analysis method) was employed to Recently, a few methods to measure automatically the
determine automatically the threshold for image binarization. characteristics of nanomaterials using image processing were
Proposed algorithm provides the numerous values of diameters proposed.
detected by horizontal and vertical scans for the binary images. A measurement method for the characteristics of
The diameter measurements using developed program were MWCNTs was proposed [3]. This method provides outer
carried out for the multi-walled CNTs (MWNTs) taken by a radius, inner radius, and a physical characteristic of
scanning electron microscopy (SEM). MWCNTs. However, the selected areas to measure MWCNTs
We have confirmed that the proposed scan algorithm detects are restricted by the image segmentation procedures.
well the parts of diameters in the wide measurement areas. We Therefore, it is not effective to evaluate the uniformity of the
have also confirmed that the measured diameters ware close to diameters of CNTs in a wide area. Other image analysis
the values estimated manually. So, we can say that the developed
method for nanocomposites contained CNTs was proposed
program is able to detect the diameters for the clearly observed
CNTs.
[4]. However, this method cannot be used to measure CNTs
We suppose that this method is useful to evaluate the
diameters.
diameters of the CNTs not only for the management of the CNTs
In this paper, we propose a diameter measurement
products but also for the optimization of CNTs preparation
algorithm for CNTs electron microscopy images by using
processes.
image processing. In the algorithm, Otsu’s method
(discriminant analysis method) [5] was employed to
I. INTRODUCTION determine the threshold for image binarization. Proposed
algorithm provides the numerous values of diameters detected
Since the discovery of CNTs [1], numerous researchers by horizontal and vertical scans for the binary images. The
have investigated on single-walled CNTs (SWNTs), multi- diameter measurement was carried out for the MWCNTs
walled CNTs (MWNTs), their growth methods, their SEM images by the developed program.
diameters, their lengths, alignment mechanism, electron
emission property, gas sensing property, and device II. PROPOSED ALGORITHM TO MEASURE THE
applications. DIAMETERS OF CNTS
It is very important to evaluate uniformity of diameters of
CNTs in order to increase the uniformity of the characteristics Fig. 1 shows a MWCNTs image taken by a SEM. The
of CNTs for the above mentioned applications. The diameters proposed program measures the diameters of CNTs in the
of CNTs are generally estimate manually using the image rectangle surrounded by four points P0-P3. The mixture area
taken by a scanning electron microscopy or a transmission of image and characters is not measured because each pixel
electron microscopy (TEM). These human visual works are cannot be distinguished between image and character. Before
extremely inefficient. the image processing is started, the white color frame is added
There is a method using radial breathing mode (RBM) of for the circumference of CNTs image. The preparations are
resonance Raman spectroscopy to estimate automatically the necessary to begin the detection for CNTs regions.
diameters of CNTs [2]. However, this method is effective
only for SWNTs and is no use to MWNTs, and estimates only
the average diameter in the region of the Raman laser
275
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_48, © Springer Science+Business Media B.V. 2010
276 IWASAKI ET AL.
Median filtering
Fig 1. A sample of CNTs image.
Image Processing
Opening operation
End
Fig 3. Program of the image processing.
Nb N w
Pixel value No
Separation metrics: V between
2
f(x-1,y)=255(white)
V 2
within
Threshold is given at the maximum of separation metrics.
Yes
Fig. 4 The algorithm of Otsu's method.
Measure CNT lengths in
five directions 1-5
Start
Return
a CNT region
x x
Direction 3
Direction 2
Direction 1
y y Direction 10
Direction 6
Direction 5
End counting
pixels and convert Direction 8
the number of
Start counting pixels into the
pixels if new length unit.
CNT region is
detected.
(a) Horizontal scan (b) Vertical scan
Fig. 7 Explanation of the procedure for searching CNTs regions in x-direction (horizontal scan) and y-direction (vertical scan).
(c) Detected diameters shown in gray of sample 1 (d) Detected diameters shown in gray of sample 2
profile if the depth information is given by other microscopy [3] C. Gommes, S. Blacher, K. Masenelli-Varlot, Ch. Bossuot, E. McRae,
A. Fonseca, J. -B. Nagy, and J. -P. Pirard, “Image Analysis
like 3-dimensional SEM.
Characterization of Multi-Walled Carbon Nanotubes,” Carbon, Vol. 41,
pp. 2561-2572, 2003.
REFERENCES [4] J. Ryszkowska, "Quantitative Image Analysis of Polyurethane/Carbon
Nanotube Composite Microstructures," Materials Characterization,
[1] S. Iijima, “Helical Microtubules of Graphitic Carbon," Nature, Vol. Vol. 60, pp. 1127-1132, 2009.
354, pp. 56-58, November 1991. [5] N. Otsu, "A Threshold Selection Method from Gray-Level
[2] A. Jorio, R. Saito, J. H. Hafner, C. M. Lieber, M. Hunter, T. McClure, Histograms,” IEEE Trans. Systems, Man, and Cybernetics, Vol. 9, No.
G. Dresselhaus, and M. S. Dresselhaus, “Structural (n, m) 1, pp. 62-66, 1979.
Determination of Isolated Single-Wall Carbon Nanotubes by Resonant [6] G. Bradski and A. Kaehler, Learning OpenCV, O'Relly Media,
Raman Scattering," Physical Review Letters, Vol. 86, No. 6, pp. 1118- September 2008.
1121, February 2001.
Object Detection in Foveated Images
1 2
Abulfazl Yavari , H.R. Pourreza
1
Payam Nour university of Kashmar, Kashmar, Khorasan Razavi, IRAN
2
Ferdowsi University of Mashad, Mashad, Khorasan Razavi, IRAN
E-mails: Abulfazl_yavari @yahoo.com, [email protected]
281
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_49, © Springer Science+Business Media B.V. 2010
282 YAVARI AND POURREZA
ferent feature maps on different scales. One difficulty with this For each feature we first compute an image pyramid from
approach is that it is not clear how bottom-up and top-down which we compute scale maps. These present saliencies on
cues compete. Since there is evidence that two distinct brain different scales for different feature types. The scale maps are
areas are associated with bottom-up and top-down mechanisms fused into feature maps representing different feature type and
in human perception [12], it might be useful to separate the these again are combined to conspicuity maps, one for each
processing also in a computational system. feature. Finally, the conspicuity maps are fused to a single sali-
The next attention system is Hamker. It aims mainly a mod- ency map, with the degree of brightness proportional to the
eling the visual attention mechanism of the human brain [14]. degree of saliency. From the saliency map, the most salient
Hamker's model, shares several aspects with the architecture of region (MSR) is extracted and the focus of attention (FOA) is
NVT: he computes contrasts for several features (intensity, directed there.
orientation, red-green, blue-yellow) and combines them in fea-
ture conspicuity maps. The conspicuity of these maps is com-
bined in a perceptual map that corresponds to the common sali- A. Feature computation
ency map.
The channel for the feature intensity extracts regions with
Another currently considerable system is VOCUS, intro- strong intensity contrasts from the input image. First, the color
duced by Frintrop [1]. It has many similarities to NVT system image is converted into gray-scale. From the gray-scale image,
and it uses feature maps and saliency and IOR, which have a Gaussian image pyramid with five different scales s0 to s4 is
been used also in many currently introduced systems. Our pro- computed. The intensity feature maps is computed by center-
posed system is very similar to this one from the implementa- surround mechanism extracting intensity differences between
tion point of view and utilizes many of its ideas. But as one of image region and their surround, Similar to ganglion cells in
the goals of vision attention systems is to simulate the human's the human visual system [12]. The cells are divided into two
perception and his vision system, and foveated pictures are types: On-center cells respond excitatorily to light at the center
used in his vision system, it’s important to note that this system and inhibitorily to light at the surround, whereas Off-center
is unable to handle foveated images, but our system is capable cells respond inhibitorily to light at the center and excitatory to
of locating the goal in such images. light at the surround [12].
III. SYSTEM DESCRIPTION According to the human system, we determine two feature
In this section we introduce details of our system that try to types for intensity: the On-center difference responding
find region of interest in foveated images. The architecture of strongly to bright region on a dark background, and the Off-
center difference responding strongly to dark regions on a
system shares the main concepts with the standard models of
visual attention, especially based on the model of Koch and bright background. Then the maps for each center-surround
Ullman. The implementation is roughly based on the VOCUS variation are summed up by across scale addition: first, all
maps are resized to scale s2. The maps are added up pixel by
(Visual Object detection with a Computational attention Sys-
tem) [1], one of the best known attention systems currently. pixel. After summing up the scales this yields 2 intensity maps.
However, there are several differences in implementation de- Similar, orientation maps are computed from oriented pyra-
tails and type of the input images (Foveated images). mids [10]. The oriented pyramid in fact consists of four pyra-
The structure of our visual attention system is shown in fig.4. mids, one for each of the orientation 0, 45, 90,135. The orienta-
Before we go into the details of the system, we present a rough tions are computed by Gabor filters detecting bar like feature,
according to a specified orientation. The orientation scale maps
overview of the structure, as illustrated in figure; the input of
the system is a foveated image, which may be generated by are summed up by across scale addition for each orientation,
software or hardware. We use software tools to generate im- yielding four orientation feature maps of scales s2, one for each
orientation.
ages with the fovea in the centre of the image. First, a pre-
processing is performed on the input image, which can be per- To compute the color feature map, the color image is firstly
formed in two ways: whether by using Gaussian filters with converted into an LAB-image; an LAB image pyramid is gen-
different sizes, so that the larger filters are used in the center of erated. From this pyramid, four color pyramids are generated
the image and as we move from the center to the periphery, the for the distinct colors red, green, blue, and yellow. On these
size of the filter decreases. Another way is to use a 3*3 Gaus- pyramids, the color contrast is computed by center surround
sian filter and use it multiple times in the center of the image mechanism. The maps of each color are rescaling to the scale
and reducing the number of its use as we go from the center to s2 and summed up into four color feature maps.
the periphery.
Next three different feature dimensions are computed: inten-
sity, orientation, and color. For each dimension, the saliencies B. Fusing saliency
are computed on different scales and for different feature types,
e.g. red, green, blue, and yellow for the feature color. Thereaf-
To combine the feature maps, we should decrease the effect
ter the maps are fused step by step, thereby strengthening im-
of less important maps. To do so, we use the formula
portant aspects and ignoring others.
284 YAVARI AND POURREZA
A. Experiment 1
X = X / m (1 )
In which X indicates the map, and m is calculated like be- This image is a tile image that in the top of it there is a black
low: for instance, for the feature of color, we first calculate the dot (fig.5).
number of pixels (Np) with a value more than the threshold t, As illustrated in figure, our system detected the defect, but
and then determine the minimum value (min), and m for every VOCUS not able to detect it. Since the resolution in the center
map is equal to Np/min. Finally, formula 1 is used for saliency of image is better than periphery, the VOSUS worked wrongly.
maps and by adding them pixel to pixel, the saliency map is
obtained. Now we should detect the salient areas in the saliency
map. This is possible through finding pixels with largest val-
ues. These regions are the region that attracts the attention.
We use a green circle to show this region on input images.
(a) (b)
Fig.5. (a) the defect is detected by our system (b) but not by the VOCUS
B. Experiment 2
This image is a cloth image that there is a vertical red defect
on it: the defect is detected by our system, but not by the
VOCUS (fig.6)
(a) (b)
Fig.6. (a) the defect is detected by our system (b) but not by the VOCUS
C. Experiment 3
This image is an artificial image (pop-out). There is a red
blob among the blue one. In this image only one feature is dif-
Fig.4 structure of the system ference: color (fig.7).
D. Experiment 4 ACKNOWLEDGEMENT
This image is also a cloth image, that there is a yellow defect The authors would like to thank Simone Frintrop for useful
on it (fig.8) comments on this paper.
REFERENCES
[1] Frintrop, S: VOVUS: a Visual Attention System for Object Detection
and Goal Directed Search. PhD thesis University of Bonn Germany.
January 9, 2006
[2] Corbetta, M.: Frontoparietal cortical networks for directing attention
and the eye to visual locations: Identical, independent, or overlapping
neural system? Proc. Of the National Academy of sciences of the
United States of America, 95:831-838. 1990
[3] Desimone, R. and Duncan, J.: Neural mechanism of selective Visual at-
tention. Annual reviews of Neuroscience, 18:193-222. 1995
[4] Coltekin, A.: Foveation for 3D visualization and Stereo Imaging. PhD
(a) (b) thesis Helsinki University of Technology Finland. February 3, 2006.
[5] Chang, E.: Foveation Techniques and Scheduling Issues in Thinwire Vi-
Fig.8. (a) horizontal line is detected by our system (b) but not by the VOCUS sualization. PhD thesis New York University. 1998.
[6] Gonzales, R. C. and Woods, R. E.: Digital image processing. Addison-
Wesley Publishing Company, 1992
[7] Koch, C. and Ullman, S.:Shifys in selective visual attention: Towards
V. CONCLUSION AND FUTURE WORKS the underlying neural circuitry. Human Neurobiology 4 (4, 1985) 219-
227.
We presented a new visual attention system which is ena- [8] Treisman, A. M. and Gelade, G.: A feature integration theory of atten-
tion. Cognitive Psychology 12 (1980) 97–136.
bling of detecting attentive areas in the images with Non- [9] Milanese, R.: Detecting Salient regions in an Image: From Bilogical
uniform resolution, something that no other currently existing Evidence to Computer Implementation. PhD thesis, University of Ge-
system is capable of. According to the standard attention model neva, Switzerland. 1993.
of Koch and Ullman, and VOCUS, we concentrate on the three [10] Milanese, R., Wechsler, H., Gil, S., Bost, J., and Pun, T.: Integration of
bottom-up and top-down cues for visual attention using non-linear re-
features (color, intensity, orientation), since they belong to the laxation. In proc. Of the IEEE Conference on Computer Vision and
basic feature of human perception and are rather easy to com- Pattern Recognition, pages 781-785. 1994.
pute. Of course, this can only approximate human behaviour [11] Itti, L., Koch, C. and Niebur, E.: A model of saliency-based Visual At-
where many other features attract attention, e.g., size, curva- tention for Rapid Scene Analysis. IEEE Trans. On PAMI 20, pages
1254-1259, 1998.
ture, and motion [1]. [12] Palmer, S. E.: Vision Science, Photons to Phenomenology the MIT
In future works, we intend to complete the top-down part of Press 1999.
the system. In addition, we should improve the system's speed [13] Greenspan, H., Belongie, S., Goodman, R., Perona, P., Rakshit, S., and
to make it possible for being used in real time cases. We will Anderson, C.: Overcomplete steerable pyramid filters and rotation in-
variance. IEEE Computer Vision and Pattern Recognition (CVPR),
also study addition of the motion feature, which has an impor- pages 222-228, 1994.
tant role in attracting the attention. [14] Hamker, F. Modeling Attention: From computational neuroscience to
computer vision. In:proc. Of WAPCV'04 (2004) 59-66.
Digital Semi-Controlled Rectifier using a
Microcontroller
Emmanuel E. Rached, Adel I. Chit, Michel J. Owayjan
American University of Science and Technology, Beirut, Lebanon
Department of Computer and Communications Engineering
[email protected], [email protected], [email protected]
Abstract- Controlled rectifiers are an important aspect in power angle, separating the control circuit from the rectifier and
electronics; they are used to modify the average DC voltage at the accurately triggering the thyristors. The proposed solution has
output. Their use is widespread and they hold an important role three main parts. The first part provides the monitoring of the
in speed control of motors. Traditional analog control for these AC sine wave and zero crossing detection. Fig. 1 shows the
rectifiers was not always accurate and some required complex
designs. This paper presents a new low cost, high accuracy digital
circuit diagram of this part. The transformer steps down the
control method using a microcontroller. The microcontroller 220 VAC source to a more manageable 6 VAC voltage. Filter
keeps track of the current AC half-wave and triggers the capacitors are used on the both sides of the transformer. The 6
corresponding thyristors based on the desired time delay through VAC is further reduced to 3 VAC by the use of a voltage
the use of a potentiometer. Experimental tests showed a high divider. The 3 VAC is required so as not to damage the
accuracy and low response time in controlling the thyristor’s MCP601 operational amplifiers [6]. The sine wave is then fed
firing angle. to two MPC601 to provide a negative and positive reference
for the microcontroller.
I. INTRODUCTION
Unlike the more commonly used 741 Operational Amplifier
Most industrial applications use thyristor-based controlled which requires a 15 V source to operate reliably, the MCP601
rectifiers, especially where high voltages and currents are can perfectly operate with the same 5 V supply as the
involved [1, 2]. Rectifiers are used industrially to provide a microcontroller allowing the use of a single power supply.
variable DC voltage from AC sources to control the speed of The lower MCP601 takes as input the sine wave and, in the
motors [3]. They are also used in high voltage converter configuration shown here with the +5 V to Vdd, ground to Vss
stations to provide long distance transmissions. Other uses and comparing to ground, outputs a constant 5 V whenever the
include light dimming, power switching, and certain circuit sine wave is in the positive half-wave. This provides a logical
breaker applications [1, 2]. Traditional analog control of “1” (ON) to the microcontroller so it can reliably know that
thyristor-based rectifiers not only requires complex designs, the signal is currently in the positive half-wave.
but also lacks accuracy [4]. Digital control using a
microprocessor can improve the accuracy of such rectifiers
that can be used to drive a motor as a load [5].
287
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_50, © Springer Science+Business Media B.V. 2010
288 RACHED ET AL.
The second stage controls the firing angle through the use a Fig. 3. Microcontroller configuration
microcontroller. A potentiometer connected between +5 V
and ground provides the microcontroller with the desired
thyristors’ firing angle. This connects to the first Analog to As seen from the graph, the firing angle increases from 0º to
Digital converter of the microcontroller. A full +5 V as input 180º as the voltage decreases from 5 V to 0 V. This was a
instructs the microcontroller to provide a trigger pulse at 0º design choice. The microcontroller’s Analog to Digital
and 180º for the positive and negative half-waves respectively, converter allows for 10 bits of precision translating into 210 =
effectively turning the controlled rectifier into a normal bridge 1024 possibilities, dividing the 180º into 1024 steps of
rectifier. On the other hand, a 0 V input results in the approximately 0.1758º each. This provides a very high
microcontroller outputting a trigger pulse at 180º and 360º for accuracy in controlling the firing angle. The 4 MHz crystal
the positive and negative half-waves respectively, effectively oscillator provides fast execution time for the microcontroller
negating any output and making it 0 V. The remainder of the and the software optimizations were performed based on it.
range allows precise control of the firing angle between 0º and
180º. The signals coming from the first stage are directed
towards pin number 17 for the negative portion and pin
number 18 for the positive part allowing the microcontroller to
correctly know in what half-wave the sine signal is currently
in. The trigger pulse is then outputted on pin 24 for the
negative half-wave and on pin 23 for the positive half-wave.
Fig. 3 shows the configuration of the microcontroller. The
other elements shown in the illustration above are standard
configuration for a PIC16F877A microcontroller [7]. The
crystal oscillator used is operating at 4 MHz. Fig. 4 shows the
potentiometer voltage versus the firing angle.
Fig. 4. Firing angle versus Voltage
DIGITAL SEMI-CONTROLLED RECTIFIER USING A MICROCONTROLLER 289
Input: negative_input(boolean)
positive_input(boolean)
delay = A/D potentiometer voltage
Output: negative_output(boolean)
positive_output(boolean)
While(positive_input = 1)
{
delay = GetADValue()
wait(delay)
positive_output = ON Fig. 6. Semi-controlled rectifier example
wait(duration)
positive_output = OFF In the configuration shown, the transistors’ output remains
} off until a pulse is sent by the microcontroller on the
corresponding pin, at which point the transistors switches ON
The negative half-wave’s pseudo code is similar to what is and remains in this state for the duration determined by the
shown above. microcontroller.
The third and final part of the solution provides separation An added simplicity for this solution is that all the active
between its various components. 1 k resistors placed elements in the circuit can be powered from a single +5 V
between the output pins of the microcontroller and the base of power supply, simplifying power requirements by allowing
the BJTs limit the current draw from the microcontroller to a the use a single 5 V power supply. A standard 5 V, Zener
maximum of 5 mA to prevent damaging it. Pulse transformers regulated power supply was also implemented to accompany
provide the necessary separation between the transistors and this solution. Fig. 6 shows an example of a semi-controlled
the thyristors as those will most likely be used in high current bridge rectifier with a motor as the load implementing the
circuit. Fig. 5 shows the diagram of this stage. The upper solution presented in this paper with the corresponding
part handles the negative half-wave while the lower one negative pulse and positive pulse indicated.
handles the positive half-wave. Freewheeling diodes are used
between the components to eliminate flyback. III. TESTS & RESULTS
TABLE 1
RESULTS PRIOR TO MODIFICATIONS
Desired Angle Actual Angle Difference
0º 3º ~300%
2º 4º 100%
25º 27º 8%
75º 76º 2%
135º 137º 1.5%
Fig. 5. Final Stage 180º 188º 4.5%
290 RACHED ET AL.
TABLE 2 REFERENCES
RESULTS AFTER MODIFICATIONS
Desired Angle Actual Angle Difference [1] J. Millman and A. Grabel, Microelectronics, 2nd ed.,
0º 0.036º ~0.3% Mc-Graw-Hill. 1987.
2º 2º 0% [2] M. Rashid, Power Electronics: Circuits, Devices, and
25º 25º 0% Applications , 3rd ed., Prentice Hall, 2003.
75º 75º 0% [3] Wikimedia Foundation Inc., “Rectifier.” [Online].
135º 135º 0% 2009. Available: http://en.wikipedia.org/wiki/Rectifier.
180º 181º 0.5% [4] D. Maksimovic, R. Zane, and R. Erickson, “Impact of
digital control in power electronics,” Proceedings of
2004 International Symposium on Power
Semiconductor Devices & ICs, Kitakyushu, 2003,
IV. CONCLUSIONS pp. 13-22.
[5] K. Rahman, M. Choudhury, and T. Zulfikar, “An
This microcontroller powered digital control method is AT89C51 microcontroller based control circuit for dual
presented in this paper to provide a low cost, high efficiency three phase controlled rectifier,”Proceedings of the 3rd
alternative to currently available methods. Programmed in International Conference on Electrical & Computer
BASIC, the implementation shows both its high efficiency and Engineering ICECE 2004, Dhaka, Bangladesh,
low response time compared to other solutions. This presents December 2004, pp. 347-350.
a potential for cost and time saving for many industries. Some [6] Microchip Technology Inc., Datasheet of
limitations do exist for this particular solution. The code was MCP601/2/3/4, [Online]. Last accessed October 2009.
written for 50 Hz AC electrical sources. Also, the Available: http://www.microchip.com. 2004.
optimizations performed were based on the use of a 4 MHz [7] Microchip Technology Inc., PIC16F87XA Datasheet,
crystal oscillator. On the other hand, these drawbacks are [Online]. Last accessed October 2009. Available:
software based and thus can be tailored to other requirements http://www.microchip.com. 2003.
by simply modifying the microcontroller’s BASIC code.
Future work could include the use of a more general code for
the ability to work with any source’s AC frequency and
further optimizations. Other improvements would be the use
of a cheaper microcontroller as the PIC16F877A was only
used since it was immediately available. These improvements
are expected to further reduce the cost of this solution and
make it more widely available.
Generalized Minimum Variance Controller with
Dynamic Pole Assignment to Improve Performance
in Industrial Applications
M.A. Paz Ramos1, S.C. Garibo Esquivel2, O.J. Santos Sánchez3, R. Fernández del Busto4,
J.F. Briones de la Torre1, J.L. Gallegos Ramirez1, J. I. Rodríguez Martínez1
1
Universidad Politécnica de Aguascalientes, Aguascalientes, Aguascalientes, México.
2
Instituto Tecnológico de Aguascalientes, Aguascalientes, Aguascalientes, México.
3
Universidad Autónoma del Estado de Hidalgo, Pachuca Hidalgo, México.
4
Tecnológico de Monterrey Campus Ciudad de México, México Distrito Federal, México.
291
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_51, © Springer Science+Business Media B.V. 2010
292 RAMOS ET AL.
∂a 1 > − p o K
(13)
∂a 2 > − p 1 K
Thanks to all the evidence that has been presented through of
this section, we confidently postulate the following Theorem:
P ( s ) = p 0 s + p1 (15)
In Figure 4(a) the output of the process controlled is showed
while in the Figure 4(b) we can see the control signal.
Q( s) = 1
Suppose now, that there is a perturbation in one of the
assuming a K fixed in (14), the stability zone before parameters, particularly in the resistance one. The nominal
perturbations in the process parameters a1 and a2 when the value is drastically perturbed to R=37.5 in the 25th second.
loop is closed considering (15), is the S zone whose Incorporating the change to (17), the new process
boundaries are defined by coefficients are a1*=4.434 and a2*=26.14, which causes for
∂a 1 > − p o K the coordinate (a1*, a2*) to stay outside of the stability zone
(16)
∂a 2 > − p 1 K S. This situation can be appreciated thanks to the simulation
results showed in Figure 5.
In the parameters space a1-a2 (see Figure 3).
GMVPA Control
3
r,y
a) 1
-1
0 5 10 15 20 25 30 35 40 45 50
time (sec.)
3
1
b)
u
0
-2
0 5 10 15 20 25 30 35 40 45 50
In order to exemplify the Theorem 1, we choose to control time (sec.)
the capacitor voltage of an RLC circuit which is the low pass
Figure 4. GMVCPA for the RLC circuit example
filter whose transfer function is
GMVPA Control
20
Eo( s ) 1 / LC
= (17) 10
Ei ( s ) s 2 + R / Ls + 1 / LC
r,y
a) 0
-10
The values of the electric components are L=8.4572H,
C=0.004523F and R=69.02119Ω. Then the transfer function -20
0 5 10 15 20 25 30 35 40 45 50
can be rewritten as time (sec.)
20
Eo ( s ) 26 .14 (18)
= 10
Ei ( s ) s 2 + 8.161s + 26 .14
b) 0
u
-10
If the desired T polynomial is defined as
-20
0 5 10 15 20 25 30 35 40 45 50
T ( s ) = s 2 + 3.68s + 9.688 (19) time (sec.)
In the GMVCPA the cost polynomials are designed off line, whenever (t1, t2 ) > 0, being t1 and t2 the coefficients of the T
from the solution of a Sylvester matrix [2]. polynomial.
r,y
asymptotic convergence of the estimation we have a) 1
aˆ 1 ≈ a1 + Δa1 (27) 0
aˆ 2 ≈ a 2 + Δa 2 -1
0 5 10 15 20 25 30 35 40 45 50
time (sec.)
2
ª 1 º
ª q1 º « » b) 1
u
«q » « 0 »
1
« » = « (t − aˆ )»
0 (28) 0
« p1 » « K 2 2
» -1
« » «1 »
0 5 10 15 20 25 30 35 40 45 50
¬ 0 ¼ « (t 1 − a1 ) »
p ˆ time (sec.)
∂a1 = a1 * −t1
1
(29) a)
θ
∂a 2 = a 2 * −t 2 0.5
0
using (29), we can see that if T(s) fulfils the Stodola 0 5 10 15 20 25 30 35 40 45 50
condition, the boundaries of the stability wrap the perturbed time (sec.)
2 p1
-2 p0
a) 3
600
2
r,y(rpm)
500
1
400
0
0 10 20 30 40 50 60 300
time (sec.)
200
6 0 20 40 60 80 100 120 140 160 180 200
time(sec.)
4
b) 8
u
2
u(volts)
0 6
-2 4
0 10 20 30 40 50 60
time (sec.)
2
0 20 40 60 80 100 120 140 160 180 200
Figure 8. GMVCPA performance when there is a parameter time(sec.)
perturbation
GMVDPA Control
Figure 10. Motor speed control using GMVCDPA
5
a)
2
1
feeding of the motor. The process output is the belt
displacement of the belt in centimeters (cm.). We use the
0
0 10 20 30 40 50 60 reference shape used in the speed example. We can see in
time (sec.)
Figure 11(a), that we have remarkable tracking. Usually for
6 position control we use a P+D Control; when the system has
attained the reference value, the control signal becomes null.
4
b)
In this practical case, the motor has a dead zone between the
u
2
-3 and 3 volts in the input process. The GMVCDPA
0 perceives this situation, and allocates the control signal under
-2 the frontier of the dead zone. We can see in Figure 11(b) that
0 10 20 30 40 50 60
time (sec.)
the control signal is barely lower than the dead zone frontier,
which allows a faster response before a sudden reference
Figure 9. GMVCDPA performance when there is a parameter
change.
perturbation
Temperature control: In this experiment, we use a prototype
which consists of in two chambers. In the first chamber there
VI. INDUSTRIAL APPLICATIONS
is a heating source and in the second one, there is a
In this section, three typical industrial applications are temperature sensor (K thermocouple). Between the chambers
presented, using GMVCDPA. First, for DC motor, speed and there is a fan that forces the air to flow between chambers
(Figure 12). The process input voltage is applied to the fan.
296 RAMOS ET AL.
Fan speed increase causes the temperature to increase in the PID Auto-tuning DC 1020 Honeywell
second chamber. The process output is the temperature in the 45
r,y(temperature °C)
second chamber (°C). The process input has a dead zone
40
between the 0 volts and 3 volts.
Position Control
100 35
80
30
60 0 100 200 300 400 500 600 700 800 900 1000
r,y(cm)
time(sec.)
40
10
20
0 5
u(volts)
0 20 40 60 80 100 120 140 160 180 200
time(sec.)
0
10
5 -5
0 100 200 300 400 500 600 700 800 900 1000
u(volts)
0 time(sec.)
35
30
0 100 200 300 400 500 600 700 800 900 1000
time (sec.)
10
8
u (volts.)
6
b)
4
Figure 12. Temperature process
2
manufactured by Honeywell. When we connect the Figure 14. Temperature control using GMVCDPA
controller to the process, we select the auto-tune mode
(ATV). The control results are showed in Figure 13(a), the VII. CONCLUSIONS
reference deviation is approximately +/- 1.3°C, which for a
general industrial application is considered a good The behavior of the GMVC and GMVCDPA which is a
performance. We can see that the control signal goes near to modified version of GMVC, was analyzed under stability and
saturation periodically; one of the causes of such behaviour is system parameter variation viewpoints. GMVCDPA provides
the dead zone of the actuator and the control structure of the better results than GMVC in these two areas. The laboratory
commercial controller. experiments that were carried out with GMVCDPA, showed
the outstanding performance of this controller.
In Figure 14 (a), we show the ability of the system to control
the temperature control using the GMVCDPA. Using this VIII. REFERENCES
self tuning controller, we improve the performance in the
application. The reference deviation for this case is +/-0.3°C. [l] Clarke, D.W., Gawthrop B.A. “Self-tuning controller”, IEE Proc.
Control Theory Appl., 122, 1975, pp. 929-934.
In Figure 14 (b) we can see that the signal control is barely [2] Wellstead, P.E., Zarrop M.B. Self-tuning systems. John Wiley & Sons,
over the dead zone, allowing a better regulation. England 1991.
[3] Paz Ramos M.A., Quintero-Mármol M.E., Fernández del Busto R.,
“Generalized Minimum Variance with Pole Assignment Controller
Modified for Practical Applications”, Proceedings of the IEEE
International Conference on Control Applications, Taipei, 2004,
pp. 1347-1352.
[4] Poulin E., Pomerleau A., Desbiens A., Hodouin D., “Development and
Evaluation of an Auto-tuning and Adaptive PID Controller”,
Automatica, Vol. 32, No. 1, 1986, pp. 71-82.
Automated Detection of Sunspots and Sunspot
Groups in Full-Disk Solar Images
Sarab F. Majed
Department Of Mathematic, University Of Western Ontario, ON, Canada
Mehmmood A. Abd
Department Of Electrical And Computer Engineering, University Of Windsor, ON, Canada
V. Zharkova
Department Of Computing And Mathematics, University Of Bradford, BD7 1DP,UK
297
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_52, © Springer Science+Business Media B.V. 2010
298 MAJED ET AL.
Fig. 1. The full disk image obtained from SOHO/MDI on 1 January 2000.
Morphological processing facilitates the sunspots region Morphological image processing are dilation and erosion.
attributes extracting. Dilation is an operation that utilized to thicken or grow the
objects in binary images. On the other hand, the erosion
shrinks the objects. The amount and the way that the objects
grow or shrink is controlled by chosen selection of the
structure element, which can be a set of 4 or 8 connected
pixels. The structuring element that chosen in this research is
an 8- connected set. However, the dilation is mathematically
expressed as [3].
A ⊕ B = z | B
§ ∧ ·
∩ A ≠ φ
½ (3)
® ¨ ¸ ¾
¯ © ¹ z ¿
Where A is the empty and B is the structuring element. The
mathematically definition of erosion is.
{
AΘB = z (B ) z A c ≠ φ } (4)
The combination of dilation and erosion is yielding closing
and opening that are higher order morphological operations.
The opening is erosion of A by B, followed by dilation.
Opening removes small regions and thins the filaments of
Fig. 3. Smoothed image. object pixels. It can be expressed as.
Meaningful shape information of geometric objects,
measuring typical characteristics area, bounding box, centroid
A $ B = ( AΘB ) ⊕ B (5)
coordinate, and the length of a perimeter. Hence, finding the While the closing morphological is dilation followed by
sunspots coordinates (x, y), area in pixels, eases a erosion. Closing removes regions and thins filaments of
determination of the sunspots location in relation to the solar background pixels. The mathematical expression of closing is
center. expressed as.
(
A • B = A ⊕ B ΘB ) (6)
Applying the mathematical morphology operations is useful in
representation and description of a region shape such as
boundaries, skeletons, and convex hull. Closing fill the small
gaps in a penumbra shadow to expose the umbra borders.
Opening operation is utilized to opens small gaps between the
touching objects. Thus, morphological operations are being
useful in breaking narrow isthmuses and eliminating small
(spurious) objects. The sunspots separation makes counting
their number much easy [3].
Moreover, binary mathematical morphology is carried out in
this study. The area, centroid, and boundaries are extracted.
Then, bounding box coordinates are calculated for each spot
Fig. 4.The masked solar image with 23 potential sunspots. and later used to segregate each group as independent object.
Moreover, area of any sunspots is utilized to estimate the area Moreover, the number of objects is utilized to find out the
in heliographic coordinates. number of sunspots in the solar disk. This number is
originated from the labeling-connected components of the
4. SUNSPOT GROUPS DETECTION binary image. All region area more than five pixels consider
as a spot otherwise not. The experimental executed on data set
Extracting sunspots and organizing them into groups phase taken by SOHO/MDI. In the image which is taken on 24
is consists of two phases: The first phase employs January 2000 at 20: 00: UT, 59 sunspots are detected. This
Morphological operations to enhance the image, and the number is very close to maximum number that detected by
second phase gathers all related sunspots into group based on Mt.Wilson [1]. However, Figure 1 shows five sunspots groups
neighborhood features of sunspots .The longitudinal of the clearly detected in an image taken by SOHO/MDI on 1
group is used to determine which sunspot belonged to which January 2000. Table 1 shows a comparison between detected
group. The MacIntosh classification catalogue is used as sunspots number in our study and EGSO (Zharkova et al.,
standard classification in this paper [11]. The Morphological 2005)[6],[13].The false detection rate (FAR) represents the
operations are used to compensate the poor quality of the sunspots that are detected in our study but not by EGSO. The
images and reduce the noise similar to Zharkov et al., False rejection rate represents the sunspots that detected by
2005a,b[8],[12],[13]. Besides that they are used to enhance EGSO by not detected in our study. However, the variation
the objects /sunspots attributes. The fundamental of between the results are because image quality. However, the
300 MAJED ET AL.
detection of sunspots number in this study around 80 percent, D–Bipolar group with penumbra on spots at both ends of the
in comparison with EGSO, which could be promising results. group, and with length ≤ 10 .
0
Finally, all the adjacent sunspots that are governed by a E-Bipolar group with penumbra on spots at both ends of
certain number, which is varying between 3 to 15 heliographic 0 0
degrees. The length of longitudinal, as that mentioned in group, and with length defined as: 10 ≤ Lenght ≤ 15 .
MacIntosh Classification, is used to group sunspots together F-Bipolar group with penumbra on spots at both ends of the
into a one group. Before computing the length of longitudinal 0
group, end length ≥ 15 . H-Unipolar group with penumbra.
the sunspots group must be determined as Unipolar or bipolar. The principal spot is usually the leader spot remaining from a
Based on the conditions: “ Unipolar group: A single spot, or a pre-existing bipolar group“[11]. However, the above-
compact cluster of spots with the largest greatest separation mentioned MacIntosh classification conditions are used as
0
between spots ≤ 3 . In the case of a group with penumbra standard in this segmentation approach.
(class H ) ,the nearest is measured between the center of the
attendant umbra and the nearest border of penumbra 5. SUNSPOT GROUPS EXTRACTION ALGORITHM
surrounding the main spot “[11]. Bipolar Groups: Two or
more spots forming an elongated cluster of length ≥ 3 .
0 Applying morphological operations enhances the sunspots
Unusually there will be space near the middle of the cluster characteristics and simplifies the way for determining
dividing it into two distinct parts. Groups with a large sunspots boundaries [3]. In this paper, 8-connected
0 0 0 components are utilized to separate each sunspot and its
principal spot must be ≥ 5 in length (i.e 2.5 plus 3 ) to be splinters in the single area, or group. After that, a leading
classified as bipolar " [11]. MacIntosh classification is used as sunspot is taken as the central spot from which the boundary
standard to determine the greatest separation between sunspots of sunspots groups is defined. The leading spot is defined by
in each group: “ A Unipolar group with no penumbra. B - it’s area and location. The ending spot is usually larger in size
Bipolar group without penumbra on any spots. and has a stronger magnetic field and located slightly closer to
the equator than the trailing spots. Moreover, it is usually
TABLE 1
THE COMPARISON OF DETECTED NUMBERS OF SUNSPOTS IN located at the western direction and the trailers /followers in
THIS STUDY AND EGSO [13]. the eastern direction. These attributes are marked with respect
Date
No# Spots No#Spots
FAR FRR to the solar center. Consequently, the sub-region of sunspots is
Our study EGSO grouped in a single larger region, or as a sunspot group.
1-July-2002 23 23 0 0
2-July-2002 22 22 0 0
Sunspots are grouped by using the length of longitudinal that
3-July-2002 39 41 0 2 confines the group as that described in MacIntosh
4-July-2002 47 53 0 6 classification. A longitude length variation for a sunspot group
5-July-2002 38 39 0 1 0
6-July-2002 37 37 0 0
is ranging in 10 ≥ legth ≥ 15 heliographic degrees. In this
7-July-2002 33 32 1 0 paper, the algorithm starts by computing the bounding box for
8-July-2002 25 26 0 1 each sunspot. The largest bounding box for a leading sunspot
9-July-2002 17 17 0 0 will be adjusted to encompass all the neighboring ones. The
10-July-2002 27 29 0 2
11-July-2002 28 28 0 0
algorithm is iterative and terminates when no more expansion
12-July-2002 30 31 0 1 occurs or the length exceeds the maximum longitude length.
13-July-2002 53 53 0 0 If the initial bounding box has vector
14-July-2002
15-July-2002
66
43
66
43
0
0
0
0
( )
V0 = x0 , y 0, xw0 , yw0 and the neighbor box has vector
16-July-2002 44 45 0 1 Vr = ( x r , y r , xwr , ywr ) then the expanded bounding box
17-July-2002 58 61 0 3
18-July-2002 57 56 1 0 has a vector V = ( x, y, xw, yw) .
19-July-2002 41 40 1 0
20-July-2002 41 40 1 0 Where x = min ( x0 , x r ) and y = min ( y 0 , y r )
xw = max ( x r + xwr − x, x r + xwr − x )
21-July-2002 41 41 0 0
22-July-2002 37 37 0 0
23-July-2002
24-July-2002
42
70
42
71
0
0
0
1
yw = ( y r + ywr − y, y r + ywr − y )
25-July-2002 86 86 0 0 This algorithm has to be repeated until no more bounding box
26-July-2002 114 114 0 0 expansion occurred. The output of this algorithm will be an
27-July-2002 133 138 0 5 independent image of each group. However, each group image
28-July-2002 141 142 0 1
is extracted with the detail of the original image, such as a
29-July-2002 118 118 0 0
30-July-2002 100 101 0 1 date and time. Figure 4 shows four sunspot groups are
31-July-2002 83 83 0 0 detected, in image that taken by SOHO/MDI on 12 July
2002,in the smoothed image in Figure 3 corresponding to five
C-Bipolar group with penumbra on one end of the group, in of Mt Wilson. However, this difference is occurred due to the
most cases surrounding the largest of the leader umbrae. image quality and the size and location of some sunspots
AUTOMATED DETECTION OF SUNSPOTS AND SUNSPOT GROUPS 301
groups, which are very small and close to the edge of the sun. Mt Wilson did not detect them) and the false rejection rate
Besides, that the automated systems is based on extracting all (FRR) (Mt. Wilson detected the sunspots and we did not
available information from single image each day, whereas detect them). However, this variance of detection results due
sunspots number changes by the time. Table 2 shows the to the some external factors such image quality, clouds or poor
accuracy of our automated sunspot groups detection and visibility. From table 2 it is clear the result’s accuracy of
extraction in comparison with Mt Wilson for July 2002. detected sunspots groups in this study around 84 percent.
6. CONCLUSION
Fig.4. Four sunspot group detected in an image that taken by SOHO/MDI on This algorithm can be applied on any type of image
12 July 2000. format with minor modification and can be applied in
biological applications. Further more, the rustles can be
TABLE 2
THE NUMBER Of SUNSPOT GROUPS THAT ARE DETECTED FOR
facilitating the automotive sunspot group detection and
JULY 2002 IN COMPARISON WITH THOSE DETECTED BY MT. classification.
WILSON [10].
#Spot
Group
#Spot
group,
REFERENCES
Date Mt Wilson Detected FAR FRR
01-July-02 5 5 0 0 [1] David H. Hathaway, Robert M. Wilson And Edwin J. Reichmann, ”
02- July -02 7 7 0 0 Group Sunspot Numbers: Sunspot Cycle Characteristics,” Solar Physics
Vol., 211,2002, pp. 357–370.
03- July -02 10 9 0 1
[2] Guseinov, M. D,” The Polarity Of The Magnetic Field In Bright Umbral
04- July -02 8 10 2 0
Points In Sunspots,” Krymskaia Astrofizicheskaia Observatories,
05- July -02 11 11 0 0
Izvestiia, Vol. 68, 1983, pp. 36-39.
06- July -02 11 10 0 1 [3] Gonzalez, R. C, Wood, R. E,”Digital Image Processing,” Prentice Hall,
07- July -02 6 6 0 0 2004 , pp. 338-446.
08- July -02 7 7 0 0 [4] Observatories Royal de Belgium (2009-9-1) Sunspots Index Data Center
09- July -02 9 9 0 0 (SIDC) [Online]. Available: http://www.icsu-fags.org/ps11sidc.htm
10- July -02 7 7 0 0 [5] Mitic, M. Djordjevic ,A. Mihajlovic,S. J. “Changes In atmosphere and
11- July -02 - - Ionosphere Caused By Solar-geomagnetic Activity ,Within Belgrade
12- July -02 5 4 0 1 Area ,” ,” EGS XXVII General Assembly, Nice,(21-4-2009) [Online].
13- July -02 7 7 0 Available: http://adsabs.harvard.edu/abs/2002EGSGA..27..631M
14- July -02 6 7 1 0 [6] Zharkova et al., Solar Physics.(1-5-2009). European Grid of Solar
15- July -02 9 9 0 0 Observation, Solar Feature Catalogues [Online]. Available:
16- July -02 9 8 0 1 http://solar.inf.brad.ac.uk/pages/ss_start.php
17- July -02 7 6 0 1 [7] M. Sezgin and B. Sankur, "Survey over image thresholding techniques
18- July -02 4 5 1 0 and quantitative performance evaluation," Journal of Electronic Imaging.
19- July -02 5 7 2 0 Vol.13 (1) ,2004, pp. 146–165
20- July -02 7 7 0 0 [8]. S. Zharkov, V. Zharkova, S. Ipson, A. Benkhalil,” The Technique For
21- July -02 8 8 0 0 Automated Recognition On Full Disk Solar ,” EURASIP Journal .Vol.
15, 2005, pp. 2573-2584
22- July -02 9 8 0 1
[9] J. J. Curto, M. Blanca and E. Martínez, ”Automatic Sunspots Detection
23- July -02 10 9 1 0
on Full-Disk Solar Images using Mathematical Morphology,” Solar
24- July -02 11 10 1 0 Physic. Vol. 250, 2008, pp. 411-429.
25- July -02 9 9 0 0 [10] Mount Wilson Observatory, (13-12-2008) [Online]. Available:
26- July -02 12 10 2 0 http://www.mtwilson.edu/hk/ARGD/Sunspot_Cycle/
27- July -02 15 11 4 0 [11]. Patrick S. MacIntosh, (1990),” The Classification Of Sunspot Groups “,
28- July -02 13 11 2 0 Solar Physics Vol. 125,1990, pp. 251-267.
29- July -02 10 8 2 0 [12]. Zharkov, S., Zharkova and V.,Ipson, S., 2005b,” Statistical properties
30- July -02 11 9 2 0 of sunspots in 1996-2004 : I. Detection ,North-South asymmetry and
31-July-02 8 8 0 0 area distribution ,” Sol. Phys. ,228, PP. 377-397.
[13]. Zharkova, V., Aboudarham, J., Zharkov, S., Ipson, S., Benkhalil, A.,
Fuller, N., “ Solar feature Catalogue in EGSO,” Sol. Phys., Vol. 228,
Table 1 shows a comparison between our method and Mt 2005, PP 361-375.
Wilson reports results. Moreover, it is worthy to compute the [14]. Solar Heliospheric Observatory (SOHO), (10-9-
false acceptance rate (FAR) (where we detect the sunspots and 2009, [Online]. Available:http://sohowww.nascom.nasa.gov/
Stochastic Model Based Approach for Biometric
Identification
Mofakharul Islam1 Sitalakshmi Venkataraman2 Mamoun Alazab3
Internet Commerce and Security Laboratory
Graduate School of Information Technology and Mathematical Sciences
University of Ballarat, Australia
[email protected]
[email protected]
[email protected]
Abstract- In this paper, we present a new stochastic model based In biometric identification, where dental characteristics
approach for enhanced image segmentation in biometric such as spacing between teeth, teeth shapes, sinus patterns and
identification systems. Biometric features such as fingerprint, root / crown morphologies are used to identify human
face, iris, hand geometry and more recently dental features are individuals, image segmentation and classification is a
being used for human identification. Image analysis of each of
these biometric features has various challenges to overcome. To
complex task due to the non-standard regions of interest (ROI)
address such contemporary problems of image segmentation, we and information present in dental X-ray images. The content
provide a novel approach based on maximum a posteriori (MAP) variations are also attributed to the quality of contrast, non-
fitting Gaussian mixture model using Expectation-Minimization uniform intensities and digitisation / high frequency noise that
(EM) algorithm within the Bayesian framework. Our new are inherent in X-ray images. Hence, traditional image
algorithm captures the pixel intensity by the likelihood term in segmentation algorithms such as edge-based, region-growing
Bayesian Networks, and a priori biasing term of the spatial and threshold techniques that have been successful in certain
location information with the help of Markov Random Fields applications are not suitable for X-ray images [7], [8].
(MRF) model. We have employed a novel approach of using Similarly, contemporary approaches, which combine region
Daubechies wavelet transform for texture feature extraction that
uses MRF model and a robust technique of determining the
growing methods with edge-detection techniques or even
number of pixel classes based on Cluster Ensembles for a reliable statistical model based approaches, are either computationally
segmentation of dental X-ray images. We present how our expensive or more suitable for colour models [9], [10].
approach could be applied in dental biometrics to achieve very In order to address the above mentioned problems, this
fast and reliable human identification. Experiments show that paper proposes a novel stochastic model based approach to
our new unsupervised image segmentation technique provides perform image segmentation effectively. The application of
accurate feature extraction and teeth segmentation for effective our proposed approach in biometric identification using dental
biometric identification. images has been experimentally tested to produce effective
and accurate results for feature extraction and image
I. INTRODUCTION
segmentation. The rest of the paper is organised as follows. In
Human physical characteristics such as fingerprint, face, Section II , we present a modified classical finite mixture
retina scans, iris patterns, hand geometry, and others have model that employs Markov Random Fields (MRF) model
been used in various personal identification systems [1], [2]. with Expectation-Minimization (EM) algorithm under
However, in order to meet the specific requirements of Bayesian framework. Section III describes our novel approach
different biometric applications, more recently, researchers of image segmentation of dental X-ray images where the gray-
have been probing into other internal features such as teeth, level intensity and texture features play crucial role in pixel
spine and chest contours [3], [4]. Since internal features such clustering. Texture feature extraction are performed using
as iris and retina consist of soft tissues that could get damaged Daubechies wavelet transform and ensemble of clustering to
or deformed, recent research studies have established that arrive at reliable number of components/segments in an image
dental features form a more reliable candidate for biometric under unsupervised framework. In Section IV, we discuss the
identification as they are more robust and could be preserved experimental results of our proposed approach applied on
for a sufficiently long period of time, even after death [5], [6]. teeth X-ray images and in Section V we present our
Another motivation is that similar to medical diagnosis, X-ray conclusions and directions for future work.
image processing approaches could be adopted as low-cost
solutions for personal identification systems using human II. MAP FITTING GAUSSIAN MIXTURE MODEL
internal features such as teeth. Hence, in this research, we Among the stochastic model based methods, Gaussian
concentrate on the image segmentation approaches suitable mixture model or the finite mixture model has received
for dental biometric identification systems that are useful not significant attention from researchers in the last decade due to
only for ante mortem human identification but also for post- its improved segmentation results when compared to other
mortem identification, due to their survivability characteristic image segmentation techniques [11], [12]. A common
for a long period of time.
303
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_53, © Springer Science+Business Media B.V. 2010
304 ISLAM ET AL.
drawback of such statistical model based methods is their high efficiently handle X-ray images towards achieving increased
computational intensity. However, adapting a finite mixture accuracy in image segmentation.
model for image segmentation is recently becoming popular We present here a mixture model with an unknown
[13]. We propose a stochastic model based on classical finite number of components K, each of which has its own vector of
mixture model that employs Markov Random Fields (MRF) to density parameter θ j for each feature. Let x i denote the
capture the neighbourhood relationships among the
neighbouring pixels. The proposed model has some observation at the ith pixel of an image (i=1, 2,…..N) modeled
advantages, especially in the case of contextually dependent as independent and identically distributed (iid). So, the
physical phenomena like images, as it incorporates the probabilities of the ith pixel belonging to the jth class label are
outstanding features of the MRF model without additional given as follows:
computational burden. In the proposed method, MRF η ij = P ( j | x i ) (1)
introduces a priori distribution that takes into account the
neighbourhood dependency or relationship among the
neighbouring pixels. We elaborate a stochastic model based We assume that the density function f ( xi | Φ, Γ) at
segmentation framework based on Gaussian mixture models any observation, xi is given by the mathematical expression
below.
as follows:
The proposed model is an introduction of MRF model K
into the classical mixture model using Expectation-
f ( xi | Φ, Γ) = ¦ p ijψ ( xi | θ j ) (2)
Minimization (EM) algorithm under Bayesian framework.
j =1
Recently, approaches using MRF models and Bayesian
methods have provided answers to various contemporary where, ψ ( xi | θ j ) is a Gaussian distribution with
problems in image segmentation. The theoretical framework
parameters θ j = ( μ j , σ j ) , and K is the number of
relies on Bayesian estimation via combinatorial optimization
(Simulated Annealing). The application of a clustering method components in the mixture.
to image segmentation has the particular characteristics that
spatial information should be taken into account. Here, in We introduce Maximum a posteriori (MAP), a prior
addition to intensity-values of pixels, the pixel location must distribution for the parameter set Φ that takes into account
also be used to determine the cluster to which each pixel is spatial information based on the Gibbs function. According to
assigned. Conceptually, in most cases, it is desired that the the Hammersley-Clifford theorem [14], [15], Gibbs
same cluster label be assigned to spatially adjacent pixels. In distribution takes the following form:
implementing these ideas, the Bayesian framework provides a P (Φ ) = Z1 exp(− β ¦ V N i (Φ )) (3)
natural approach to address the problem of image
segmentation. Pixel intensity information is captured by the where, Φ is a vector of features, β is called the
likelihood term in Bayesian Networks, while a priori biasing regularization parameter and the normalizing constant Z
term captures the spatial location information with the help of is called a partition function. V N i (Φ ) denotes clique
MRF model.
m
In dental biometrics, where dental features are required to potential of the label configuration p within the
be extracted from X-ray teeth images, the existing image
neighbourhood N i of the ith pixel which can be
segmentation algorithms are ineffective if they are unable to
handle the inherent lighting and imaging conditions of X-ray calculated as;
images. We propose an adaptive Gaussian mixture model that
employs MRF to handle these changing conditions in X-rays. V N i (Φ ) = ¦ g (u i ,m ) (4)
The conceptual idea behind the MRF model is that the m∈N i
neighbouring pixels must have influences on a pixel. What we where, u i , m denotes the distance between the two label
capture through the MRF model is a pixel's interaction with its
i m
neighbouring pixels. In other words, we adapt the MRF model vectors p and p . The function g(u) must be gradually
to capture the spatial relationship among the neighbouring increasing and non-negative. The neighbourhood
pixels. As a result, it determines the probability of a pixel to
remain in the same cluster with its adjacent pixels. In this way,
N i consists of horizontal and vertical neighbours of i.
the spatial information captured addresses the various X-ray
imaging artefacts such as noise and non-homogeneity in To calculate the posterior probability distribution of one
intensity, as the neighbourhood information gathered is used random variable while the value of another is given, we make
to achieve a higher degree of discriminative ability in use of Bayes theorem. According to Bayes theorem [14], the
determining different pixel clusters or segments. We provide posterior probability density function (PDF) is defined as
below our mathematical model in adapting the MRF approach follows:
through the introduction of neighbourhood relationship among
the pixels in addition to the pixel features in order to Posterior PDF = Likelihood Term x Prior Probability
STOCHASTIC MODEL BASED APPROACH FOR BIOMETRIC IDENTIFICATION 305
Normalising Constant intensity and texture could play a significant role in achieving
this goal. Hence, in this research work, we have effectively
Then, Posterior PDF for a random variable X, given that the utilized the pixel’s gray level intensities and texture features
data Y = y takes the form as follows: for clustering purpose. Here, Daubechies wavelet transform
has been employed as texture descriptor. We combine i) the
gray level intensity, ii) Daubechies transform wavelet
descriptor, iii) Cluster Ensembles, a robust number of
component / segment finding tool, where the SNOB [17], [10]
where, fX(x) is the prior density of X and the likelihood is used as a starting point, and iv) the proposed MAP fitting
function as a function of x is given by Gaussian mixture model, into an integrated segmentation
model. The main purpose of such a novel image segmentation
approach is to achieve greater accuracy in meaningfully
,
segmented components that is required for a wide variety of
industrial images.
is the normalizing constant, We adopt the following four steps for the experimental
and is the posterior density of X, given investigation of our proposed dental X-ray image
the data Y = y. segmentation model:
Step 1: Extract the texture features,
Now, we can derive a posterior log density function as follows: Step 2: Find the number of components,
N Step 3: Determine the cluster ensembles, and
¦
P (Φ, Γ | X ) = log( x i | Φ, Γ) + log P (Φ )
i =1
(5) Step 4: Segment the image using the proposed MAP Fitting
Gaussian Mixture Model
The Expectation-Maximization (EM) algorithm requires that
i
Step 1: Extract the Texture Features
the computation of the conditional expectation value, z j of Texture characterizes local variations of colour or
the hidden variables must be at the Expectation Step (E-step). intensity in an image. The neighbourhood property is a
The E-step computes the expected value using current common similarity among all the texture descriptions.
estimate of parameter and the observed data in the image [16]. Statistical methods consider texture as random phenomena
Due to the application of the MAP estimation of the with identifiable local statistics. So, we can define texture as
parameters { p j } and {θ j } , the mathematical expression for an image feature which is characterized by the spatial
i
distribution of gray values (for X-ray images) or colour
z ij takes the form as given below: patterns (for colour images) in a neighbourhood surrounding
the pixel.
i(t ) i (t ) As the segmentation quality greatly depends on the
(t ) p j ψ ( x |θ j )
z i
= ( )
choice of feature (intensity, colour, texture, and co-ordinate)
¦
j K t
(6) descriptor, in our proposed approach, we have utilized pixel’s
p i ψ ( χ i |θ (j t ) )
l =1 l gray level intensity along with texture feature in the analysis
Next, the maximization of the following log-likelihood of contextually dependent physical phenomena of X-ray
corresponding to the complete data set is performed in the images. Hence, in this research work, we adopt a more context
Maximization Step (M-step) with t denoting the iteration step: sensitive image segmentation technique using wavelet
QMAP (Φ, Γ | Φ ( t ) Γ ( t ) ) (7) transform to produce reliable and accurate segmentation of X-
ray images based on subtle intensity and texture variation.
Although, wavelet transform has made significant
The M-step determines the maximum likelihood estimate of
contribution in several areas of image processing such as
the parameter using the data from the E-step [16]. For each
image enhancement, image compression, and image
parameter, QMAP can be maximized independently, which registration, its use in image segmentation is very limited and
yields the next iteration update equations for parameter of the is gaining attention recently [18-21]. Islam et al. [21] have
component densities μ j . Thus, the problem used wavelet transforms in their image segmentation approach
( t +1)
and [σ j ]
2 ( t +1)
to extract texture features from natural colour images and
of local minima in the EM algorithm has been successfully have reported promising results. However, image analysis in
addressed by using Simulated Annealing. dental biometrics demands a more sensitive texture descriptor
that could provide an efficient way to capture the subtle
III. PROPOSED DENTAL X-RAY IMAGE SEGMENTATION
variations in texture. The Daubechies wavelet transform has
APPROACH
the capability to better represent image semantics such as
In different imaging applications like medical diagnosis, intensity variations and segment configurations [22], [23].
biometrics, security and surveillance, mineral and mining, and Hence, in this research, we have used Daubechies wavelet
material science, where the primary goal is accuracy in image transform as the texture descriptor under Markovian
segmentation, making use of subtle information related to framework for the purpose of image segmentation as it is
306 ISLAM ET AL.
found to be more accurate and sensitive in capturing texture the image. Then the Expectation Maximization (EM)
details in an X-ray image. algorithm is used for estimating the model parameters. In fact,
model parameter estimation, incorporating the neighbourhood
Step 2: Find the Number of Components relationships among pixels and segmentation of pixels are run
Existing approaches are found to be either incapable of simultaneously within MAP estimation under Markovian
handling data with noise or do not perform well in handling framework. The whole process is being done using a
discrete data like image data [7], [10]. Our experimental study particular iterative procedure in an interleaved manner. In
indicates that an unsupervised data clustering approach is doing so, MAP classifies the pixels of the input images into
ideal for dental X-ray images and is found to be more different pixel classes on the basis of intensity and texture
effective in handling noisy image data that have diverse features. Here, the function QMAP plays a crucial role in
varieties of faults and imperfections.
After the extraction of the texture features, in this step, determining the actual membership of a class. In fact, the
intensity and texture features of pixel directly influence the
we input all these features along with pixel’s intensity value
into the Snob program [17] to determine the number of classes maximization of the QMAP function. That means every
present in the X-ray image. We have used the Snob program individual feature has some influence on the maximization
only as a starting point to find the initial number of process of the QMAP function. Some features have greater
components. In general, Snob has been found quite effective
influences while others have minimum effect. If the influence
in identifying the number of components in an image but
of a particular feature is greater than others in the
sometimes it produces spurious results due to its limitations to
handle some imaging artefacts like noise, shadow, and maximization process of the QMAP function for a certain
intensity non-homogeneity. Hence, we refine the number of pixel, then the clustering of that pixel is greatly influenced by
components by determining the cluster ensembles in the next that particular feature.
step.
IV. MODEL SIGNIFICANCE AND EXPERIMENTAL RESULTS
Step 3: Determine the Cluster Ensembles Recognition of a particular object in an image is a task
In this step, we devise a robust unsupervised technique to that is carried out as pixel classification and the most
find the number of components in an image automatically. An significant aspect in pixel classification is the feature
emerging approach like Ensembles of Clusterings could play a extraction. Feature extraction is a process of measurement and
significant role to augment the technique for finding accurate it forms the main role in the process of object recognition or
and authenticated number of components in an image. image segmentation. The performance of image segmentation
Ensembles of clustering are basically a combination of in an experimental study depends strongly on the choice of
multiple clusterings used to arrive at a consolidated clustering feature extraction approach.
without accessing the original data and algorithms. Here, only Our approach in image segmentation is different from
symbolic cluster labels are mutually shared. Multiple previous work in a few major ways. We find in existing
clusterings can be produced either using a single clustering literature, approaches that adopt pre-processing procedures
algorithm by varying the different parameters or using such as gum-line detection, pixel classification and adaptive
different clustering algorithms on the same data set and thresholding are based on classical techniques like
parameters [24], [25]. With Snob as a starting point for getting segmentation using edge, line, contour and region [26-28].
initial clusterings, we base upon Ensembles of Clusterings to The algorithms that are proposed in other previous work are
produce a consolidated clustering. In our experimental exclusively based on high level features using classical
procedure, we produced multiple clusterings for different approaches that are computationally more expensive. They
parameter values and then selected the stable solution found in have not used any stochastic or statistical models that provide
the plot of the number of clusters. The convergence criteria or maximum accuracy. In our novel approach, we have
number of iterations of MRF model is the main parameter that developed a stochastic model to make it efficient for pixel
we have varied to get multiple clusterings. based image segmentation. Though some research has been
conducted in using stochastic model based approach for
Step 4: Segment the Image Using the Proposed MAP Fitting
processing magnetic resonance (MR) images of human organs
Gaussian Mixture Model
for medical diagnosis [29], most of these existing approaches
Next, the four original features, such as intensity / gray are greatly influenced by similarity in intensity. In dental
scale level, and texture, along with the number of classes images, mere similarity in pixel’s intensity does not really
obtained from Ensembles of Clusterings, become the input to mean they belong to the same object. In the front of
our segmentation model as discussed in Section II. background intensity of dental images, many objects of
Accordingly, clustering of pixels is done on the basis of similar intensity may lie, which can be characterized through
homogeneity in intensity and texture properties. As a result, variation of their textures.
regions are obtained where there is a discontinuity in either In our proposed approach, we have used the Daubechies
intensity or texture. wavelet transform for texture feature extraction as it has the
In our segmentation model, the K-means algorithm ability to capture subtle texture information with better
determines the centres of the clusters for each component of
STOCHASTIC MODEL BASED APPROACH FOR BIOMETRIC IDENTIFICATION 307
V. CONCLUSIONS AND FUTURE WORK [10] C.S. Wallace, and D..L. Dow, “MML Clustering of Multi-state,
Poisson, Von Mises Circular and Gaussian Distribution”, Statistics and
In this paper, we have proposed a novel dental image Computing, Vol. 10, No. 1, pp. 73-83, 2000.
segmentation approach and have investigated its use in dental [11] G. McLachlan and D. Peel, Finite Mixture Models. New York: John
biometric identification. Our stochastic model based approach Wiley&Sons, 2000.
[12] P. Meinicke and H. Ritter, “Resolution-Based Complexity Control for
utilizes a Markov Random Field model to capture the Gaussian Mixture Models,” Neural Computation, vol. 13, no. 2,
relationships among the neighboring pixels and integrate that pp. 453-475, 2001.
information into the Expectation Maximization model fitting [13] M. Alf`o, L. Nieddu and D. Vicari, “A Finite Mixture Model for Image
Maximum a Posteriori algorithm. The paper has described Segmentation”, Statistics and Computing,, Vol. 18, No. 2, pp. 137 –
150, 2008.
unsupervised means of automatic discovery of classes or [14] S. Geman and D. Geman, “Stochastic Relaxation, Gibbs distribution,
clusters in images to produce precise segmentation of dental and the Bayesian Restoration of Images”, IEEE Transactions on Pattern
X-ray images using intensity and texture information along Analysis and Machine Intelligence, Vol. 6, No. 6, pp. 721–741, 1984.
with neighbourhood relationships among image pixels. Such [15] P. Clifford, “Markov Random Fields in Statistics”, In: Geoffrey
Grimmett and Domnic Welsh (Eds.), Disorder in Physical Systems: A
an approach has provided the required accuracy in Volume in Honour of John M. Hammersley, pp. 19–32. Oxford
segmentation as well as in finding the number of components University Press, 1990.
in an image automatically through the use of Cluster [16] T.K. Moon, “The Expectation Maximization Algorithm”, IEEE Signal
Ensembles and MRF model. In addition, our results are Processing Magazine, Vol. 13, No. 6, pp. 47-60 1996.
[17] D.M Boulton,. and C.S Wallace, “A Program for Numerical
enhanced by the use of Daubechies wavelet transform for Classification”, The Computer Journal, Vol. 13, No.1, pp. 63–69, 1970.
increasing the content sensitivity of the segmentation model, [18] D. Heric and B. Potocnik, “Image Enhancement by Using Directional
thereby addressing even noise in images. Finally, the Wavelet Transform”, Journal of Computing and Information
iterations of the algorithm have been converged using specific Technology- Vol. 14, No. 4, pp. 299–305 .2006.
[19] J. Feng, N. Xiong and B. Shuoben, “X-ray Image Enhancement Based
criterion on model parameters as they were calculated for on Wavelet Transform,”, Proceedings of IEEE Asia-Pacific Services
arriving at the resulting image segments of dental X-rays in an Computing Conference, APSCC, pp. 1568-1573, 2008.
interleaved manner. A pilot experimental study has tested the [20] M. Islam, J. Yearwood and P. Vamplew ”Unsupervised Color Textured
algorithm on real-life X-ray images and has provided Image Segmentation Using Cluster Ensemble and MRF Model” ,
Advances in Computer and Information Sciences and Engineering,
convincing results. Springer Netherlands, ISBN 978-1-4020-8740-0 (Print) 978-1-4020-
Our future work will test the performance of our proposed 8741-7 (Online), pp. 323-328, 2008.
approach described in this paper using large data sets so as to [21] M. Islam, P. Vamplew and Yearwood J., “MRF Model Based
determine the threshold level of noise in image that could still Unsupervised Color Textured Image Segmentation Using
Multidimensional Spatially Variant Finite Mixture Model”,
be catered within the model. Future work also entails in International Joint Conferences on Computer, Information, and
addressing the important research question with regard to the Systems Sciences, and Engineering (CISSE 2008), Bridgeport, USA,
practical implementation of human dental identification in 2008.
real-life biometric applications. [22] I. Daubechies, “Orthogonal Bases for Compactly Supported Wavelets”,
Communications on Pure and Applied Mathematics, Vol. 41, pp. 773-
REFERENCES 789, 1988.
[23] J.Z. Wang, G. Wiederhold, O. Firschein and S.X. Wei, “Content-based
[1] J. D. Woodward, N.M Orlans and P.T Higgins, Biometrics: Identity Image Indexing and Searching Using Daubechies’ Wavelets”,
Assurance in the Information Age, Berkeley, CA: McGraw-Hill/Osborne, International Journal of Digital Library, 1, pp. 311-328, 1997.
2003 [24] A.L.N. Fred, and A.K. Jain, “Data Clustering Using Evidence
[2] J. R. Vacca, Biometric technologies and verification system. Oxford: Accumulation”, Proceedings of the 16th International Conference. on
Butterworth-Heinemann, 2007. Pattern Recognition 'ICPR'02', Quebec, Canada, 2002.
[3] L.R. Long, S. Antania, D.J. Leeb, D.M. Krainakc, and G.R. Thomaa, [25] A. Strehl, and J. Ghosh, “Cluster Ensemble - A Knowledge Reuse
“Biomedical Information from a National Collection of Spine X-rays – Framework for Combining Multiple Partitions”, Journal of Machine
Film to Content-based Retrieval”, Proceedings SPIE , 2003 Learning Research, Vol. 2, pp. 583-617, 2002.
[4] A. K. Jain and H. Chen, ``Matching of Dental X-ray Images for Human [26] H. Chen and A. K. Jain, “ Tooth Contour Extraction for Matching
Identification,'' Pattern Recognition, Vol. 37, No. 7, pp. 1519-1532, Dental Radiographs,” Proceedings of 17th International Conference.
2004. on Pattern Recognition 'ICPR'04,’ Cambridge, UK, 2004.
[5] G. Fahmy, D. M. Nassar, E. Haj Said, H. Chen, O. Nomir, J. Zhou, R. [27] J. Zhou and M. Abdel-Mottaleb, ``Automatic Human Identification
Howell, H. H. Ammar,M. Abdel-Mottaleb, and A. K. Jain, “Toward an Based on Dental X-ray Images,'' Proceedings of SPIE Technologies for
Automated Dental Identification System,” Journal of Electronic Homeland Security and Law Enforcement Conference, the Biometric
Imaging, Vol. 14, No. 4, pp. 1–13, 2005. Technology for Human Identification Conference, Orlando, USA, 2004.
[6] S. Purushotham, and M. Anouncia, “Enhanced Human Identification [28] E. Haj Said, D. M. Nassar, and H. H. Ammar, “Image Segmentation for
System Using Dental Biometrics”, Proceedings of the 10th WSEAS Automated Dental Identification,” SPIE Electronic Imaging, San Jose,
International Conference on Neural Networks, Prague, Czech Republic, CA, 2006.
2009, pp. 120 -125, 2009. [29] Y. Zhang, Michel Brady and S. Smith, ‘‘Segmentation of Brain MR
[7] C.S. Wallace, and D.M. Boulton, “An Information Measure for Images Through a Hidden Markov Random Field Model and the
Classification”, Computer Journal, Vol. 11, No. 2, pp. 185-194, 1968. Expectation Maximization Algorithm,’’ IEEE Transactions on Medical
[8] N. R. Pal and S. K. Pal, “A Review on Image Segmentation Imaging, Vol. 20, No. 1, pp. 45- 57, 2001.
Techniques”, Pattern Recognition, Vol. 26, No. 9, pp. 1277 - 1294, [30] Islam, M., “Unsupervised Color Image Segmentation Using Markov
1993. Random fields Model”, Master by Research Thesis, Graduate School of
[9] Shih, F.Y. and Cheng, S., “Automatic Seeded Region Growing for Information Technology and Mathematical Sciences, University of
Color Image Segmentation”, Image and Vision Computing, Vol. 23, No. Ballarat, Australia, 2008.
10, pp. 844–886, 2005.
A New Stochastic Model Based Approach for Object
Identification and Segmentation in Textured Color
Image
Mofakharul Islam1 Paul A. Watters2
Internet Commerce and Security Laboratory
Graduate School of Information Technology and Mathematical Sciences
University of Ballarat, Australia
[email protected]
[email protected]
309
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_54, © Springer Science+Business Media B.V. 2010
310 ISLAM AND WATTERS
texture, and coordinates) simultaneously. Further, Markov where, α is called the regularization parameter, the
Random Fields (MRF) model is introduced to capture the normalizing constant Z is called a partition function,
spatial relationship among neighboring pixels in terms of its and VS x denote clique potentials of the label
color and texture features [1]. In doing so, it can efficiently
handle specific situations like recognition of different objects configuration within the neighborhood S x of
in an image which is usually associated with numerous the xth pixel.
artifacts such as noise and intensity inhomogeneity. So, the An MRF is characterized by its local property whereas a
proposed model has some advantageous features, especially in GRF (Gibbs Random Field) is characterized by its global
the case of contextually dependent physical phenomena like property. At this point, we need a global or joint probability
images as it incorporates the outstanding features of the MRF distribution to represent a whole image. The Hammersly-
model without additional computational burden. Cliford theorem establishes the equivalence of these two types
The remainder of this paper proceeds as follows: In of distributions. The theorem states that any conditional
Section II we present an EM Algorithm Fitting MAP distribution has a joint distribution which is Gibbs if the
Classifier followed by the proposed image segmentation following conditions are satisfied;
model using the aforesaid MAP classifier in Section III. In Positivity: Pr( X = x) > 0
Section IV, we present previous work in this specific field.
Section V presents experimental results demonstrating the Locality: Pr( X s = xs | X t = xt , t ≠ s, t ∈ S )
effectiveness and accuracy of the proposed approach and a
= Pr( X s = xs | X t = xt , t ≠ s, t ∈ N s )
discussion on the experimental results and finally in Section
VI, we present our conclusion. where, N s is the neighborhood of site s.
II. STOCHASTIC MODEL BASED MAP Homogeneity: Pr( X s = xs | X t = xt , t ≠ s, t ∈ N s ) is
CLASSIFIER the same for all sites s.
Clique potentials of the neighborhood can be calculated
Let us assume i x is the observation at the xth pixel of an as;
observation i x can be expressed as: of the hidden variables must be at the Expectation step due to
N MAP estimation of the parameters {ξ y } and
x
{λ j } ;
f (i x | Γ, Ω) = ¦ ξ yx Φ (i x | λ y ) (2)
x( t )
y =1 (t ) ξ Φ (i | Ω )
x (t )
From the above two roots, we select the root with positive color intensity and texture information are captured by the
sign + as it yields ξ yx ≥ 0 which provides a straight forward likelihood term in Bayesian Networks, while a prior biasing
term captures the spatial location information with the help of
update for the values of label parameters ξ yx of each pixel x at Markov Random Fields (MRF) model.
the M-step of each iteration of the EM algorithm. To satisfy
¦
N A. Color Feature Extraction
the other constraints i.e., 0 ≤ ξ y ≤ 1 and
x
y =1
ξ yx = 1 , we Color is an important dimension of human visual perception
utilize an active set method that use Lagrange multiplies. that allows discrimination and recognition of visual
information. Extracting and matching color features are
III PROPOSED COLOR IMAGE SEGMENTATION relatively easy. In addition to that, color features have been
APPROACH found to be effective in color image segmentation [1], [4], [5],
In this particular work, we have used YCrCb and Daubechies [6]. In this particular work, we prefer YCrCb as it has got
wavelet transforms as color and texture descriptors
312 ISLAM AND WATTERS
some specific advantage in vision and graphics as mentioned like Ensembles of Clusterings can plays a significant role to
earlier. augment the technique for finding a faithful and authenticated
number of components in an image. Ensembles of clustering
B. Texture Feature Extraction are basically a combination of multiple clusterings to get a
Texture characterizes local variations of image color or consolidated clustering without accessing the original data and
intensity. There is no formal or unique definition of texture algorithms. Here, only symbolic cluster labels are mutually
though texture based methods are commonly being used in shared. Multiple clusterings can be produced either using a
computer vision and graphics. Each texture analysis method single clustering algorithm by varying the different parameter
defines texture according to its own need. Texture can be or using different clustering algorithms on same data set and
defined as a local statistical pattern of texture primitives in parameter [11], [12]. We use SNOB as a starting point for
observer's domain of interest. Texture is often describe as getting multiple clusterings based on which Ensembles of
consisting of primitives that are arranged according to Clusterings produce a consolidated clustering. The procedure
placement rule which gives rise to structural methods of we have followed for determining the number of components
texture analysis that explicitly attempts to recover the is as follows: produce multiple clusterings for different
primitives and the replacement rules. On the other hand, parameter values and select the stable solution found in the
statistical methods consider texture as random phenomena plot of the number of clusters. The convergence criteria or
with identifiable local statistics. The neighborhood property is number of iteration of MRF model is the only parameter that
a common similarity among all the texture descriptions. So, we have varied to get multiple clusterings.
we can define texture as an image feature which is
characterized by the gray value or color pattern in a E: Segmentation of Image
neighborhood surrounding the pixel. After extraction of the six colors and texture features, we
We are using Wavelet transform to extract texture feature. put all the features into the Cluster Ensembles where the
Using Wavelet transform is a recent trend in image SNOB acts as a starting point to determine the number of
processing. Although, wavelet transform has made significant classes in the input image on the basis of these color and
contribution in the several areas of image processing such as texture features. Now, the six original features along with the
image enhancement, image compression, and image number of classes to be obtained from Ensembles of
registration, its use in image segmentation is very limited. Clusterings, are the input of our proposed MAP classifier as
Generally, Wavelet transforms performs better than other described in Section 2, where clustering of pixels will be done
texture descriptors like Gabor and MRSAR filters [10], [16]. on the basis of homogeneity in color and texture properties
Gabor filters are found efficient in capturing the strong along with the neighborhood relationships among the
ordered or coarser texture information while MRSAR filters neighboring pixels. As a result, regions are obtained where
works well in capturing weak ordered or finer texture details. there is a discontinuity in either color or texture and
Wavelet transform computation involves recursive filtering neighborhood.
and sub-sampling and each level, it decomposes a 2D signal .
into four sub-bands, which are often referred to as LL, LH, IV. RELATED PREVIOUS WORK
HL, and HH (L=Low, H=High) according to their frequency Recently, we [1] proposed a MRF model based
characteristics. In this particular work, we use Daubechies unsupervised color image segmentation approach, which is
wavelet transforms, a family of wavelet transforms discovered mainly focused on textured color image segmentation using
by Ingrid Daubechies to add more texture sensitiveness into Haar wavelet transform, CIE-Luv color space, and
our proposed approach. It has similar concepts like Haar neighborhood influences of neighboring pixels and found
wavelet but differs in how scaling functions and wavelets are promising results while applied on relatively less complex
defined. images. In this work, first order neighborhood has been
applied to capture the neighborhood influences among the
C. Finding The Number of Components neighboring pixels.
Considering the pros and cons, we have selected SNOB [2], Kato and Pong suggested a model based color image
[3], an unsupervised data clustering approach which has been segmentation approach that aims at combining color and
found more effective in handling noisy color image data that texture using Markov Random Field model theory [4]. The
have a diverse varieties of faults and imperfections. approach relies on Bayesian classification associated with
combinatorial optimization or simulated annealing where all
D. Cluster Ensembles the pixels are classified into different groups to form
In general SNOB has been found more or less effective in segments. The authors have used the perceptually uniform
identifying the number of components in an image but CIE-Luv color space and Gabor filters as color and texture
sometimes it produces spurious results due to its limitations to descriptors respectively. It appears from their experiments that
handle some imaging artifacts like noise, shadow, and they have tested their algorithm on synthetic and simple
intensity inhomogeneity. So, there is a pressing need to devise natural images only. Experimental results reveal that this
a robust unsupervised technique to find the number of approach is able to produce coarser (most prominent objects
components in an image automatically. An emerging approach only) segmentation on very simple images only. Further, no
A NEW STOCHASTIC MODEL BASED APPROACH FOR OBJECT IDENTIFICATION 313
model parameter estimation and number of components extract different objects from an image with the highest
estimation technique has been suggested by the authors which accuracy.
makes the approach primarily supervised in nature. Regarding the proposed approach, our overall aim was to
In 2003, Kato et al. proposed a similar approach using a find an efficient and appropriate machine vision technique to
multi-layer MRF model for unsupervised segmentation that identify and extract individual meaningful objects in textured
automatically determines the number of components in the color images. The results of the proposed algorithm have been
Gaussian mixture and the associated model parameters [6]. To compared with that of one of our similar previous work [1] in
find out the number of components in the mixture, they have order to make a direct comparison of the segmentation
introduced a mathematical term to the energy function of each accuracy between these approaches. Our proposed approach
layer. The paper does not provide any detail about how it differs with the existing method in four major ways; 1) a
works. Gaussian parameter estimation has been done by using second order neighborhood is employed to capture the
an adaptive segmentation scheme where at every 10th neighborhood relationships among the neighboring pixels
iteration, they simply recomputed the mean values and more efficiently, whereas in our existing method [1], first
covariance matrices based on the current labeling and the order neighborhood was employed, 3) to have better
initial estimates at feature layers are obtained via mean shift performance in texture feature extraction especially in terms
clustering. In this paper, they have used perceptually uniform of strong texture, we use Daubechies wavelet, whereas in our
CIE-Luv color values as color features and a set of Gabor and existing method we employed the Haar wavelet only, 3)
Multi-Resolution Simultaneous Auto-Regressive (MRSAR) as YCrCb color space is employed to get a better human
texture features. perceptual color features, whereas in our existing method, we
Kato and Pong suggested another stochastic model based employed CIE-Luv. Further, in our previous work, MRF
color image segmentation method based on MRF model, model, SVFM model, Simulated Annealing, and EM
which aims at combining color and texture features [5]. The algorithm were loosely coupled. Our new framework uses an
theoretical frame works relies on Bayesian estimation via integrated model to tightly couple these techniques.
combinatorial optimization. This work is basically a The comparison of results reveals that our proposed
comparative study among the segmentation results produced approach outperform the existing similar approach in terms of
by color and texture features independently and in segmentation accuracy and has higher discriminating ability to
combination. The authors have tested their algorithm on a identify disjointed segments or regions.
variety of color images including synthetic images and have
compared their results for supervised and unsupervised
segmentation using color only, texture only and combined
features. No technique for automatic determination of number
of components has been employed; instead they manually set
the number. Using Expectation-Maximization (EM) algorithm,
they have estimated the model parameters only which makes it
semi-unsupervised in nature. So, their claims regarding the
comparison of results between supervised and unsupervised
segmentation is invalid as the algorithm they have utilized for
unsupervised segmentation was not completely unsupervised
in nature.
Yiming et al. [10] suggested another unsupervised color
image segmentation approach based on stochastic model Fig. 1. Segmentation Results with our proposed approach
where they employed Minimum Message Length (MML)
principle for finding the number of components in images and In Figure 2, the existing method has found 2 segments in
Maximum Likelihood Estimate (MLE) for pixel classification. the top image but our proposed approach able to identify 3
This is completely a color feature based approach where segments (Figure 1). If we look carefully to the image, in
neither texture feature nor has neighborhood relationship been existing approach wooden balcony in the image is merged
considered. with the building though there is a significant color and
texture difference in between them, while the proposed
V. DISCUSSSION ON EXPERIMENTAL approach able to identify the wooden balcony as disjoint
RESULTS region or segments. Our approach is able to capture the subtle
To qualitatively evaluate the performance of the proposed similarities in color and texture and able to extract or segment
approach we tested this algorithm on a wide variety of natural the region on the basis of these similarity. In Figure 1, our
images comprising different number of regions or objects and approach has identified 3 different objects in the middle image
get promising results but due to space constraint we would not while the existing approach found 4 segments or objects
be able to present all of them. The images are sourced from (Figure 2). In this case, texture features plays the dominant
Bakerely Segmentation Dataset [11]. role in the segmentation process so ignores the dissimilarity in
Figure 1 and Figure 2 show how our proposed approach color and considers the region as a single segment. In our
takes advantage of the discriminating ability to identify and
314 ISLAM AND WATTERS
REFERENCES
315
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_55, © Springer Science+Business Media B.V. 2010
316 ISLAM AND WATTERS
MAP classifier that we use for image segmentation purpose. each color plane and interaction between different color
In Section III, we present previous work in this specific field. planes. They proposed an Agglomerative Hierarchical
Experimental results demonstrating the accuracy and Clustering for pixel segmentation. The model parameter has
efficiency of the proposed approach are discussed in Section been estimated using a Maximum pseudolikelihood scheme
IV and finally in Section V, we present our conclusion and but no automatic technique has been used for estimating the
future work. number of mixture components.
II. RELATED PREVIOUS WORK
Sanjay-Gopal and Herbert [1] proposed a novel approach III. PROPOSED IMAGE SEGMENTATION MODEL
for pixel labeling and gray scale image segmentation based on
a Bayesian framework using a finite mixture model where an Our proposed image segmentation model composed of
EM algorithm is used for maximum likelihood estimation of following three major sequential phases;
the pixel labels and the parameters of the mixture densities. 1. Image feature extraction
The Bayesian framework provides us with a natural approach 2. Determination of number of components in the
to implement the idea that the application of clustering image using a robust number of object/component
methods to image segmentation has the particular finding tool
characteristic, which takes spatial information of the pixel into 3. Pixel classification using a stochastic model based
consideration in addition to its intensity values. According to MAP classifier
this formulation, a likelihood term is introduced that is based
exclusively on pixel intensity information while spatial A Image Feature Extraction
location information is captured by a prior biasing term that Color textured image has different features like intensity,
uses MRF model theory. color, texture and coordinates. In our project, we have used
Recently, we [2] proposed a finite model based color and texture features for the segmentation purpose. To
unsupervised color image segmentation approach, which is segment an image into objects, nine features are extracted
mainly focused on textured color image segmentation using from the image. Three features are color features and the rest
Haar wavelet transform, CIE-Luv color space, and are texture features. CIE-Lab color space has been chosen for
neighborhood influences of neighboring pixels and found color feature extraction considering its several advantages as
promising results while applied on relatively less complex discussed in [11]. To obtain the other six features, we apply
images. the Haar wavelet transform to the L components (number of
In [3], a novel approach for segmentation of color pixels) of the image to get three features and Gabor filters
textured images based on Markov Random Fields has been under three different orientations for the rest three features.
proposed where they use CIE-Luv color values as color
features and a set of Gabor filters as texture features. The B Color Feature Extraction
proposed approach is supervised in nature. As such, no model Choice of a suitable color space is one of the main aspects
parameter estimation technique has been used and the number of color feature extraction. Most of the color spaces are three
of components was selected manually. In 2003, Kato et al. [7] dimensional where different dimensions represent the different
proposed an algorithm using a multi-layer MRF model for color components. The most common color space in computer
unsupervised segmentation that automatically estimates the graphics is RGB. Unfortunately, this 3D color space does not
number of components in the Gaussian mixture and the correspond to equal perception of color dissimilarity. As a
associated model parameters. To find out the number of result, alternative color spaces are generated by transformation
components in the mixture, they introduced a mathematical of the RGB color space. The color spaces like HSV, CIE-Luv
term to the energy function of each layer. Gaussian parameter and CIE-Lab can be generated by non-linear transformation of
estimation has been done by using an adaptive segmentation the RGB space. We prefer CIE-Lab as it represents the three
scheme where at every 10th iteration, they simply recomputed characteristics (hue, lightness and saturation) that characterize
the mean values and covariance matrices based on the current the perceptually uniform color efficiently.
labeling and the initial estimates at feature layers are obtained
via mean shift clustering. In this paper, they have used C Texture Feature Extraction
perceptually uniform CIE-Luv color values as color feature The texture features are constructed in such a way that
and set of Gabor and MRSAR filters as texture feature. similar texture represents similar intensities. As a result, a well
Further, Kato and Pong [5] proposed another similar algorithm determined value with some variance is assigned to the pixels
where they used CIE-Luv color values as color feature and a with a given texture. The image textures are being extracted
set of Gabor filters as texture feature. They utilized the EM through the wavelet transform as already mentioned earlier.
algorithm for estimation of the model parameter. They set the Wavelet transform computation involves recursive filtering
number of components (pixel class) manually. In another and subsampling; and each level, it decomposes a 2D signal
approach [8], an unsupervised segmentation algorithm which into four sub-bands, which are often referred to as LL, LH,
uses Gaussian MRF model for color textures. Their model HL, and HH (L=Low, H=High) according to their frequency
characterizes a texture in terms of spatial interaction within characteristics [9].
AN UNSUPERVISED STOCHASTIC MODEL FOR DETECTION AND IDENTIFICATION 317
the channels are represented by a bank of real-valued, even- t=iteration step, Φ =set of probability vector {r , r ,....r }for
1 2 n
4 i =0 j =0
k + i ,l + j
conditional expectation value of the hidden variables ρ ij is
The reason for choosing Haar wavelet is that it has better derived as follows;
reflection of texture properties where the coefficients in (t )
rji δ ( xni | θ j )
(t )
i( t )
different frequency bands signal variations in different ρ = (4)
¦ i( t )
j K (t )
directions, such as horizontal, vertical, and diagonal [4]. In
k =1 k
r δ ( xni | θ k )
addition to that, Haar transform require less computation
The density function f ( xn | Φ, Ω) at an observation xn
i i
compared to other wavelet transform with longer filters. We
extract three texture features using Haar wavelet transform. is defined as;
Rest of the three texture features are being extracted using K
Gabor filters under different orientations. Different f ( xni | Φ, Ω) = ¦ rjiδ ( xni | θ j ) (5)
orientations can be obtained by rotating the coordinate system. j =1
In our tests, we used three orientations: 0o ; 60o ; 135o . Where, δ ( xni | θ j ) is a Gaussian distribution with
parameters θ = {μ j , σ j } .
j
D Determination of number of classes
We have tried several data classification approaches to In order to incorporate the neighborhood influences
identify the number of components automatically but most of among the neighborhood pixels, Maximum a priori (MAP)
them has been found inappropriate for handling image data introduces a prior distribution for the parameter set - that takes
properly and need major adjustment through a huge series of into account spatial information based on the Markov Random
trial and error. In that perspective, AutoClass C [10], an Fields (MRF) Model theory. An MRF is characterized by its
unsupervised Bayesian classification approach from NASA local property. At this point, we need a global or joint
has been found effective in automatic determination of the probability distribution to represent a whole image. The
number of classes in the input images on the basis of supplied Hammersly-Cliford theorem establishes the equivalence of
color and texture features. these two types of distributions. The theorem states that any
conditional distribution has a joint distribution which is Gibbs
E Stochastic Model Based MAP Classifier if the following conditions are satisfied;
Let X i = ( x1 , x2 ,....xn ) denote the observation of ith pixel
i i i
Positivity: Pr( X = x) > 0
i
of the image, where x n represents pixel values in the nth Locality: Pr( X s = xs | X t = xt , t ≠ s, t ∈ S )
dimensional plane. So, the probabilities of the ith pixel
belonging to the jth class in the nth plane are given as:
= Pr( X s = xs | X t = xt , t ∈ N s ) , where
rji = P( j | xni ) (2) N s is the neighborhood of sites s.
The objective function of our proposed MAP classifier is Homogeneity: Pr( X s = xs | X t = xt , t ∈ N s ) is the
defined as; same for all sites s.
N
The following Gibbs function takes into account spatial
MAP(Φ, Ω | Φ Ω ) = U − α ¦
t t
¦ q(l i ,h ) (3) information and express it in the form of prior
i =1 h∈Ni
distribution p (Φ ) ;
318 ISLAM AND WATTERS
i =1 Therefore,
constant, ® is a regularization parameter,
1
CNi (Φ ) denotes the clique potential function of the pixel [¦ h∈N q(li , h ) + ρ ij [¦ h∈N q (li ,h )]]
rji
( t +1)
=F±
i α i
label vectors { p } within the neighborhood N i of the ith 2[¦ h∈N q(li , h )]
h
i
pixel which can be defined as ;
(10)
C N i (Φ ) = ¦ q(l
h∈Ni
i ,h ) (7) From the above two roots, we select the root with positive sign
+ as it yields rj ≥ 0 which provides a straight forward update
i
¦
h 2 K h 2
Where li , h =| r − r | = (r − r ) and N i is
i i
i
j =1 j j for the values of label parameters rj of each pixel i at the M-
the set containing pixels that are either horizontally or step of each iteration of the EM algorithm. To satisfy the other
vertically adjacent to the pixel i. Here, we have adopted
¦
K
the function q (l ) = (1 + l )
−1 −1
from [17], which is more
constraints i.e., 0 ≤ rj ≤ 1 and
i i
j =1 j
r = 1 , different well
robust to outliers. studied mathematical approaches are available. In this
particular work, we utilize an active set method that use
Now, based on the above prior density, a posteriori log- Lagrange multiplies.
density function can be defined as;
N
p (Φ, Ω) = ¦ log f ( xni | Φ, Ω) + log p(Φ) (8) IV.
DISCUSSSION ON EXPERIMENTAL
i =1 RESULTS
Recognition of a particular object in an image is a task
To estimate the parameter rj and
i
θh for MAP, the that is carried out as pixel classification, which in many cases
Expectation-Maximization algorithm needs the conditional follows recognition of the individual object on the image. The
expectation values ρ ij as defined in Eqn. (4) at the most significant aspect in pixel classification is feature
extraction. Feature extraction is a process of measurement that
Expectation step while Maximization step maximizes the performs the main role in the process of object recognition.
objective function of our MAP classifier as defined in Eqn. The performance at segmentation depends strongly on the
(3). choice of feature extraction.
However, maximization of the function MAP with respect In this work, our overall aim was to find an efficient and
i
to the label parameters rj does not provide a closed form of appropriate machine vision technique to identify individual
objects in color textured images, which is capable to take into
update equations. In addition, the maximization procedure
account the subtle variations in color and texture and produce
must also take into account the constraints 0 ≤ rj ≤ 1 and
i
precise segmentation.
A series of image segmentation experiments on a wide
¦
K
r = 1.
i
j =1 j variety of color textured images have been carried out to
To address these difficulties, we take some new measures evaluate and compare the effectiveness of the proposed
in the M-step of EM algorithm. We set derivative of Eq. (8) approach. Since the main contribution of our work is
equal to zero to get the expression in the following quadratic development of a new stochastic model based image
form; segmentation approach, we compared our proposed approach
Suppose, with one of our earlier work that employs a stochastic
approach based spatially variant finite mixture model [2]. Our
E = 4α [¦ h∈N q(li ,h )](rji
( t +1)
)2 proposed approach differs with the existing method in four
i
Sarab F. Majed
Abstract__A new effective technique is presented for automatic group of spots. Sunspot groups can have infinite different
classification of sunspot groups on full disk white light (WL) formations and size, ranging from solo spots to the giant
solar images. This technique is implemented on images taken groups of sunspots with complex structures [1]. Despite such
from the Solar Oscillations Investigation Michelson Doppler complexity and diversity, astronomers have been able to
image (SOI/MDI) aboard the Solar Heliospheric observatory
(SOHO). The technique focuses on employing Support Vector
classify the sunspots classes based on The McIntosh Sunspot
Machines (SVMs) as effective classification tool. In addition to Classification Scheme and modified Zurich Scheme [2]. A
applying SVMs the problem of extracting sunspots and sunspot sunspot is recognized by its two main features ; the central
groups from solar image is solved in an efficient and different dark features ‘umbra’; and the outer lighter area which is
way from the ones previously employed. This technique proceeds called ‘penumbra’ (See Figure 1).Therefore,most
in several consequence phases. The first phase involves solar disk computerized recognition techniques rely on these properties .
image extracting. The second phase involves Binarization and Many manual sunsopts in different formats are produced
smoothing of an extracted solar image disk. The third phase all over the world , such as the Meudon Observatory in France
involves unsupervised segmentation of sunspots groups. This ,the Locarno Solar Observatory in Switzerland ,the Mmount
phase consists of two subphases: a) extracting the spots and, b)
combining together the sunspots that belong to the same group.
Wilson Observatory in the United States Of America and
The fourth phase involves attributes extraction of each sunspots many others . Zurich classification is based on the Locarno
group. The final phase is the classification phase using SVMs, by Catalogue which is considered a primary indicator of solar
applying a one–against–all technique to classify the sunspots activity [2][3]. The Zurich Classification is categories into
groups. The proposed technique has been tested using different seven classes of sunspots groups numbered from A to H [2]
sets of sunspots achieving approximately 88.56% recognition rate .(see Figure 2 and Table 1 ). Because of the effect of the solar
for Zurich classification of sunspot groups. activity on the Earth’s atmosphere, human activity, businesses
and even climate, the automated detection and verification
Index Term__Solar Disk, Sunspots Groups, Support Vector Machines, techniques become essential for a reliable forecast of the
Unsupervised Segmentation.
future effects on human lives and businesses. This requires
stringent requirments on the classification accuracy of any
I. INTRODUCTION
automated recognition system. Many researchers have been
S unspots have been a subject of interest to scientists and
researchers for many years. A sunspot is a region in the
developing automated reliable recognition systems with
varying recognition rate and techniques [3].
sun’s photosphere that is darker by comparison with its
photosphere background and has intense magnetic field. A
sunspot may consist of one or more umbra which often are
surrounded by a less dark area called penumbra. Sunspot
numbers increase and decrease with an irregular cycle with a
length of approximately 11 years [1]. Sunspots can be
observed in the visible continuous spectrum, also known as “
white light ‘. Larger sunspots can be observed by using Call
Fig.. 1. A high-resolution SOHO/MDI image showing sunspots with umbra
K1 absorption line image and H and Call K3 absorption line (black spots) and penumbra (dark gray) above the background (gray).
images . Sunspot shapes are varied from individual spots to a
321
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_56, © Springer Science+Business Media B.V. 2010
322 ABD ET AL.
II. PREPROCESSING
Preprocessing consists of image acquisitions, edge
detection, noise removing and feature enhancing, and
Binarization. The input images are taken from the solar
Oscillations Investigation Michelson Doppler Imager
(SOI/MDI) aboard of the solar Heliospheric Observatory
(SOHO)[11]. Figure 3 shows an example of the input white
light full disc image .The image has dimensions of 1024x1024
pixels with 16 different intensities. The resolution is 2.6
arces/pixel. Edge detection, noise removing and Binarization
are curried out as that reported by Sarab et al [4].
III. SEGMENTATION TECHNIQUE four different sunspot group images where A represents the
pores and groups B, C, and D respectively [2].
In this paper an unsupervised segmentation method was
employed based on the similarity measure. The sunspot’s is IV. FEATURE EXTRACTION
extracted /cut from a full disk solar disk image by computing
the bounding rectangle (box) that contains it. In this phase the attributes of each sunspots groups are
extracted and formed into a vector of N (7 x 1) based on the
Modified Zurich classification attributes. After extracting all
the attributes the vector is fed to the SVMs classifier [6], [5].
The features consist of a number of sunspots/umbra, size,
penumbra, penumbra’s position, bipolar, unipolar, and major
longitudinal axis. The seven classes of modified Zurich
classification are used as standard classes in this paper. For
further details see [2]. The classes are numbered as A, B, C,
D, E, F, and H. To make the classification more reliable the
class letters are converted to numbers from 1 to 7. Thus, the
class A is represented by one (1) and B by (2) and so on. The
attributes of each sunspot or sunspots groups are formed into
vector and fed to the SVMs. Multi-class SVMs are employed
in this paper.
Fig. 4. Black and white image produced from the preprocessing phase.
i=1,2…N, the ith SVM is trained with all the samples in the ith NS
(6)
class considered as positive examples and the samples from all w = ¦a i y i (x i )
other classes considered as negative examples. Given {(x1, y1), i =1
(x2, y2),…,(xk, yk)} as the training data set, where xj ∈ Rn are Where Ns is the number of support vectors and ai Lagragian
the n-dimensional samples (feature vectors) and yj ∈ multiplier for any training point [5].
{1,2,…N}are the corresponding class labels j=1,2,…,k, the ith
SV solves the following problem [6], [9]: Replacing the inner product between Φ in (5) with a function
simplifies things, where in kernel functions are introduced. In
Min
{1
k
w i ,b i , i
2
(w ) i T
w i
+ C ¦ ij } this work, Gaussian Radial Basis Functions (GRBF) are
j=1
(1) employed as kernel functions:
(w ) (x ) + b
i T
j
i
≥ 1 − ij , if y j = i § 2·
(7)
¨ − x −y ¸
(w ) (x ) + b
i T
j
i
≥ 1 + ij , if y j ≠ i K ( xi , x j ) =e
¨
©
i j ¸
¹, >0
ij ≥ 0, j = 1,........ .k The linear kernel is a special case of GRBF as Kreethi and Lin
have shown in [9]. Sigmoid kernels produce results generally
where 1/2 (Wi)T.Wi is the regularization term (objective comparable to GRBF kernels. However, GRBF has less
function), numerical complexities in comparison to other kernels [7].
GRBF has a control parameter, , which along with the cost
k
factor C are the two tunable parameters in the SVM
C ¦ j = 1
i
j
(2)
optimization problem. As suggested in [9], a useful way to
determine these is to perform grid-search with cross-
validation. A two-stage procedure can also be followed, where
is a cost term (training error) used to constrain the solution in after identifying a good region in the (C, ) space, a finer
cases of non-separable data ( are ‘slack’ variables introduced search can be conducted which may improve the results
for this purpose [5]), function is used to map the training ϕ
(w ) T
(x ) + b
The attributes are extracted from the images and
formatted into two text files. One of the text files is used as a
and x is said to belong to the class which gives the largest training resource and the other as a test resource. In the final
value for the decision function, i.e., testing stage, ten-fold cross validation is used to determine
good C and gamma parameter values, followed by actual
i=1,..., w (x) + b
class of x≡ argmax i i
(( )
T
) (4) classification using these values. The recognition rates
achieved are in the range of 80-88 % correct classification.
In practice the dual problem, formulated in terms of Lagrange
multipliers, is solved, i.e., maximize the follows VII. CONCLUSION
REFERENCES [8] Kreethi, S. S. and Lin C. J., ”Asymptotic behaviors of support vector
machines with gaussian kernel”, Neural Computation, Vol. 15 No.7,
pp. 1667-1689, 2003.
[1] T. Nguyen, C. P. Willis, D. J. Paddon,”Learning sunspot classification,“
[9] C-W. Hsu, C-C. Chang, and C-J. Lin, “A practical guide to support
Fundamental Informaticae XX, vol.72, 2006, pp. 295-309.
vector classification”, [Online]. Available:
[2] P. McIntosh,” Solar Physics,”, pp. 125-251, 1990.
[3] S. Zharkov, V. Zharkova, S. Ipson, A. Benkhalil,”Technique for
www.csie.ntu.edu.tw/~cjlin/papers/guide/guide.pdf.
automated recognition of sunspots on full disk solar,”, EURASIP [10] M. Abd and G. Paschos,” Effective arabic character recognition using
Journal on Applied Signal Processing, vol. 15 , pp. 2573–2584 2005. support vector machines,” In Proc.2007, Innovations and Advanced
[4] S. Majed, M.Abd and V.Zharkova,” Automated detection of sunspots techniques in Computer and Information Sciences and Engineering,
and sunspot groups in full-disk solar images,” ,Accepted in CISSE09 Springer, ISBN 978-14020-6267-4, 2007, PP. 7-11.
[5] Christopher J.C., ”A tutorial on support vector machines for pattern [11]. Solar Heliospheric Observatory (SOHO), (10-9- 2009),[Online].
recognition”, [Online] Available: Available: http://sohowww.nascom.nasa.gov/
http://aya.technion.ac.il/karniel/CMCC/SVM-tutorial.pdf. [12] M. Sezgin and B. Sankur, "Survey over image thresholding techniques
[6] Chih-Wei Hsu and Chih-Jen Lin, ”A comparison of methods for multi- and quantitative performance evaluation," Journal of Electronic Imaging.
class support vector machines”, IEEE Transactions on Neural Networks, Vol.13 (1), pp.146–165, 2004.
Vol. 13,No.2, pp. 415-425,March 2.002.
[7] Eugene A. Borovikov, ”An evaluation of support vector machines as a
pattern recognition tool”, University of Maryland at College Park,
2005., [Online]. Available:www.umiacs.umd.edu/users/yab/
SVMForPatternRecognition/report.pdf,
A Novel Ballistics Imaging System for Firearm
Identification
Dongguang Li
School of Computer and Security Science, Edith Cowan University, Australia
Email: [email protected]
327
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_57, © Springer Science+Business Media B.V. 2010
328 LI
the online system is convenient for end-user and more • “B” type characteristics were defined as broach series
independent on the client computer. characteristics, with B1 and B2 subclasses that occur on
The rest of the paper is organized as follows: Section 2 the edges of land impressions.
gives an introduction to firearm identification. In Section 3, Generally, Firearm identification is concerned with two
we illustrate the ballistics imaging system. An intelligent types of characteristics, i.e. class characteristics (C type)
cartridge case segmentation algorithm is described in and individual characteristics (A type). While broach series
Section 4. Section 5 concludes this paper. characteristic (B type) is only applied in particular types of
rifles. Class characteristics are useful in determining what
2. FIREARM IDENTIFICATION brand of firearm was used, but not helpful in identifying a
individual firearm. Individual characteristics can directly
Firearm identification aims to provide a link between the match a bullet to a firearm. That is, the match between
suspect firearm and the forensic ballistics specimens based individual characteristics and a bullet can ensure that one
on the depressions, scratches and markings on the certain firearm was the only firearm that could have fired
specimens of cartridge cases and projectiles. These various the bullet. These characteristics are detailed in Table 1.
marks are called ‘toolmark’ and it is thought that the
toolmark is unique to itself [5]. In this regard, it is possible 3. BALLISTICS IMAGING SYSTEM
to identify the suspect firearm by toolmark left on ballistics
specimens. Due to several drawbacks of workstation system including
Firearm identification mainly includes the following low data sharing and security, expensive cost of install and
stages [12]: maintenance, inconvenient application environment etc, we
• The crime scene weapon is examined and test fired for developed an Online Ballistics Imaging System (OBIS)
a test specimen; based upon Browse-Server (BS) structure. Fig. 1 shows the
• The characteristic marks on the cartridge cases and structure of the system. By a computer with internet access
projectiles are found and recorded; and a web browser, users can access OBIS anywhere and
• Features of the characteristic marks are investigated; anytime.
• Comparison of marks between the test specimen and
the crime scene specimen;
• Assessment of the comparison between the test and
crime scene specimens.
• Storage of the characteristic marks for further
comparisons.
cast of the peaks and ridges of striations. The high quality of 4. Repeat step 1, 2, and 3 until the whole surface of
shadow cast makes discernible features for the identification the projectile specimen is scanned.
process. According to Smith’s guideline, we test point 5. Combine all arrays in the buffer one after another
source illuminator and LED ring-light illuminator. Although so that a 2D flat image for the whole surface of
point source illuminator can produce better shadow cast, the the projectile is produced.
shadow cast of the image may vary with the rotational
position of the cartridge case about its central axis. While
the shadow cast of the image is invariant in LED ring-light
illuminator. Therefore, LED ring-light illuminator is
selected. Fig. 2 shows the high contrast image of head of
cartridge case using LED ring-light illuminator.
coded with respect to this position. While, for centrefire methods because of the heavy shadow in the input image.
cartridge case the ejector and extractor mark position are To acquire the satisfactory segmented image of cartridge
orientated as they would be in the firearm. These case, a novel approach based on color invariant properties
characteristics can be easily measured and saved by mouse [15] and geometrical shape of cartridge case is proposed.
click on the graphic user interface. The method is composed of two steps. That is, remove the
It is difficult to accurately measure the class background and shadows using color invariant properties;
characteristics on projectile due to its cylindrical shape. On detect and segment the circular cartridge case using Hough
the contrary, these characteristics can be easily measured on transform [16].
2D images acquired by line-scan imaging system. Fig. 5
shows that these characteristics can be measured by
calculating the difference of coordinates of pixels. For
example, “Land mark width” can be measured using the
following steps: firstly, draw a horizontal line on “Land
mark” in the image by mouse; then calculating the number
of pixels between the starting pixel and the ending pixel of
the line. Clearly, the product multiplying the number by the
length of each pixel is the value of length of “Land mark (a) (b)
width”.
(c) (d)
Abstract—Automatic seal imprint identification system is high- rotate C to make it have the same orientation as R, and com-
ly demanded in oriental countries. Even though several seal pare the corresponding details between them. If all of the de-
identification techniques have been proposed, it is seldom to tails are satisfactory to the technician’s subjective judgment,
find the papers on the recovery of lost seal imprint strokes the candidate seal imprint C is regarded as a genuine one.
caused by superimposition. Most identification methods are
conducted by matching the reference seal imprint and candi-
According to the process mentioned above, the hard-
date one pixel by pixel. In this paper, a robust color seal seg- matching consists of two steps, i.e. segment seal imprint
mentation and identification approach is proposed. This ap- from the input image, and then identify whether the seg-
proach segments the seal imprint from the input image in mented one is genuine or forgery. There are three main is-
terms of the adaptive thresholds. The lost seal imprint strokes sues on computerizing this procedure. Firstly, the input seal
are recovered based on the text stroke width that can be de- imprint is usually superimposed by other information such
tected automatically. In addition, the moment-based seal iden- as signature or background text, consequently, the segmen-
tification is to compare the reference seal imprint and the re- tation of seal imprint is difficulty[5]. Secondly, the differ-
covered one. Experimental results show that the proposed me- ence between the genuine seal imprint and the forgery one is
thod is able to correctly segment seal imprint and efficiently
discriminate the genuine and forgery seal imprint.
small, so the results are very sensitive to noise. Thirdly, it is
very difficult to make the candidate seal imprint have exact
Keywords-Seal segmentation, Seal identification, Image same orientation and size as the reference one.
binarization, Moment invariant Automatic seal segmentation have been studied in
many literatures [6-9]. These methods were mostly imple-
I. INTRODUCTION mented based upon RGB or HSV color space and can ac-
quire satisfactory results. However, none of the methods
Seals are widely used in oriental countries such as Chi- focuses on how to detect and recover losses of seal imprint
nese, Japanese and Korea to identify a person, a social strokes in segmented seals. The losses of seal imprint
group, or an organization etc[1, 2]. Currently, a variety of strokes often exist in segmented seal imprint due to supe-
seals must be manually registered, which relies on the exper- rimposition among the seal, signature and background text.
tise and experience of professional technicians. The process The result of seal identification may be strongly affected by
of seal identification not only consumes so much time, but losses of seal imprint strokes. As a consequence, it is of
also inevitably introduces some misidentification. Therefore, great importance in automatic seal imprint identification.
a seal imprint verification system is highly demanded to re- Several seal identification methods have also been pro-
solve this issue automatically, speedily and reliably. posed for decades [10-14]. These methods mainly include
Seals can be made of various materials e.g. wood, stone, four steps: 1) Rotate the candidate seal imprint to the same
marble, plastics, ivory, crystal, metal. The shape and size of orientation as the reference one; 2) Compute the centroid to
seals are also different. Basically, there are three shapes, i.e., the candidate imprint and reference one; 3) Overlap the ro-
circular, rectangular, and elliptic[3]. Generally, the color of tated candidate seal imprint and the reference one in terms
seal imprint is pure red or pure blue. Seal identification can be of the centroid; 4) Match them pixel by pixel. From these
regarded as a template match between the reference seal im- steps, we can see that the performance and robustness of
print and the candidate one. To exactly match the seals, a seal identification strongly depend upon the precision to
variety of features have to be considered, including the cha- detect the angle of rotation and position of centroid. Owing
racters in the imprint, the scales, distributions, relative spatial to the difficulty in exactly overlapping the input seal imprint
positions, strokes etc. As a result, the matching is called and reference one, the method matching candidate seal im-
“hard-matching” [4]. Currently, hard-matching is usually print and reference one pixel by pixel is difficult to use.
performed by human visual inspection. To describe the Clearly, the exact detection of rotated angle is very compli-
process of seal identification in detail, we define two va- cated and difficult.
riables R (Reference seal imprint) and C (Candidate seal im- In this paper, a robust approach for seal imprint seg-
print). To identify whether C is genuine or forgery, a profes- mentation and identification is proposed. To improve the
sional technician firstly overlap the centroid of R and C, then precision and robustness to segment seal imprint, an adap-
333
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_58, © Springer Science+Business Media B.V. 2010
334 Li
tive thresholding method based on RGB components is pre- seal imprint, signature and background. Seal imprint is ex-
sented. This segmentation method is very suitable for both tracted based on the cluster index. Since K-means method
seal segmentation and extracting a particular color cluster divides the pixels in terms of Euclidean distance, the com-
from an image. The recoveries for the losses of seal imprint pletely different color may be divided into a same cluster.
strokes are considered for minimizing the errors in seal im- For example, the color (R=255, G=0, B=0) and (R=0,
print identification. A moment-based method is employed G=255, B=0) are red and green, but they have the same
because it is independent on the image translation, rotation Euclidean distance to black (R=0, G=0, G=0). This method
and scaling. As a result, the proposed method not only may also get incorrect result when noise is strong. Liang[6]
avoids the complicated match processing but also improves divided the RGB space into eight subspaces corresponding
the robustness and precision to identify seal imprints. The to color index 0, 1, 2 ,…, 7 by dichotomy, and then compute
flowchart of the proposed method is shown in Fig. 1. It con- the color index for each pixels in input image in terms of
sists of five major steps: 1) Segment the seal imprint in equation (2.1), where “[]” represents integer Round opera-
terms of the adaptive thresholds; 2) Recover the losses of tor, r, g, b represent respectively red, green and blue com-
seal imprint strokes in segmented seal imprint; 3) Binarize ponents in color normalized image. Finally, extract the seal
the segmented seal imprint and the reference one using Otsu imprint by index (e.g. the index of red is 1). This method
algorithm[15]; 4) Compute the Hu’s moment invariants[16] may also be easily attacked by noise.
to the segmented seal imprint and the reference one; 5) Index= [b/128] × 4 + [g/128] ×2 +[r/128] (2.1)
Identify them based on Euclidean distance of moment inva- Shih-Hsu [12] proposed a register method in terms of
riants. the feature points. This method includes four major steps: 1)
Feature points extraction; 2) Estimate the rotation and trans-
lation parameters using the point matching technique; 3)
Seal registration using the affine transformation by the esti-
mated parameters; and 4) Fake detection from the registered
seals. However, the seal imprints with different orientation
stamped by the same seal may generate different feature
points that may strongly affect the stability of this algo-
rithm. To rotate the input seal imprint and the reference one
to the same orientation, Yung-Sheng[10] used the contour
analysis to find the principal orientation of a seal imprint,
and then rotate the input seal imprint to the same orientation
with the reference one. The algorithm may misidentify if the
seal imprint does not exist principal orientation such as the
Fig. 1. Seal imprint identification system seal imprint displayed in Fig. 2(b).
In this paper, we employ the Otsu algorithm [15] for
The rest of the paper is organized as follows: Section 2 the adaptive seal segmentation. Otsu is a typical and popular
gives an introduction to related works. In Section 3, we illu- image binary algorithm based on global threshold. This me-
strate the proposed approach. The experimental results are thod divides input data into two classes by threshold. The
presented and analyzed in Section 4. Section 5 concludes basic idea of Otsu is to find an optimal threshold that gives
this paper. the best separation between the two classes in terms of their
values. The threshold can be estimated automatically on an
II. RELATED WORK iterative scheme, in order to find the maximum variance
A variety of seal identification techniques have been between the two classes. It works well in situations where
proposed. In general, the studies mainly focus on the seal there is a reasonably clear valley between the modes of the
segmentation and seal registration. Seal segmentation deals histogram related to the two classes.
with how to exactly extract seal imprint from the input im- Hu’s moment invariants[16] are very effective on identi-
age such as bank check image. Seal registration is to identi- fying the input seal imprint. Moment invariants have been
widely applied to image pattern recognition in a variety of
fy that the input seal imprint is genuine or forgery by com-
applications due to its invariant features on image transla-
paring the input and the reference seal imprint. tion, scaling and rotation. Two-dimensional (p+q)th order
Katsuhiko[9] extracted the signature and seal imprint moments are defined in equation (2.2).
by projecting orthogonally all pixels of a bank check image ∞ ∞
on to an appropriate axis in RGB color space[17]. This me- m pq = ³ ³ x p y q f ( x, y )dxdy
thod may fail if the center of three clusters is all in or near a −∞ −∞ (2.2)
straight line. In [14], the input image is first transformed p, q = 0,1,2,
into HSV (Hue, Saturation, and Intensity-Value ) perceptual To achieve the independence for image translation and
space[18], and then K-means algorithm[19] is employed to rotation, the equation (2.2) is transformed into central mo-
divide the input image into three clusters corresponding to ment shown in equation (2.3)
COLOR SEAL SEGMENTATION AND IDENTIFICATION 335
∞ ∞ A. Seal Segmentation
μ pq = ³ ³ ( x − x ) p ( y − y ) q f ( x, y)dxdy
(2.3)
−∞ −∞ Seal segmentation aims to extract the seal imprint from
p, q = 0,1,2, the input image and recover the losses of seal imprint
strokes caused by superimposition. It mainly contains the
where x =m10/m00 and y =m01/m00, which are the cen- following steps: 1) Estimate the thresholds to segment the
seal imprint, 2) Segment the seal imprint and text from the
troid of the image. Scale invariance can be obtained by nor-
malization. The normalized central moments are defined in input image, 3) Recover the losses of seal imprint strokes, 4)
equation (2.4) Binarize the segmented seal imprint and reference one.
μ pq 1. Estimate Thresholds
η pq = , γ = ( p + q + 2) / 2, p + q = 2,3, (2.4)
μ 00γ In our demonstration, the color of seal imprint is red,
i.e. the red component of seal imprint is stronger than
Based on normalized central moments, Hu[16] intro- green and blue components. For three color clusters, i.e.
duced seven moment invariants: red (seal imprint), white (background) and black (text), the
R (Red), G (Green) and B (Blue) components are equal for
ϕ1 = η 20 − η 02 white(R=G=B=255) and black(R=G=B=0), The red seal
φ2 = (η 20 −η 02 ) 2 + 4η112 imprint can be easily segmented under the conditions R>G
and R>B. Fig. 3 shows the segmented result only based on
φ3 = (η 30 − 3η12 ) 2 + (3η 21 − μ 03 ) 2 the conditions R>G and R>B. The result shows that so
φ4 = (η 30 + η12 ) 2 + (η 21 + μ 03 ) 2 many pixels belonging to background or text are also meet
the conditions in practice because of various noisy factors
φ5 = (η30 − 3η12)(η30 +η12 )[(η30 +η12)2 − 3(η21 +η03)2 ] such as camera, illumination.
+ (3η21 −η03)(η21 +η03) [3(η30 +η12 )2 − (η21 +η03)2 ]
φ6 = (η 20 −η 02 )[(η 30 +η 12 ) 2 - (η 21+η 03 ) 2 ]
+ 4η 11(η 30 +η 12 )(η 21+η 03 )
φ7 = (3η21 −η 03)(η 30+η12)[(η 30+η12) 2 - 3(η 21+η 03) 2 ]
- (η 30-3η12)(η 21+η 03)[(3(η 30+η12) 2 - (η 21+η 03) 2 ]
This seven moment invariants are useful properties of be-
ing unchanged under image scaling, translation and rotation. Fig. 3. Segmented Seal with Noise
III. PROPOSED APPROACH
In general, the shapes of seal may be circular, rectangu-
lar, or elliptic. The colors of seal imprint may be red or blue.
For simplicity, only circular seal and red seal imprint are
selected in our demonstration. The genuine seal imprint is
shown in Fig. 2.(a). Obviously, the image is composed of
three color clusters, i.e. the red seal imprint, the black text
and the white background. Fig. 2.(b) displays the reference
seal imprint in which only includes white background and
red seal imprint. The forgery seal imprint is displayed in Fig.
2.(c).
Fig. 4. Histograms for the values of R/G and R/B
We can found that the ratios of R/G and R/B for noise
pixels are far less than the ratios of seal imprint pixels. It is
possible to find the thresholds of R/G and R/B to divide the
pixels which meet the conditions (R>G and R>B) into two
classes, i.e. the noise pixels and the seal imprint pixels. To
(a) Genuine seal imprint (b) Reference seal imprint (c) Forgery seal imprint
find the optimal thresholds for R/G and R/B, the histo-
Fig. 2. The input images for demonstration
grams for the ratios of R/G and R/B on pixels whose R>G
and R>B are constructed, as shown in Fig. 4. Since both
336 Li
histograms of R/G and R/B have evident bimodality, the 3. Recover the losses of seal imprint strokes
optimal thresholds can be acquired using Otsu algorithm
[15] detailed in Section 2. For Fig. 2(a), the optimal thre- Since the brightness and color vary in superimposition
sholds of R/G and R/B calculated by Otsu are respectively of text and seal imprint inks, the character strokes of seal
2.1 and 1.8. imprint in superimposition may lose in segmented seal im-
print. This may cause incorrect result in seal imprint regis-
2. Segment Seal Imprint and Text tration process. To avoid this, a method for detecting and
After acquiring the thresholds, the seal imprint can be recovering losses of seal imprint strokes is proposed in this
segmented easily. Since the text in input image will be section. Comparing the Fig. 7(a) and Fig. 5(b), we can see
used in subsequent process, the seal imprint and text are that the losses of seal imprint strokes basically take place in
segmented in the same time. The steps of segmentation are superimposition between the seal imprint and text. Further-
detailed in algorithm 1. more, the width of lost stokes is very close to the width of
stokes corresponding text. Accordingly, we apply the fol-
Algorithm 1 Segmenting Seal Imprint and Text from lowing steps to recover the lost strokes of seal imprint.
the Input Image • Estimate the width of the strokes of black text;
Input: an image with red seal imprint, black text and • Recover the lost strokes in seal imprint image in
white background terms of the width of the strokes of black text.
Output: an image of red seal imprint, an image of To estimate the stroke width of black text, we scan
black text each row of black text image (shown in Fig. 5(b)) to calcu-
late the number of each continuous black pixels, and then
i: Estimate the threshold THRG and THRB using method construct the stroke histogram using all of the number of
described above; continuous black pixels. The position of maximum in histo-
ii: Scan input image, get values of the red, green and gram is considered as the stoke width of black text. For the
blue component in each pixel and save in variable image shown in Fig. 5(b), the histogram is displayed in Fig.
r, g and b respectively; 6. Evidently, the stroke width of black text is 4 pixels.
iii: if r/g> THRG and r/b> THRB
The pixel is belong to seal imprint Iseal
Else
The pixel is belong to background or text IBT
End if;
iv: Binarize the pixels in IBT using Otsu algorithm [15]
and then get the text image Itext;
v: Remove the black block in Itext whose size less than
5 pixels;
vi: Return the seal imprint Iseal and black text Itext.
Fig. 6. Histogram of stroke width on black text
The segmented seal imprint is displayed in Fig. 7(a).
And the black text with noise is shown in Fig. 5(a). To re- After getting the stroke width of black text Wstroke, we
move the noise, the size of each black block is computed. can recover the segmented seal imprint using the following
And then remove the small block whose size less than 5 procedures. Scan the black text image shown in Fig. 5(b).
pixels. After this operation, a clear text image is acquired For each black pixel, detect its Wstroke hWstroke neighbor
and showed in Fig. 5(b). pixels, if the pixels located at up and down position or left
and right position of the black pixel are all red, the white
pixel with the same position as the black text image in the
seal imprint image (shown in Fig. 7(a)) is converted into
black. The recovered seal imprint is displayed in Fig. 7(b).
Fig. 5. Images of black text (a) Seal imprint without strokes recovery (b) Recovered seal imprint
B. Seal Registration D are circular seals (Type 1), Seals E to H are rectangular
Normally, the registration includes two steps: 1) Ac- seals (Type 2), and Seal I to L are elliptic seals (Type 3).
cord the orientation and size of candidate seal imprint with Each type includes two genuine seals and two forgery seals.
the reference one; 2) Overlap the candidate seal imprint and All seals are engraved by one maker, therefore the differ-
the reference one in terms of their centroid, and then match ence between the genuine seal and forgery one are very
them pixel by pixel. Since exact adjustment on the orienta- small. The experimental steps are describes as follows.
tion of candidate seal imprint is very difficult or even im- • Stamp each genuine seal on white paper, and convert it
possible, especially to circular seal, the precision of registra- into a digital image by CANON IXUS 90 IS camera as
tion may be strongly affected by adjusted precision. We a reference seal imprint. Then transform them into bi-
propose a novel moment-based approach to address the dif- nary images and calculate the moment invariants on the
ficulty. The moment invariants are independent of transla- binary images;
tion, rotation and scaling, so the complicated adjustment in • Arbitrarily stamp each of twelve seals on five different
orientation and size can be avoided. documents. And then convert them into digital images
by CANON IXUS 90 IS camera. Thus, sixty candidate
seal imprints are applied to examine the proposed algo-
rithm;
• Segment each seal imprint from input images with seal
imprint and recover losses of seal imprint strokes. Then,
transform them into binary images;
• Calculate the moment invariants on each image from
(a) Genuine seal imprint (b) Reference seal imprint (c) Forgery seal imprint last step.
Fig. 8. Binarized seal imprint • Calculate the Euclidean distance between the reference
seal imprint and corresponding candidate seal imprints.
To minimize the effect of luminance, we transform all • Make sure whether the candidate seal imprint is ge-
seal imprints into binary images with black background and nuine or not in terms of the selected threshold.
white seal imprint. The sample seal imprints are displayed
in Fig. 8. Seal registration is achieved by computing the
seven moment invariants on all seal imprints and Euclidean
distance (shown in equation (3.1)) based on these invariants.
7
Ci − Ri
D=¦ (3.1) (A) Genuine (B) Forgery (C) Genuine (D) Forgery
i =1 Ri
where Ci, Ri respectively present the seven moment in-
variants of candidate seal imprint and reference one. Table 1
displays the results of comparison between the reference
seal imprint and genuine or forgery one. Although it is very
difficult to manually distinguish the reference seal imprint (E) Genuine (F) Forgery (G) Genuine (H) Forgery
from the forgery one (displayed in Fig. 8), the Euclidean
distance between the reference seal imprint and forgery one
is evidently greater than the values between the reference
seal imprint and genuine one. Consequently, the forgery seal
imprint can be easily identified by the Euclidean distance. (I) Genuine (J) Forgery (K) Genuine (L) Forgery
Fig. 9. Twelve seals for experiments
Table 1. Comparison of moment invariants for refer-
ence, genuine and forgery seal imprint Fig. 10 shows the results of recovered seal imprints
segmented from the input image ( only Seal A and B are
displayed). Although the orientation of seal imprints are
different from each other, which not affects the results of
identification due to the moment invariants are independent
to rotation. The experimental results are displayed in Table
IV. EXPERIMENTAL RESULTS 2. From this table, we can see that the Euclidean distance
To evaluate the performance of the proposed seal seg- between genuine and reference seal imprints are all less than
mentation and identification algorithm, extensive experi- 1, and the Euclidean distance between the reference and
ments are conducted based on sample images. Twelve seals forgery seal imprints are all greater than 1.8. In real applica-
are used in the experiments (displayed in Fig. 9). Seals A to tion, the errors to identify a forgery seal imprint as a ge-
338 Li
nuine one is not acceptant. Therefore, in our experience, we Euclidean distance to identify seal imprints. In addition,
select 1 as the threshold to identify whether the seal imprint more experiments will be conducted to check whether the
is genuine or forgery. approach is able to be applied to more extensive conditions.
VI. ACKNOWLEDGEMENT
Mr Z. Huang, as a visiting fellow at ECU, helped with
digital image preparation.
REFERENCES
[1] H. Leung, "Analysis of traditional Chinese seals and synthesis of
personalized seals," in Multimedia and Expo, 2004. ICME '04. 2004
IEEE International Conference on, 2004, pp. 1283-1286 Vol. 2.
[2] J. J. Brault and R. Plamondon, "Segmentation handwritten signatures
at their perceptually important points," IEEE Trans. Pattern. Anal.
Fig. 10. Five seal imprints of A (row 1) and five seal im- Mach. Intell., vol. 15, pp. 953-957, 1993.
prints of B (row 2) [3] Chou Ju-Pei and Chen Hu-Sheng, "Chinese seal identification," in
Security Technology, 1990. Crime Countermeasures, Proceedings.
IEEE 1990 International Carnahan Conference on, 1990, pp. 51-53.
Table 2. Comparison of moment invariants for twelve seals [4] Yung-Sheng Chen, "Computer Processing on the Identification of a
Chinese Seal Image," in Third International Conference on Docu-
ment Analysis and Recognition (ICDAR'95), 1995, pp. 422-425.
[5] Yung-Sheng Chen, "Approach to the Chinese Seal Registration," in
Advances in Communication Systems and Electrical Engineering.
vol. 4, 2008, pp. 529-542.
[6] Liang Cai and Li Mei, "A Robust Registration and Detection Method
for Color Seal Verification," in Advances in Intelligent Computing.
vol. 3644, 2005, pp. 97-106.
[7] Ching-Liang Su, "Chinese-Seal-Print Recognition by Color Image
Our Proposed approach has been implemented using Dilating, Extraction, and Gray Scale Image Geometry Comparison,"
Visual Basic .NET, and running on a PC (CPU P4 2.0G, Journal of Intelligent and Robotic Systems vol. 44, pp. 349-359,
2005.
RAM 1G). The processing time to segment and recover the [8] Aureli Soria-Frisch, "Color Seal Extraction from Documents: Ro-
seal imprint is about 0.9 second for 500*500 color image. bustness through Soft Data Fusion," EURASIP Journal on Applied
And the processing speed for calculating the moment inva- Signal Processing, vol. 2005:13, pp. 2146-2152, 2005.
riants is about 0.8 second. As a result, the whole computa- [9] K. Ueda, "Extraction of signature and seal imprint from bankchecks
by using color information," in Document Analysis and Recognition,
tional time is less than 2 seconds. 1995., Proceedings of the Third International Conference on, 1995,
pp. 665-668 vol. 2.
V. CONCLUSION [10] Yung-Sheng Chen, "Registration of Seal Images Using Contour
In this paper, a robust segmentation and identification Analysis " in Lecture Notes in Computer Science. vol. 2749: Springer
Berlin / Heidelberg, 2003, pp. 255-261.
algorithm for color seal imprint is proposed. This method is [11] Ching-Liang Su, "Edge distance and gray level extraction and orien-
composed of segmentation of seal imprint, recovery of tation invariant transform for Chinese seal recognition," Applied
losses of seal imprint strokes, and identification of seal im- Mathematics and Computation, vol. 193, pp. 325-334, 2007.
print. The segmentation automatically extract seal imprint [12] Shih-Hsu Chang, Fang-Hsuan Cheng, et al., "Automatic seal identifi-
cation by using point matching and fake detection," Pattern
from the color input image in terms of the adaptive thre- Recognition Letters, vol. 20, pp. 1545-1552, 1999.
sholds. To minimize the errors, the losses of seal imprint [13] T . Horiuchi, "Automatic seal verification by evaluating positive
strokes are recovered by detecting the width of black text cost," in Document Analysis and Recognition, 2001. Proceed-
stroke and superimposition pixels between the red seal im- ings. Sixth International Conference on, 2001, pp. 572-576.
[14] K . Ueda, T. M utoh, et al., "Automatic verification system for seal
print strokes and the black text strokes. To avoid compli- imprints on Japanese bankchecks," in Pattern Recognition,
cated adjustment on orientation and size between the refer- 1998. Proceedings. Fourteenth International Conference on,
ence seal imprint and candidate one, the moment-based 1998, pp. 629-632 vol. 1.
technique is applied to seal imprint identification. [15] N . Otsu, "Threshold Selection Method from Gray-Level Histo-
grams," IEEE Transcations on Systems Man and Cybernetics,
Experimental results showed that the proposed seg- vol. 9, pp. 62-66, 1979.
mentation method is able to segment seal imprint from color [16] H u Ming-Kuei, "Visual pattern recognition by moment invariants,"
input image, and the losses of seal imprint strokes can be Information Theory, IRE Transactions, vol. 8, pp. 179-187, 1962.
correctly recovered. Moreover, the identification method [17] Rafael C. Gonzalez and Richard E. Woods, Digital Image
has high performance in exactly discriminating the genuine Processing (Third Edition): Prentice Hall, 2007.
[18] AR . . Smith, " Color gamut Transformation Pairs," Computers
and forgery seal imprint. The computational time is less Graphics, vol. 12, pp. 12-19, 1978.
than 2 seconds on general PC. Therefore, the proposed ap- [19] Y. Linde, A.Buzo, et al., "An algorithm for vector quantizer design,"
proach is fast and feasible. In future, we will find more fea- IEEE Trans. on Communications, vol. 28, pp. 84-95, 1980.
tures on seal imprints, and employ neural network instead of
Efficient SOPC-based Multicore System Design
using NOC
Vignesh Veerapandian, Xingguo Xiong
Department of Electrical and Computer Engineering, University of Bridgeport, Bridgeport, CT 06604, USA
Email: [email protected], [email protected]
Abstract — Due to the advancement of VLSI (Very Large Scale As a result‚ around 1999 several research groups have
Integrated Circuits) technologies, we can put more cores on a started to investigate systematic approaches to the design of
chip, resulting in the emergence of a multicore embedded the communication part of SOCs. It soon turned out that the
system. This also brings great challenges to the traditional
parallel processing as to how we can improve the performance
of the system with increased number of cores. In this paper, we
meet the new challenges using a novel approach. Specifically,
Problem has to be addressed at all levels from the physical
we propose a SOPC (System on a Programmable Chip) design to the architectural to the operating system and application
based on multicore embedded system. Under our proposed level. Hence‚ the term Network on Chip (NOC) is today
scheme, in addition to conventional processor cores, we used mostly in a very broad meaning‚ encompassing the
introduce dynamically reconfigurable accelerator cores to hardware communication infra-structure‚ the middleware
boost the performance of the system. We have built the and operating system communication services and a design
prototype of the system using FPGAs (Field-Programmable methodology and tools to map applications onto a NOC. All
Gate Arrays). Simulation results demonstrate significant this together can be called a NOC platform. [4] Networks on
system efficiency of the proposed system in terms of
Chip (NOCs) have emerged as a viable option for designing
computation and power consumption. Our approach is to
develop a highly flexible and scalable network design that
scalable communication architectures. For multiprocessor
easily accommodates the various needs. This paper presents System-on-Chips (MPSOCs), on-chip micro networks are
the design of our NOC (Network on Chip) which is a part of used to interconnect the various cores. The main idea with
the platform that we are developing for a reconfigurable NOCs, besides the solutions to the physical issues, is the
system. The major drawback of SOPC based systems lies in the possibility for more cores to communicate simultaneously,
routing of the various on-chip cores. Since it is technically leading to larger on-chip bandwidths. The adoption of NOC
difficult to integrate more than one core on a single chip, we architecture is driven by several forces: from a physical
come across several routing problems which lead to inefficient design viewpoint, in nanometer CMOS technology
functioning. Thus we implemented several NOC based routing
interconnects dominate both performance and dynamic
algorithms which considerably improve accessing speed and
enhance the system efficiency.
power dissipation, as signal propagation in wires across the
chip requires 2 multiple clock cycles. NOC links can reduce
Keywords: Multicore system, System on a Programmable the complexity of designing wires for predictable speed,
Chip (SOPC), Network on Chip (NOC), Multiprocessor power, noise, reliability, etc., thanks to their regular, well
System-on-Chip (MPSOC). controlled structure. From a system design viewpoint, with
the advent of multi-core processor systems, a network is a
I. INTRODUCTION natural architectural choice.
During the 1990s more and more processor cores and large NOC can provide separation between computation and
reusable components have been integrated on a single communication; support modularity and IP reuse via
silicon die‚ which has become known under the label of standard interfaces; handle Synchronization issues; serve as
System-on-Chip (SOC). Buses and point-to-point a platform for system test and hence increase engineering
connections were the main means to connect the productivity.
components, Hence they can be used very cost efficiently.
As silicon technology advances further‚ several problems
related to buses have appeared. [1][5] Buses can efficiently II. DIFFERENTNOCTOPOLOGIES
connect 3-10 communication partners but they do not scale
to higher numbers. The Network-on-Chip (NOC) architecture, as outlined in
Figure 1, provides the communication infrastructure for the
resources. In this way it is possible to develop the hardware
of resources independently as stand-alone blocks and create
the NOC by connecting the blocks as elements in the
network.
339
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_59, © Springer Science+Business Media B.V. 2010
340 VEERAPANDIAN AND XIONG
.
Figure.1. Network on chip [7]
A number of different NOC topologies have been proposed. Figure.2. A honey comb structure for NOC [7]
They all have in common that they connect resources to
each other through networks and that information is sent as 2.2 Mesh Topology
packets over the networks [7]. Network on Chip (NOC) has
evolved as an important research topic during the last few NOC is a scalable packet switched communication platform
years. The idea is that scalable switched networks are used for single chip systems. The NOC architecture consists of a
for on-chip communication among processing units, in order mesh of switches together with some resources which are
to cope with design of continuously growing systems. placed on slots formed by the switches.[2] Figure 3 shows
Design complexity promotes reuse of investment in earlier NOC architecture with 16 resources. Each switch is
designs as well as purchase of outside intellectual property connected to four neighboring switches and one resource.
(IP). However, in larger designs, communication among Resources are heterogeneous. A resource can be a processor
components will become a bottleneck using traditional core, a memory block, a FPGA, custom hardware block or
techniques like common buses. NOC is one solution to any other intellectual property (IP) block, which fits into the
address this issue because packet switched communication available slot and complies with the interface with the NOC
can provide higher flexibility, throughput and reusability. switch. We assume switches in NOC have buffers to
To gain full advantage when using this concept in NOC manage data traffic.
architecture design, the size of resources should be similar
and the communication facilities should be homogeneous.
Every resource has a unique address and is connected to a Figure.4 shows an FPGA design that includes an SOPC
switch in the network via a resource network interface Builder system and custom logic modules. We can integrate
(RNI). The NOC platform defines four protocol layers: the custom logic inside or outside the SOPC Builder system. In
physical layer, the data link layer, the network layer, and the this example, the custom component inside the SOPC
transport layer. The RNI implements all the four layers, Builder system communicates with other modules through
whereas every switch to switch interface implements the an Avalon-MM master interface. The custom logic outside
three of four layers except physical layer. The NOC of the SOPC Builder system is connected to the SOPC
architecture also has a concept of region allows us to handle Builder system through a PIO interface. The SOPC Builder
physically larger resources and can be used to provide fault system includes two SOPC Builder components with
tolerance. A typical NOC architecture will provide a Avalon-ST source and sinks interfaces. The system
scalable communication infrastructure for interconnecting interconnect fabric shown below in Figure.5 connects all of
cores. The area of multi-media is a very suitable candidate the SOPC Builder components using the Avalon-MM or
for using this high computing capacity of NOCs. NOC is a Avalon-ST system interconnects as appropriate.
general paradigm and one needs to specialize a NOC based
architecture for every application area.
Figure.6. Chip planner tool bar from Quartus II Software 3.6. Delay Calculation
We can view the timing delays for the highlighted
With the Chip Planner, we can view critical path connections when generating connections between
information, physical timing estimates, and routing elements. For example, you can view the delay between two
congestion. We can also perform assignment changes with logic resources or between a logic resource and a routing
the Chip Planner, such as creating and deleting resource resource.
assignments, and post-compilation changes like creating,
moving, and deleting logic cells and I/O atoms. By using the 3.7. Viewing Assignments in the Chip Planner
Chip Planner in conjunction with the Resource Property
Editor, we can change connections between resources and
make post-compilation changes to the properties of logic Location assignments can be viewed by selecting the
cells, I/O elements, PLLs, and RAM and digital signal appropriate layer set from the tool. To view location
processing (DSP) blocks. With the Chip Planner, we can assignments in the Chip Planner, select the Floor plan
view and create assignments for a design floor plan, perform Editing (Assignment) task or any custom task with
power and design analyses, and implement ECOs in a single Assignment editing mode. The Chip Planner shows
tool. location assignments graphically, by displaying assigned
resources in a particular color (gray, by default). We can
3.2. Viewing Routing Congestion create or move an assignment by dragging the selected
resource to a new location.
The Routing Congestion view allows us to determine the
percentage of routing resources used after a compilation. IV. RESULTS
This feature identifies where there is a lack of routing
resources. This information helps us to make decisions Using SOPC Builder in Quartus II tool, we designed and
about design changes that might be necessary to ease the simulated efficient SOPC-based Multicore System, and the
routing congestion and thus meet design requirements. The results are listed in Figure 7-11.
congestion is visually represented by the color and shading Figure 7 shows the screenshot of the developed SOPC
of logic resources. The darker shading represents greater builder system. It is done using SOPC builder, this build
routing resource utilization. We can set a routing congestion system has a design of multicore system with 2 CPU’s,
threshold to identify areas of high routing congestion. After Avalon tri state bridge, flash memory, LCD and PIO’s. It’s
selecting the Routing Utilization layer setting, click on the a FPGA design which includes SOPC builder system and
Routing Congestion icon on the taskbar. custom logic modules. We can integrate custom logic inside
or outside the SOPC builder system. In this design the
3.3. Viewing I/O Banks custom logic modules inside the SOPC builder system
communicates with other modules through an Avalon-MM-
The Chip Planner can show all of the I/O banks of the master interface. The custom logic modules outside of the
device. To see the I/O bank map of the device, click the
EFFICIENT SOPC-BASED MULTICORE SYSTEM DESIGN USING NOC 343
SOPC builder system is connected to SOPC system through designed system has no errors so in order to test the system
a PIO interface. we compile our system. It shows 100% full compilation
during synthesize of our SOPC system.
Figure.9. Compilation
Figure.7. SOPC builder for system building
The chip planner view of the compiled design is shown in
The block diagram of the symmetric file is shown in Figure Figure 10. In this screenshot we can see the build
8. SOPC builder allows us to design the structure of a components placed in this chip planner. We can view
hardware system. The GUI allows adding components to a critical path information, physical timing estimation and
system configure the components and specify the routing congestion. The Chip Planner uses a hierarchical
connectivity. After adding and parameterize components, zoom viewer that shows various abstraction levels of the
SOPC Builder generates the system interconnect fabric, targeted Altera device. As we increase the zoom level, the
outputs HDL files and .BDF during system generation. level of abstraction decreases, thus revealing more detail
This .BDF file shown in Figure8 represents the top –level about our design.
SOPC system for use in Quartus II.
The routing utilization is shown in Figure 11. It allows us to bandwidth and reduces routing congestion so that the
determine the percentage of routing resources used after system efficiency is enhanced. Our designed multicore
compilation. This information helps us to make decisions system has two CPU cores which are individually optimized
about design changes which is necessary to ease the routing to the particular computational characteristics of different
congestion to meet the design requirements. The routing application fields, complementing each other to deliver high
congestion is visually represented by the color and shading performance levels with high flexibility at reduced cost.
of logic resources. In Figure 11 we can see some areas are The research focus is shifting from implementation of NOC
dark and some areas are bright. The dark regions represent to investigation of its optimal use. The research problems in
greater routing resource utilization and the bright regions NOC design are identified as synthesis of communication
represent no routing congestion. infrastructure, choice of communication paradigm,
application mapping and optimization. In the future, we will
continue to design an efficient multicore SOPC with
optimized timing constraints, reduced latency and improved
programmability. We will also develop highly embedded,
multi-core systems with more number of cores which in turn
increases the system performance and many applications
can run at the same time.
REFERENCES
Tampere University of Technology, Department of Automation Science and Engineering, P.O.BOX 692, FI 33101
Tampere, Finland Tel: 03-31154203, Fax 03-31152340, email: [email protected]
Abstract. This paper presents OPCUA binary protocol integration of function block based basic automation
from telecommunication point of view by using OSI- information. OPC UA focuses on providing support for
network model as a basis. Presented architecture is based complex data models and flexible address space. OPC UA
directly on OPCUA specifications and Java reference abandons inherent need for Windows operating system and
implementation. Practical interest is focused on OPCUA tight coupling with Windows security settings present in
security and its feasibility for real world applications. The OPC/DCOM. The somewhat complex programming with
impact of enabling security in OPCUA is measured by COM technology is also replaced with modern software
benchmarking performance of current Java default development approach. OPC UA also increases the number of
implementation. possible types of integration solutions for which OPC can be
used by basing on standard technologies to implement Service
Oriented Architecture (SOA), namely web services (WS).
I. INTRODUCTION Even though OPC UA has significantly evolved from
OPC/DCOM in both data model and implementation
architecture backward compatibility has not been neglected.
Integration of industrial information systems has become a
OPC Foundation has realized that clear and working way to
common means to increase the overall efficiency of
emigrate from legacy OPC/DCOM is requirement for OPC
production. Information system integration is even exceeding
UA success. This transformation can be accomplished by
plant and company borders forming a collaborative network
using wrapper technology for which OPC Foundation
between businesses. The current trend of increasing demands
provides sample solutions.
posed for integration calls for unified base on which the
The goal of this paper is to present structure of OPC UA
solutions are built on. This significantly reduces the cost and
communication stack as it has been implemented in JAVA
time of integration in the long term.
reference implementation. Java implementation is closely
One widely adopted solution to evolving needs of industrial
related but not identical to that of .NET/C# implementation
integration is OPC Foundation specifications. The term OPC
developed by OPC Foundation. The performance of OPC UA
was originally acronym of concept OLE (Microsoft Object
protocol stack is also studied based on measurements done
Linking and Embedding) for Process Control. However as
with released version of the Java stack. Migration path to
OLE is no longer relevant technology for OPC UA the
OPC UA is further studied in papers [1] and in book [3] pages
interpretation has been changed to Openness, Productivity
283-293. Capabilities of OPC UA in low level device
and Collaboration. Over the years OPC has become the de
integration are studied in references [4] and [5].
facto standard for industrial integration and process
information sharing.
The original OPC was based on Microsoft DCOM
(Distributed Component Object Model) which is no longer II. BACKGROUND
actively developed by Microsoft. Traditional OPC is also
facing integration needs that it can no longer fulfill due static This paper is based on results and expertise obtained in the
information and communication models based on. These OPC Unified Architecture Java reference implementation
reasons have leaded the OPC Foundation to reinvent OPC (OPCUA Java, 2006-2007) and OPC UA Information model
replacing the core components and technologies with modern, (OPCUA Java 2, 2008 - 2009) projects. The goals of these
vendor independent solutions. Backward compatibility has projects were to contribute to the OPC Foundation’s Early
not been forgotten and OPC UA profiles detail use cases Adopter Program, develop an OPC UA Java reference
similar to those of legacy OPC systems. The new implementation and assess the feasibility of the OPC UA
specification is called OPC Unified Architecture (OPC UA). specification in the Java programming environment.
New characteristics of OPC UA are quite different from old Final results were handed over to the OPC Foundation to be
OPC/DCOM which was meant for straight forward published officially for OPC Foundation members.
345
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_60, © Springer Science+Business Media B.V. 2010
346 SALMENPERÄ AND SALONEN
References [1] and [2] details further results of these projects. good features from existing protocols in order to be able to
The projects were funded by Tekes (Finnish Funding Agency utilize existing security libraries and architectures. This
for Technology and Innovation), as well as Finnish approach also attempts to bypass caveats present in designing
companies Metso Automation, Prosys PMS, Wapice and secure and robust communication protocol from scratch.
Kone. The Technical Research centre of Finland (VTT) and Presented model of OPC UA communication stack is based
Tampere University of Technology (TUT) participated in the on the specification parts [7], [9], [10] and [11], OPC
development of the Java implementation and Helsinki Developers Conference materials [12] and the knowledge
University of Technology (HUT) provided expertise. acquired from above mentioned projects.
UA high level tool API signifies a need for 3rd party toolkit Policies have been defined in UA specification part seven to
that expedites the development of UA applications. It respond to the information security need for the different
provides abstraction of UA stack and possibly handles routine application fields, as comprehensively as possible. For
functionality such as certificates, sessions and access control. example, the 128-bit encryption algorithms are sufficient to
This layer is no required by specification but rather is secure the normal applications at the moment, but UA also
implicitly present in all UA applications. Above this layer is defines algorithms that use 256-bit encryption algorithms.
UA Vendor application which is the actual client/server Also, more UA Security Policies can be defined to UA
software. according to demands. [3]
In addition to the Security Policies, the level of security in
UA may be tuned using the concept of
IV. OPC UA BINARY PROTOCOL SECURITY FEATURES MessageSecurityMode. MessageSecurityMode defines
whether the messages are signed and encrypted or only
One of the new features for the OPC UA is the security model signed, before sending them to the network. There is also a
that is a built-in property of the new protocol. The main ‘None’ mode defined, meaning that all the security features
requirement for the security model is to offer a flexible directed to messages are turned off. [3]
security approach as OPC UA may be used in such diverging
environments. As a result, the UA applications must be able
to mitigate several threat types and security requirements.
OPC UA defines a secure way to transfer the data between
UA applications where the generality is added with the help
of UA Security Policies. The following describes the security
features that are available when the UA Secure Conversation
(UASC) protocol is used.
The security of the messages that use the binary protocol can
be improved by using the UASC protocol. By using the
protocol, messages between the UA applications may be
encrypted and signed according to the predefined encryption
algorithms. As a result, the integrity and confidentiality of
messages are assured.
In OPC UA, the principal factor of the security is the concept
of Secure Channel, which is implemented as an independent
layer to the OPC UA protocol stack. The Secure Channel is a
long-term logical connection between OPC UA client Figure 3: UA Create Session sequence
application and OPC UA server that enables the
confidentiality and integrity of the messages. This secure B. The Security handshake
connection must be opened between the applications before
the actual device and system data may be exchanged between As discussed earlier, the connections between UA
UA applications. The Secure Channel can be opened by applications are secured when the connection is established.
calling an OpenSecureChannel method specified in Secure This part of the UA communication is called “Security
Channel Service Set which defines the methods for the Secure handshake” and it is performed according to the sequence
Channel. The secured channel is formed according to the which is depicted in Figures 2 and 3. The encryption and
security handshake discussed in the next chapter. The signature of messages take place by using asymmetric
messages that are related to the security handshake may be encryption and digital signature algorithms. The Security
both encrypted and signed or they can be sent without Handshake contains several phases: Firstly, the Secure
security procedures. Naturally, the channel is secure only if Channel is created. Secondly, an UA Session is created and
the encryption has been used. Thereby, Secure Channel is thirdly, the Session is activated. [3]
created even if no security or encryption is used. [8][9] The Secure Channel may be constructed using the
The security of the Secure Channel is based on certificates, OpenSecureChannel service. The Session is created and
chosen UA Security Policy and fine-tuning of security level activated by service methods defined in Session Service Set
(Message security mode). The certificates are used to verify described in UA specification part 4 in reference [9].[9]
the parties’ identities. Also, the keys contained in the The OpenSecureChannel request is always made by the UA
certificate can be used for encryption and signature purposes. Client application that first signs the request by using its
The UA Security Policy specifies the algorithms which are private key, after which the request is encrypted with the help
used to encrypt and sign the messages. These Security
348 SALMENPERÄ AND SALONEN
of the public key to be obtained from the receiver's certificate. Instance Certificate can also be used to sign and encrypt the
In this case, the receiver is the UA Server chosen to be messages that use the asymmetric algorithms. [3]
connected to. The Software Certificate is used to identify the particular
When an OpenSecureChannel request arrives at the UA OPC UA product. The Software Certificate contains the
server from the UA client (and the message has been information about the profiles supported by the product. In
encrypted), the server will first decrypt the message using the practice, the OPC UA server returns in the response of the
private key of the server. Also, the integrity of the message is Session service to the UA client program the Software
verified by the server. After that, the server verifies the trust Certificates list on the basis of which the client program can
to the sender by checking the sender’s certificate that is conclude the supported services of the server. When a session
received as one part of the OpenSecureChannel request. is activated, the client program has to transmit its own
Before the UA server may send the response message back to Software Certificates to the UA server. The server can forbid
UA client, the server also processes other parameters of the the activating of the session, if necessary, in case the Software
request and calculates the derived symmetric keys to itself Certificate does not meet the demands of the server. The
with the help of generated nonce. Finally, the server sends a supplier of the UA product always delivers the Software
response to the OpenSecureChannel request. The response Certificates with the product. [3][9]
contains along with other information, an unambiguous The User Certificates can be used, when activating a session,
identifier of the channel determined by the server as well as to identify the user of the UA application. OPC UA also
the nonce of the server that the client program requires in provides other opportunities to identify the user, for example
order to calculate the symmetrical keys. [3][11] the combination of username and password or web service
The level of the information security of the response is the security token instead of the User Certificate. UA also
same as in the request. Correspondingly, the possible provides a possibility to use the client applications
encryption and signature take place like in the case of the anonymously. [3] In that case, the user is not identified at all.
request. After the Secure Channel has been opened between
the UA applications, the possible encryption and signature of V. IMPACT OF ENABLING SECURITY
messages take place with the help of symmetrical algorithms.
However, before an actual use of UA, a Session must be The usage of encryption and signature ascertain the
created and activated. This can be achieved by using the confidentiality and integrity of the messages. However, the
Session Service Set that defines the methods for the Session. encryption and signature of messages require considerably
With the help of the Session, a person may be authenticated more resources than those of the messages that are sent
and authorized to call the actual UA services like Browse, without securing them.
Read or Write. The Session creation and activation part of the The Department of Automation Science and Engineering of
security handshake has been presented in a sequence [Figure Tampere University of Technology (TUT) began a
3]. [3][9] preliminary efficiency testing concerning the security in
February 2009. According to the plan, the purpose was to test
the effect of different encryption algorithms on the capacity
C. Role of Certificates for the Java UA protocol stack. The OPC UA stack
implemented using Java programming language is developed
A digital certificate is a structure that associates a public key in Finland as the part of TEKES funded project, ”OPC
with an identity. As a result, the digital certificates may be Unified Architecture Information Modeling and Java
used to identify a party that may be for example a person, an Implementation”.
application, a product or a machine. A certificate always The tests are performed by sending messages between the UA
contains a public and private key with some other applications. The messages are protected by using the several
identification information depending on a certificate type. The different UA Security Policies and security modes supported
keys combined to the certificate may be used for message by the Java OPC UA Stack. The UA applications function in
encryption and signing. [3] separate virtual machines and the measured data was
OPC UA uses X.509 certificates. The certificates in UA are collected by using the tcpdump program. In tests, the time
used in three different purposes depending on the type of from the client program’s request (first TCP packet sent) to
certificate: to verify the UA application (Application Instance the time the server returns the response (first TCP packet
Certificate), to verify UA product (Software Certificate) and sent) was measured. 20 requests were carried out in one
to identify a user of UA client program (User Certificate). measurement for each message size towards every supported
The usage of the two first mentioned certificate types is security level. Every test was performed three times, resulting
mandatory in UA, whereas the usage of the User Certificate is in total of 60 requests.
optional. [3] Table 1 depicts the averages calculated from the results. In
The Application Instance Certificate of UA is a digital figures 4, 5 and 6 the effect on the capacity of the message
certificate which can be used to identify the individual UA sizes and encryption types is seen. Figure 4 will show
application, thus every separate UA application (installation) measurement results when a payload of 1024 bytes is
requires its own application certificate. The Application transported. In Figure 5, payload was 10240 bytes, and Figure
6 outlines 102400 bytes payload transfer times.
ARCHITECTURE AND INITIAL PERFORMANCE INDICATORS 349
Table 1: Averages in milliseconds using the 256-bit encryption and signature algorithms. The
Payload Rsa128 Rsa128 RSA256 None Effect difference with regard to the non-secure transmission of the
size in Enc & Sign Enc & of message was about 10-fold. It is true that the message sizes
Bytes Sign Sign security that were tested are large considering the purpose of the
1024 2,0 0,8 1,7 1,0 2X UASC protocol, but the main target in these tests was to see
the effects of the different security policies. As seen from
10240 31 1,8 27 1,7 18 X Table 1, the differences were more distinct when the amount
102400 65 14 88 8,1 11 X of data was increased.
One can perceive from the results how much both the size of VI. CONCLUSIONS AND FUTURE WORK
the message and the encryption algorithm used affects to the
processing times. With the smallest amount of the data, the
OPC foundation has chosen a new approach to integration.
128-bit encryption was the slowest, since the encryption and
The clear and fundamental change of OPC UA compared
decryption of the messages took about twice a time longer
with previous OPC specification is neutrality. OPC UA is no
than processing the message sent without any security. The
longer coupled to any 3rd party technology. It is also platform
biggest payload size (102400 bytes) was the slowest when
and programming language independent. The advantages of
350 SALMENPERÄ AND SALONEN
this choice are significant as OPC UA is targeted for Industrial Informatics (INDIN 2008), Daejeon, Korea, July
integration in industrial environments, where life time of 13-16 , 2008.
applications exceeds conventional office application by 2. Implementing OPC UA in Java; Mikko Salmenperä, Seppo
decades. On the other hand OPC specifications have grown Kuikka, Jyrki Peltoniemi, Tommi Karhela, Marko Mattila,
significantly in both size and complexity. Jouni Aro, Olli Snellman, Esa Salonen; Automaatio 07
OPC UA resides completely on OSI model’s application Seminar; Finnish Society of Automation; 2007; ISBN 978-
layer even though parts of it could have been moved to lower 952-5183-33-7
OSI levels in OSI model. This would have reduced the size in 3. OPC Unified Architecture, Wolfgang Mahnke, Stefan-
OPC UA stack and spared resources for other important new Helmut Leitner, Matthias Damm, 2009, Springer
aspects of OPC UA, such as data model. Security 4. Providing device integration with OPC UA, Hadlich, T.,
implementation is notoriously vulnerable to bugs and need Industrial Informatics, 2006 IEEE International Conference
constant updates and upkeep. It has been said that when on, Aug. 2006.
security protocol passes initial quality assurance and audit it 5. Bender K., Grossmann, D., Danzer, B, Future Device
still needs to be used in a real world for at least five years Integration, V1.1, TUM, ITM, 2007.
before any creditability can be accounted to it. 6. X.200: Information technology - Open Systems
Performance of OPC UA stack binary protocol Interconnection - Basic Reference Model: The basic model.
implementation can be considered good and even the initial ITU-T Recommendation X.200. 1994.
test shown in this paper clearly demonstrate that it provides a 7. OPC Foundation, ‘OPC Unified Architecture Specification
good tool for both vertical and horizontal integration. Effect Part 1: Concepts’, Version: Release 1.01, February 2009
of security becomes apparent with larger messages but as it 8. OPC Foundation, ‘OPC Unified Architecture Specification
can be used only where necessary there should not be major Part 2: Security Model, Version: Release 1.01, February 2009
problems for even embedded environments. 9. OPC Foundation, ‘OPC Unified Architecture Specification
Part 4: Services, Version: Release 1.01, February 2009
10. OPC Foundation, ‘OPC Unified Architecture Specification
REFERENCES Part 5: Information Model’, Version: Release 1.01, February
2009
11. OPC Foundation, ‘OPC Unified Architecture Specification
1. Hannelius, T., Salmenperä, M., Kuikka, S., "Roadmap to
Part 6: Mappings, Version: Release 1.00, February 2009
Adopting OPC UA", in 6th International IEEE Conference on
12. OPC Foundation, DevCon 2007, OPC Unified
Architecture, Conference Materials, OPC Foundation, 2007
Automatic Segmentation of Cardiac Cavity Images
Using Collinear and Triangle Equation
Riyanto Sigit, Mohd. Marzuki Mustafa, Aini Hussain, Oteh Maskon, and Ika Faizura Mohd Noh
Dept. of Electrical, Electronic and Systems Engineering,
Faculty of Engineering and Built Environment,
Universiti Kebangsaan Malaysia
Bangi, Malaysia
[email protected], [email protected], [email protected]
Abstract- Automatic segmentation of cardiac cavity images used for segmentation of cardiac cavity [8]. Accordingly, we
using collinear and triangle equation algorithms to detect and have devised a technique which involves the use of collinear
reconstruct the boundary of cardiac cavity is afforded in this and triangular equations. Our proposed technique employs the
paper. Firstly, high boost filtering is used to enhance the high high boost filter to enhance the high frequency component
frequency component without affecting the low frequency
component. Next, to eliminate noise and convert the image into
while keeping the low frequency component unaffected. This
binary format, the morphological and thresholding operators are is followed by applying the morphological and thresholding
applied to the image. Then, using negative Laplacian filter operators to the image to eliminate noise and convert it to the
followed by region filtering, the edge detection is performed. binary format. Next, Laplacian filter and region filtering are
Finally, the collinear and triangle equations are used to detect applied in order to perform the edge detection process.
and reconstruct the more precise cavity boundary. Results The final step is the implementation of the collinear and
obtained showed an improvement in the detection boundary of triangle equation which is used to detect and reconstruct an
cardiac cavity which implies the suitability of the proposed improved cardiac activity border. It is therefore, the aim of
technique in performing the boundary detection task of cardiac this paper to present our proposed technique which we believe
cavity from echocardiographic images.
can be used to perform segmentation of cardiac cavity very
Keywords: Cardiac cavity, high boost filter, morphology,
precisely. The following section describes the development of
negative Laplacian, region filter, collinear and triangle equation. the proposed technique further and outcomes of the
implementation are shown in the results and discussion
section.
I. INTRODUCTION
II. PROPOSED AUTOMATIC SEGMENTATION METHOD
The two-dimensional (2D) short axis echocardiography The schematic diagram of the proposed technique is shown
images have been used by medical practitioners to diagnose in Fig.1. As shown, the collinear and triangle equations
patient cardiac condition by studying the cardiac cavity. As implementation represents the last step in our proposed
such, various researches and methods have been conducted to automatic segmentation technique to detect edge and
develop computerized technique to detect cardiac cavity [1-9]. reconstruct the border. Prior to it, the echocardiography image
Although many techniques have been developed, there is still is subjected to a 3-step procedures involving high boost
room for innovation and development of such methods and
algorithms. For cardiac activity detection, researchers have Viewed Image containing cardiac
used either the short axis images [1-6] or long axis images i
[7&8] implemented in semi-automatic mode [1, 4 & 6] or
fully automatic [3 & 7]. In addition, active contour models or High Boost Filtering
snakes have also been used to segment cardiac cavity [5 & 8].
J. W. Klinger et al in [1] have applied segmentation of
Morphology & Thresholding
echocardiography images using mathematical morphology.
On the other hand, Laine and Zong in [2] have used border
identification that depended on the shape modeling and border Negative Laplacian & Region Filtering
reconstruction from a set of images. Detection of left
ventricular endocardium using ternary threshold method was
Collinear & Triangle Equations
introduced by Ohyama et al [3], in [1 & 4] the semi automatic
detection method was employed and the multiple active
contour models for cardiac boundary detection by Chalana et Cardiac cavity boundary detected in
viewed image
al is presented in [5]. In a more recent work, watershed pre-
segmentation snake was used for boundary detection and to Fig.1. Schematic Diagram of the algorithm
find its center point [6 & 7]. Of late, applied radial-search was
351
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_61, © Springer Science+Business Media B.V. 2010
352 SIGIT ET AL.
filtering, morphological and thresholding and a double On the contrary, the closing algorithm is when image A is
filtering using the Laplacian and region filters. dilated and eroded by B. The mathematical notation of the
closing algorithm is shown in equation 2.
A. Cardiac Short Axis Image View
(2)
There are two main standard views used in cardiac imaging
that is the short axis and long axis. In this research, we have D. Negative Laplacian & Region Filter
used the short axis images to test out our proposed
segmentation technique. An example of the image comprising Next, the negative Laplacian filter is implemented. It is a
the left ventricular cardiac cavity from echocardiography is derivative filter that is used to find areas of rapid change (i.e.
shown Fig. 2. edges) in an image. There are different ways to find an
approximate discrete convolution kernel that can approximate
the effect of the Laplacian. A possible kernel is as shown in
Fig.4 below.
0 1 0
1 -4 1
0 1 0
Fig. 4. Kernel used for negative Laplacian.
c a
A b C
Fig.5. A triangle
a 2 = b 2 + c 2 − 2bc(cos A) (7)
Fig. 6. Sample result from high boost filter implementation
A = a cos((b 2 + c 2 − a 2 ) / 2bc ) (8)
IV. CONCLUSION
ACKNOWLEDGMENT
REFERENCES
[1] J. W. Klinger, C. L. Vaughan, and T. D. Fraker, “Segmentation of
echocardiographic images using mathematical morphology,” IEEE
Trans.Biomed. Eng., vol. 35, 1988, pp. 925–934.
[2] A. Laine, and X. Zong, “Border Identification of Echocardiograms via
multiscale edge detection and shape modeling,” Proc. of the IEEE Int.
Conf. on Image Processing, vol. 3, Sep. 1996, pp. 287 - 290 vol.3.
[3] W. Ohyama, T. Wakabayashi, F. Kimura, S. Tsuruoka, and K. Sekioka,
(b)
“Automatic Left Ventricular Endocardium Detection in
Echocardiograms Based on Ternary Thresholding Method”, in 15th
International Conference on Pattern Recognition (ICPR’00), Barcelona,
Spain, 2000, pp. 320-323.
[4] Maria do Carmo dos Reis, Adson F. da Rocha, Daniel F. Vasconcelos, etc,
“Semi-Automatic Detection of the Left Ventricular Border “, 30th
Annual International IEEE EMBS Conference Vancouver, British
Columbia, Canada, August 20-24, 2008.
[5] V. Chalana, D. T. Linker, D. R. Haynor and Y. Kim, “A Multiple Active
Contour Model for Cardiac Boundary Detection on Echocardiography
(c) Sequences”, IEEE Trans. on Medical Imaging, vol.15, no. 3, June 1996.
[6] Cheng, J, Foo, S.W. & Krishnan, S. M. “Automatic Detection of Region of
Interest and Center Point of Left Ventricle using Watershed
Segmentation,” IEEE Int. Symposium on Circuits and Systems, vol. 1, n.
2, May 2005, pp. 149-151.
[7] Cheng, J, Foo, S.W. & Krishnan, S. M. “Watershed-Pre-segmented Snake
for Boundary Detection and Tracking of Left Ventricle in
Echocardiographic Images,” IEEE Trans. on Information Technology in
Biomedicine, vol. 10, n. 2, 2006, pp. 414-416.
Fig.11.The step-by-step results to eliminate small contours [8] Lacerda SG, da Rocha AF, Vasconcelos DF, de Carvalho JL, Sene IG Jr &
using the OpenCV library functions. Camapum JF, “Left Ventricle Segmentation in Echocardiography Using
a Radial Search Based Image Processing Algorithm”, 30th Annual
Step-by step results using the triangle equation to eliminate International IEEE EMBS Conference Vancouver, British Columbia,
small contours by small corner angle computation are shown Canada, August 20-24, 2008
in Fig. 11(a), (b) and (c). Figure 11(c) depicts the final cardiac [9] M. Kass, A.Witkin, and D. Terzopoulos, “Snakes: Active contour
models”, presented at the Int. Conf. Computer Vision, ICCV’87,
cavity boundary that has been successfully detected and London, U.K., 1987.
precisely reconstructed. [10] Gary Bradski and Adrian Kaehler, “Learning OpenCV”, O’Reilly Media,
September 2008.
AUTOMATIC SEGMENTATION OF CARDIAC CAVITY IMAGES 355
Abstract - The paper presents the design of the I.I THE FIRST ALTERNATIVE
principle of multiple output both side differential
transmissions with integrated harmonic transmission. The Transmission of revolutions and performance is
presented principle enables more coincidental coaxial possible through the input eccentric shaft I, where
outputs with the alternative of input and output fulcrum wheel „2“ and „7” is located with gearing z2 and
interchange. z9 and firmly connected gearing z5 and z7, meshing with
gearing z6 eventually z8 connected with the output shafts
INTRODUCTION II, V.
The next output is possible either through the
Department of technological equipment design of flexible member (disk wheel) „3“ to the output shaft III,
the FVT of the TU in Košice within the process of or through the crown wheel „4“ connected with the output
searching for alternatives to the so-called high precision shaft IV. The second output is done through the flexible
transmissions of the constructions based on cycloidal member „3“ with the internal gear z3a and external gear
teeth with the relatively complicated transmission of z3b.
movement from satellite to centric has aimed it research at Thus, gearing z2 and z9 of the wheel „2“ and „7”
the new principles for design with relatively simple meshes with internal gear z3a of the flexible member „3“
transmission of movement as well as presented and consequently the flexible member by means of
transmissions regarding the number and range of external gear z3b meshes with the internal gear z4 of the
transmission ratios. non-flexible member of the crown wheel„4“. If the crown
The principle of integrated harmonic transmission wheel „4“ is braked, corresponding performance of the
with the multiple both side outputs has been suggested. third output is transmitted to the shaft III by means of the
The paper also presents the kinematic analysis of flexible member „3“. If the flexible member „3“ is braked
the designed construction with transmission ratios for the , performance is transmitted by means of the member „4“
certain levels. to the shaft IV.
357
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_62, © Springer Science+Business Media B.V. 2010
358 HAL’KO
From the kinematic ratios shown in fig.2 it follows that v A 2O = v A 2 E + vE O = v A3O (2)
for the velocity vector at point „A“ the following is
valid: From the equation (1) it follows that at the point A the
following is valid:
v A 2 O = v A 3O (1)
ω A 2 O = ω A3O (3)
After substitution the velocity vectors at the point A
v A3O = ω A3O . r3 ae , v A 2 E = ω 2 E .r2 , v E O = ω E O . re
and if we take into consideration the supposed direction of
revolution regarding fig. 2 the following is valid:
ω A 3 O . r 3ae = ω 2 E . r 2 + ω E O . r e (4)
Regarding the fact that body „3“ is in the part of gearing
z3a , z3b flexible, ω A 3 O it does not always equal the angle
velocity of the body „3“, thus ω A3O ≠ ω 3O . Angle
velocity of the non-deformed body „3“ equals angle
velocity of the input shaft III. With constant ω I = ω EO
where u 3
I −2 - is transmission ratio between shaft If ω5 = ω 2 (gearings z 5 and z 2 are firmly gripped on
I and wheel „2“ when member „3“ is braked. the wheel „2“) then we can substitute from the equation
m z 3a m z2 (9) ω 2 = ω 5 to the equation (12). From the transmission
If we substitute r 3 a = and r2 = , (m- design it follows that angle velocity of the shaft V equals
2 2
3 the angle velocity of the gear wheel z 4 , because of the
module gearing z2,z3a) then transmission ratio u I − 2 is
as follows:
firm connection with the shaft „V“. Thus ω 6 = ωV .
3
z2 ωI After substitution we will obtain:
u I −2 = = (8)
z 2 − z 3a ω2 § r −r ·
r6ωV = r5 ¨¨ ω i 2 3a ¸¸ + (r6 − r5 )ω I
b) Transmission ratio between the shafts I and IV when © r2 ¹
member „3“ is braked. ωI
Ratio is the transmission ratio between shaft
Output shaft V has internal gear wheel. ωV
From the equation (6) and (8) for kinematic ratios of gear I and V when member „3“ is braked. Let us denote it by
wheels z 2 and z 3 of the harmonic transmission for angle u I3−V and angle velocity ωV by ωV3 (index 3 – braked
velocity of the internal gear wheel „2“ it follows member „3“ , and consequently we will obtain the
r1 − r 2 following:
ω2 = ωI (9) ωI r6 r2
r2 u I3−V = = (13)
ωV r5 (r2 − r3a ) + r2 (r6 − r5 )
3
If annual gears z 2 and z8 are on the same body of the
After the radius through teeth number has been defined:
gear wheel the following is valid: ω 2 = ω8 . ωI z6 z 2
Eccentricity e at the same time equals the radius of u I3−V = = (14)
revolution of the wheel axis „2“ (of the point E) around ωV z 5 ( z 2 − z 3a ) + z 2 ( z 6 − z 5 )
3
4
where ω III - is the angle velocity of the shaft III when III.II. TRANSMISSION RATIOS WHEN MEMBER „3“
IS BRAKED AND n3 = 0
member„4“ is braked.
a) Transmission ratio between the shafts II and I
b) Transmission ratio between the shaft I and V
For the angle velocity of the shaft V the following is 1
u II3 − I =
valid: ω =ω +ω
V
4
V
3 4
III (17) u I3− II
where b) Transmission ratio between the shafts II and IV
ωV4 - is the angle velocity of the shaft V when member u I3− IV
u II3 − IV =
„4“ is braked u I3− II
ω V
3
- is the angle velocity of the shaft V when member c) Transmission ratio between the shafts II and V
„3“ is braked u I3 − V
Let us denote transmission ratio between the shaft I and u II3 − V =
V when member „4„ is braked by u I 4− V , then u I3 − II
ωI ω CONCLUSION
u I 4− V = = 3 I 4 (18)
ωV ωV + ω III
4
The objective of the paper is to present
principally new design of multiple output both side
After expression ωV from
3
the equation (14) and ω III4 transmission with the possibility of input and output
from the equation (16) and after substitution to the interchange with the alternatives of the possible three
equation (18) we will obtain: outputs from the given transmission mechanism. Based on
z 6 z 2 z 3b the kinematic scheme, corresponding transmission ratios
u I4−V = have been derived. Within the further research, force
[z 5 (z 2 − z 3a ) + z 2 (z 6 − z 5 )]z 3b + (z 3b − z 4 )z 6 z 2 ratios analysis and possible loadings of the mentioned
Relationship foR transmission ratio u I 4− II between the mechanism, eventually the other corresponding solutions
[1], [4], [9] will be given great attention. This
shafts I and II when member 4 is braked is of the
contribution forms a part of the solution of the grant task
analogical form , although when the corresponding
VEGA 1/4156/07 and KEGA 3/6279/08.
number of teeth of the gearing are substituted.
measurement or detection problem, while others constitute a An aspect of the science of photic fields relates to the
sub-system of a larger design. The specific implementation of interaction of the photic field power P(), which in general
an intelligent optical system also depends on if its may be a complex function of wavelength with matter of
functionality is pre-specified or if some part of it can be different kinds, via its responsivity R(), to produce a physical
learned or modified during operation. or chemical effect V [2] according to:
In many intelligent optical monitoring systems intensity
modulation offers the advantages of inherent simplicity. ª º
m
361
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_63, © Springer Science+Business Media B.V. 2010
362 TOMTSIS
Computer ª § V i − V min
H = 120 « m i + ¨
·º
¸»
¨ V + V − 2V ¸ (4)
¬« © i j min ¹ ¼»
Sensor
Camera a) with mi=0 for i=1, j=2, Vmin=V3, b) with mi=1 for i=2,
j=3, Vmin=V1, c) with mi=2 for i=3, j=1, Vmin=V2
where P(λ) is the spectral power distribution in the optical III. EXPERIMENTAL SYSTEM
signal and Rn(λ) is the wavelength responsivity of the nth
detector and is the wavelength. The color model The optical monitoring system is computer based and is
representing this mathematical formalism is generally called broken down into three distinct and important elements: a
RGB and it is widely used in self-luminous display camera, a color frame grabber card (CFG), and a host personal
technologies [4]. computer.
Each detector output may also be intensity normalised The camera converts the spatially distributed optical
according to: information into analog electric signals which are accepted by the
CFG. The signal is then digitised and stored in the card's frame
memory. The C.F.G. card is the interface between the camera
u n = V n / ¦ n VT (3) and the host personal computer which is used to access and
program the card's registers, frame memory, and buffers. Fig. 1
and chromaticity maps may be formed in terms of the illustrates graphically the configuration of the system.
coordinates u1, u2, … u(n-1) (un becomes redundant since Although many measurands may be fully specified by
¦ n uT = 1 ). distimulus detection [9] and dominant wavelength alone, others
The special case of n=3 leads to a two-dimensional require tri-stimulus detection [10] to completely and
chromaticity map (u1:u2) on which changes in optical signals unambiguously determine their state. The suggested intelligent
may be traced. The special case when R1(), R2(), R3() optical monitoring system is tri-stimulus in nature and its
correspond to the responsivities of the human eye leads to the hardware and software components are fully described in the
chromaticity map reducing to the CIE diagram of color following sections. The built-in intelligent nature of the
science [5], [6]. monitoring system results from its ability to automatically
The color model representing this mathematical formalism compensate for any lightness variation as well as its ability to
is called LXY and provides the relative magnitudes of the decide the gamma correction process based on the surrounding
tristimulus values (i.e. x=u1; y=u2; z=u3). environment and the event being monitored.
Although the above method of displaying chromatic A. Optical Frequency Response
information has proved the information compression
The optical frequency response of the CCD camera was
capabilities of the chromatic approach, it does not easily lead
measured using the optical monitoring system and the
to the identification of signal changes in terms of fundamental
apparatus shown in fig. 2.
signal properties. An alternative approach to the processing of
the signals from such chromatic detectors overcomes this
limitation. For the tristimulus case (n=3) this approach utilizes
INTELLIGENT PLASMA MONITORING SYSTEMS 363
Output (a.u)
Green
100
Blue
A lign m en t
S l its O sc illoscope 50
Red
R o ta tab le
S tand
O SA 0
420 440 460 480 500 520 540 560 580 600 620 640 660
Wavelength (nm)
Fig. 2. Apparatus used for the optical frequency response of the color CCD
camera. Fig. 3. Optical frequency response curves of the Red, Green and Blue
camera detectors (tri-stimulus detection).
The required part of the spectrum was selected with the The procedure adopted in this work is similar to the Von
alignment slits and was directed to both the camera and the Kries transformation [11] with the difference being that
optical spectrum analyser (OSA) using the half silvered instead of using a white color (or greys of the same
mirror. The oscilloscope was used to display information chromaticity) as a reference, any carefully selected color can
about the light intensity as a function of the selected be used for the same purpose. In this procedure if a stimulus
wavelength incident to the camera and the OSA. Different gives rise to the r,g, and b color responses then the resulting
wavelengths were selected by rotating the diffraction grating responses will depend on r/rc, g/gc, b/bc, where rc, gc, and bc
on the rotating stand. The resulting information was then are the chromatic coordinates of the reference color.
processed by the host personal computer to provide color For a stimulus in the state considered to have the same color
quantification that correlated with the wavelength monitored appearance as the same stimulus in a reference state it is
by the OSA. necessary that:
The CCD array response of the color CCD camera was
measured in a dark environment to avoid the unwanted effects r / rc = r ' / rc' , g / g c = g ' / g c' , b / bc = b ' / bc' (7)
of external lighting. The camera was used with its focus and
aperture lens removed so that the light reflected from the
where r’c, g’c and b’c are the color responses of the reference
grating was incident directly on the CCD element.
color in the reference state. Hence
Measurements were taken at 25 nm intervals.
The calibration results of the camera are shown in fig. 3.
The obtained frequency response curves show that the CCD r ' = ( rc' / rc ) r , g ' = ( g c' / g c ) g , b ' = ( bc' / bc )b (8)
sensor is composed of three detector elements (tri-stimulus
detection), with peak responses at 595, 525 and 445 nm for The resulting adapted variables r’, g’, b’, obtained from (8),
the red, green and blue detectors respectively, and spectral can then be used to calculate the LXY tristimulus values as well
bandwidth 170 nm approximately for each detector. as their HLS representation.
The frequency response curve of each detector is utilised by Hence, a calibration method for correcting the variations in
eq. 1 and corresponds to the Rn(λ) wavelength responsivity RGB chromatic values was developed which concentrated on
parameter. removing the RGB errors introduced by non-uniform
illumination and external light fluctuations of the external
B. Chromatic Calibration environment.
The chromatic information received by the intelligent The RGB values, the chromatic coordinates X, Y and the
monitoring system depends on the lighting conditions of the corresponding HLS values of a custom color patch were
surrounding environment. This means that the chromatic co- recorded under controlled laboratory lighting conditions [12],
ordinates of an object under daylight conditions will be which have been used as reference information for any further
different than those under artificial light conditions. Other processing.
parameters that affect the chromatic measurements are
lighting fluctuations, secondary light sources and shadowing. C. Gamma compensation
In a chromatic based system, this ambiguity which is involved An important characteristic of the human visual system is
with the measurement of an objects chromaticity makes it that it is sensitive to ratios of intensity levels rather than to
difficult to relate the optical emissions received from an object to absolute values of intensity. Therefore, the intensity levels
the measurand. The RGB, Lxy and HLS color models are should be placed logarithmically rather than linearly to
designed to be used with applications where the lighting achieve equal steps in brightness [13]. Gamma compensation is
conditions are not too different from daylight. Under different a facility available on most CCD cameras and it is applied to the
lighting conditions alternative procedures are usually adopted R, G, and B detector responses to simulate the non-linear
and incorporated into these models to allow for lightness
adaptation.
364 TOMTSIS
0,7
0,6
0,5
Chromaticity Y
0,4
Gamma Off
0,3
0,2
0,1 Gamma On
0
0,3 0,4 0,5 0,6 0,7 0,8 0,9 1
Chromaticity X
REFERENCES
[1] D. Tomtsis, V. Kodogiannis, “Optical pH Measurement Using
Chromatic Modulation”, International Conference of Computational
Methods in Sciences and Engineering, ICCMSE 2003 Kastoria, Greece
12-16 September 2003.
[2] P. Moon and D.E. Spencer, “The Photic Field”, Cambridge, MA:MIT,
1981.
[3] D. Tomtsis, V. Kodogiannis, “Digital Image Processing of Axis-
Symmetric Electric Arc Plasma Based on Chromatic Modulation
Methods”, 6th WSES /IEEE Multi-conference CSCC-MCP-MCME
2002, Crete, July 2002.
[4] D. Tomtsis, K. Sapalidis, “A CCD-Based Tristimulus Colorimeter for
Fig. 7. Location on a chromaticity diagram of the emission from the plasma Fast and Improved Measurement of Chromaticity Coordinates Of
of Argon, Hydrogen and Nitrogen gases (diamond) and the white light point Displays by Matrix Correction Methods”, WSEAS Transactions on
(square). Information Science and Applications, Issue 6, Volume 1, December
2004, pp. 1606-1611, ISSN: 1790-0832.
[5] CIE, Colorimetry Publication 15, 1971.
366 TOMTSIS
[6] CIE, Colorimetry Publication 15.2, 1986a [11] R.W. Hunt, “Measuring Color”, J. Wiley and Sons, New York, 1987.
[7] P.C. Russell, J. Cosgrave, D. Tomtsis, A. Vourdas, L. Stergioulas and [12] C. S. McCamy, H. Marcus and J.G. Davidson, “A color-rendition chart”,
G.R. Jones, “Extraction of Information from Acoustic Vibration Signals Soc. Photog. Scientists Engineers, vol. 2, no 3, 1976, pp. 95-99.
Using Gabor Transform Type Devices”, Measurement Science and [13] J.D. Foley, A. Van Dam, K.S. Feiner and J.F. Hughes, “Computer
Technology, Vol. 9, August 1998, pp. 1282-1290. Graphics: Principles and Practice”, Addison-Wesley, 1996.
[8] J.D. Foley, A.V. Dam, S.K. Feiner, J.F. Hughes JF, “Computer graphics: [14] G.R. Jones, “High Pressure Arcs in Industrial Devices: Diagnostic and
Principles and practice”, 1995, Addison-Wesley. Monitoring Techniques”, Cambridge University Press, Cambridge,
[9] D. Tomtsis, V. Kodogiannis, P. Chountas, “Distimulus Chromatic 1988.
Measurement Systems”, WSEAS Transactions on Circuits and Systems, [15] G.R. Jones and M.T.C. Fang, “The Physics of High Power Arcs”, Rep.
Vol. 3, No. 2, April 2004, pp. 211-214, 2004. ISSN: 1109-2734. Prog. Phys., 43, pp. 1415-1465, 1980.
[10] D. Tomtsis, K. Sapalidis, C. Katsanos, “Electronic Tristimulus [16] C.H. Flurscheim, “Power Circuit Breaker Theory and Design”, Peter
Chromatic Measurement Systems”, WSEAS Transactions on Circuits Peregrinus Ltd., Herts, 1975.
and Systems, Issue 9, Volume 3, November 2004, pp. 1835-1840, ISSN
1109-2734.
A Neoteric Chaotic & Adaptive Watermarking Scheme for
Colored Images
Navita Srivastava* & Achintya Singhal †
*
Department of Computer Science †Department of Computer Science
Awadesh Pratap Singh University, Rewa Banaras Hindu University, Varanasi
E-mail: [email protected] E-mail: [email protected]
Abstract- Today users are enjoying the convenience and This paper uses an image oriented chaotic factor for hiding
advantages of advancement in technology. At present both text and colored logo watermarks in the targeted
technology provides high quality multimedia processing, multimedia data as per algorithm outlined in [5]. The
acquisition, flexibility, and reliability at significantly lower effective size of the image watermark is significantly reduced
costs and ease. Digital multimedia documents authentication before embedding. The proposed algorithm extracts a
is thus one of the most important and investigated security graceful watermark from the authenticated image which is
applications. Watermarking schemes for the purpose has resilient to high quality JPEG compression, image format
become evergreen research area. This paper outlines the transformations like PNG, TIFF etc and image enhancements
previous work carried before by the authors and then like color correction and filtering like high pass, Gaussian
proposes an image adaptive chaotic algorithm for noise etc. This is obtained by avoiding the lossless entropy
watermarking of digital mages. The algorithm is tested coding step of common image compressing algorithms. The
against both image and text watermarks and found resilient paper is organized into sections: section II describes the
to compression and filtering. watermark embedding and extraction process followed by
results and discussion in section III and conclusions in
Keywords: watermarking; data hiding; image authentication; section IV.
multimedia security; chaotic, henon map. II. The Process
I. Introduction A. The background
Quality work entails a lot of hard work, coupled with The watermark insertion process is done in transform
talent and use of intricate expensive equipments. But, the domain using Discrete Cosine Transformation (DCT) where
declining technological barriers, as a means to protect a reversible, linear transform action is used to map an image
intellectual property, have emerged as the greatest threat. It is into a set of transformed coefficients, which are then
therefore obligatory to provide copyright protection for such quantized and coded. The watermark insertion strength is
work and retain confidence of brainpower. Cryptography calculated chaotically depending on the host image block
tools available provide digital content integrity verification pixel value. Henon map is used.
bit by bit but multimedia content authentication requirement For image processing, the definition of two-dimensional
is much different from it where by content enhancement/ DCT, block-sized n×n, is given by [23]:
worsening is allowed, without compromising the semantic ª f ( j, k ) ∝ (u ) ∝ (v) × º
n −1 n −1
« »
content. This resulted use of watermarking techniques to hide F (u, v) = ¦¦ « ª ( 2 j + 1) uπ º ª ( 2k + 1) vπ º »
j =0 k = 0 cos « » cos « »»
the authentication information just inside the digital content « 2n 2n
¬ ¬ ¼ ¬ ¼¼
because watermark becomes the part of the multimedia data
and can be recovered to prove authenticity but can’t be ° 1
u ,v = 0
removed from it. The vital condition for watermarking is that
the embedding technique should be such that it should
∝ (u ), ∝ (v) = ® n
2
produce the same results (multimedia document) even after °̄ n
u ,v = 1, 2 ,...,n −1
367
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_64, © Springer Science+Business Media B.V. 2010
368 SRIVASTAVA AND SINGHAL
chaotic systems appears to be random. In general, the chaotic Edge detection algorithm mark the points in the digital
model system is given as: image at which the luminous intensity changes sharply.
x ( n) = f ( x ( n − 1)) Sharp changes in image properties usually reflect important
events and changes in properties of the world. Hence,
where x(n) is a chaotic sequence generated by the
generating an edge map of the image and breaking it into
nonlinear map f (⋅), x(0) is the initial condition. equi-sized blocks (8x8), gives blocks with high activity
We propose the use of model introduced by French blocks as discussed in [1], [2], and [3]. Also, the image is
astronomer Michel Hénon called Hénon Map [21]. The quantized using DCT and JPEG quantization matrix. This
Hénon map is a discrete-time dynamical system and takes a helps in identifying blocks resistant to JPEG compression
point (x, y) in the plane and maps it to a new point and also high activity blocks. Common blocks from the two
processes are then selected.
x n +1 = y n + 1 − ax n2
B. The Embedding Process
y n +1 = bx n The graphical representation of the embedding process is
The map depends on two parameters, a and b. For values given in figure 3. The original image and the watermark is
of a and b the map may be chaotic, intermittent, or converge DCT transformed and the blocks suitable for watermarking
to a periodic orbit. But, generally, for smaller iterations, the are segregated as discussed above. For each suitable block,
model has chaotic behavior. The figure 1 shows the Henon watermark embedding strength factor is calculated using
plot generated for a particular image block. Figure 2 shows Hénon map. The watermark to be incorporated is also DCT
the incorporation strength factor generated for each pixel, to transformed. Incase of image watermarks, the watermark is
be used for watermarking in the block (8 X 8). lossy compressed and doubly DCT transformed before
embedding. With the calculated strength factor and the
provided watermark, the watermark is embedded as
discussed in [2].
Figure 8:
Watermark Used
Figure 10: Watermark extracted JPEG compressed (100% quality) watermarked image
Extracted Watermark 1 2 3 4 5 6
NC 0.9210 0.99323 0.96215 0.9997 0.97752 0.99652
Q 0.74085 0.96754 0.65941 0.95702 0.91934 0.95946
PSNR 16.578 26.03 14.503 24.617 22.035 24.827
PSNR-HVS 61.252 73.772 59.024 71.071 67.342 71.357
Figure 11: Watermark extracted from JPEG compressed (70% quality) watermarked image
Extracted Watermark 1 2 3 4 5 6
NC 0.85414 0.99265 0.94215 0.99247 0.98831 0.99541
Q 0.51453 0.8469 0.49458 0.84481 0.82711 0.86554
PSNR 13.167 18.965 11.979 18.858 18.269 19.050
PSNR-HVS 57.889 64.311 56.7 64.124 63.304 64.278
Figure 12: Watermark extracted from JPEG compressed (25% quality) watermarked image
Extracted
1 2 3 4 5 6
Watermark
NC 0.96133 0.88003 0.90673 0.93168 0.90451 0.9888
Q 0.38554 0.27369 0.26757 0.35738 0.31649 0.46923
PSNR 9.8404 9.6085 8.6569 10.037 9.5879 10.812
PSNR-HVS 54.496 54.391 53.578 55.044 54.305 55.757
370 SRIVASTAVA AND SINGHAL
Figure 16: Watermark extracted from the original Figure 17: Watermark extracted from JPEG compressed
watermarked image (100% quality) watermarked image
A NEOTERIC CHAOTIC & ADAPTIVE WATERMARKING SCHEME 371
IV. Conclusion
In this proposed watermarking scheme, the strength factor
is computed chaotically and dynamically for each block and
each pixel wherever the insertions have to be made. Further,
the watermark is embedded only in sensitive areas of the
image thus securing the actual concerned content of the
image. For identifying sensitive areas, image feature
extraction using edge detection and JPEG quantization
matrix is deployed. Moreover, the concept exploited only the
Y component of YCbCr color space for watermarking as Cb
and Cr components are more prone to loss during JPEG
compression as our human visual system is dull for changes
in these components [22]. The quality of extracted
watermark (colored logo) from the unhampered image
deteriorated in comparison to the algorithms proposed in [1],
[2], [3], and [4] due to compression of the watermark used.
Figure 18: Watermark extracted from JPEG compressed But, the compression helped in achieving more insertions in
(91% quality) watermarked image lesser space thus retrieval in case of cropping is more
promising.
The concept of double DCT introduced for colored
watermarks to incorporate greater sustainability and provides
cryptic sense to blind attacks being made. The algorithm
proved to be generic and the results depict good sustenance
to compression too with both image and text watermarks.
The quality metrics of the extracted watermark depicted
better results than previous algorithms [1], [2], [3], and [4].
Adobe Photoshop CS and Adobe ImageReady CS have been
used for JPEG compression and various attacks.
The values of NC, Q, PSNR, PSNR-HVS reflect the
imperceptibility of the watermark as well as the quality of
retrieved watermark samples. The use of Hénon map
guarantees that the watermarking strength for each pixel
becomes inadvertent. Redundant inclusions and chaotic
strength factor provides significant robustness and helps in
obeying Kerckhoff’s principle [27].
Figure 19: Watermark extracted from JPEG compressed V. References
(85% quality) watermarked image [1] A. Singhal, N. Srivastava and R. Mishra, “Edge sensitive
The scheme, with colored watermark, shows excellent frequency domain watermarking”, in Proc. International
resistance to various attacks like JPEG compression up to Information Technology Conference, IITC, Colombo, Sri
25% quality, color correction, brightness and contrast Lanka, pp. 127-134, Oct. 2006.
alterations, diffusion, high pass filter, uniform and Gaussian [2] A. Singhal, N. Srivastava and R. Mishra, “Hiding
noise, rotations with 90 and 180 degrees etc. Using text as a signature in colored image”, in Proc. IEEE International
watermark has got steganographic value more than digital Symposium on Communication and Information
watermarking. Its prime objective is to convey the hidden Technologies, ISCIT, Bangkok, Thailand, pp. 446-450, Oct.
message to the receiver of the file. Its sustenance even with 2006.
JPEG compression increases it application. [3] A. Singhal, N. Srivastava and R. Mishra, “Robust and
The proposed algorithm used meaningful colored sensitive watermarking scheme for Colored Images”, in
watermark as well as text watermark in contrast to most of Proc. 41st annual convention CSI, Kolkata,, India, pp. 13-17,
the previous work where either random number bit stream Nov. 2006.
[14], or some known bit stream [16], or binary image [9]. [4] A. Singhal and N. Srivastava, “Digital Watermarking of
The algorithm incorporated 8 pixels per selected block (8x8) Colored Images Using DCT-HSI Transformation”, in Proc.
for colored watermarks and 8 bits per selected block (8x8) 40th annual convention CSI, Hyderabad, India, pp. 201-204,
for text watermarks which is significantly high in Nov. 2005.
comparison to 3 bits per block (8x8) in [17], first three AC
[5] A. Singhal and N. Srivastava, “Chaotic Watermarking
components of selected block (8x8) in [14] and 3840 pixels
Scheme for Colored Images”, The Icfai University Journal of
from 768x512 size image [16]. Further, the use of chaotic
Information Technology, vol. 4, no. 4, pp. 23-33, Dec. 2008.
technique helped in attaining so called randomness in
strength factor in comparison to fixed strength factor [9], [6] C. H. Chang, M. Zhang, and Z. Ye, “A content-dependent
[12], [14], and [16]. robust and fragile watermarking scheme”, in Proc. 2nd
International Conference on Visualization, Imaging and
Image Processing, pp. 201–206, Sep. 2002.
372 SRIVASTAVA AND SINGHAL
[7] C. H. Chang, Z. Ye, and M. Zhang, “Fuzzy-ART based Proceedings on Vision, Signal and Image Processing, vol.
digital watermarking scheme”, in Proc. IEEE Asia Pacific 143, no. 4, pp. 250-256, Aug 1996.
Conference on Circuits and Systems, APCCAS, vol. 1, [17] J. Zhao and E. Koch, “Embedding robust labels into
pp. 425–426, Dec. 2002. images for copyright protection”, in Proc. Int. Congress IPR
[8] C. Rey and J. L. Dujelay, “Blind detection of malicious for Specialized Information, Knowledge and New
alterations on still images using robust watermarks”, in Proc. Technologies, Vienna, Austria, pp. 242–251, Aug. 1995.
IEE Secure Images and Image Authentication Colloquium, [18] K. Egiazarian., J. Astola, N. Ponomarenko, V. Lukin, F.
London, U.K., pp. 7/1–7/6, Apr. 2000. Battisti, and M. Carli, “New full-reference quality metrics
[9] C. T. Hsu and J. L. Wu, “Hidden digital watermarks in based on HVS”, in Proc. Second International Workshop on
images”, IEEE Trans. Image Process., vol. 8, no. 1, pp. 58–68, Video Processing and Quality Metrics, Scottsdale, 4 p, 2006.
Jan. 1999. [19] M. A. Suhail and M. S. Obaidat, “A robust digital
[10] Chun-Shien Lu, “Multimedia security: steganography watermarking technique”, IEEE, pp. 629–632, 2000.
and digital watermarking Techniques for Protection of [20] M. Barni, F. Bartolini, V. Cappellini, and A. Piva, “A
Intellectual Property”, Idea Group Publishing, 2004. DCT-domain system for robust image watermarking”, Signal
[11] E. Izquierdo, “Using invariant image features for Processing, vol. 66, no. 3, pp. 357–372, 1998.
synchronization in spread spectrum image watermarking”, [21] M. Henon, “A two-dimensional mapping with a strange
EURASIP Journal of Applied Signal Processing, vol. 4, attractor”, Communications in Mathematical Physics, vol.
pp. 410–417, 2002. 50, pp. 69-77, 1976.
[12] H. Inoue, A. Miyazaki, and T. Katsura, “Wavelet-based [22] M. Ejim and A. Miyazaki, “A wavelet-based
watermarking for tamper proofing of still images”, watermarking for digital images and video”, in Proc.
International Conference on Image Processing Proceedings, International Conference on Image Processing, ICIP, vol. 3,
ICIP 2000, pp. 88-91, Oct 2000. pp. 678-681, Oct. 2000.
[13] J. Cox, J. Kilian, F. T. Leighton, and T. Shamoon, [23] R. C. Gonzalez and R. E. Woods, “Digital Image
“Secure spread spectrum watermarking for multimedia”, Processing”, Pearsons Higher Education 2/E, 2002.
IEEE Transactions on Image Processing, vol. 6, no. 12, pp. [24] R. G. Schyndel, A. Z. Tirkel, and C. F. Osborne, “A
1673–1687, Dec. 1997. digital watermark”, in Proc. IEEE International Conference
[14] J. Huang, and Y. Q. Shi. “Adaptive image watermarking on Image Processing, vol. 2, pp. 86–90, Nov. 1994.
scheme based on visual masking”, Electronics Letters, vol. [25] W. Bender, D. Gruhl, N. Morimoto, and A. Lu,
34, no. 8, pp. 748-750, Apr 1998. “Techniques for data hiding”, IBM System Journal, vol. 35,
[15] J. J. Eggers and J. K. Su, “A blind watermarking scheme no. 3&4, pp. 313–336, 1996.
based on structured codebooks”, in Proc. IEE Secure Images [26] Z. Wang and A. C. Bovik, “A universal image quality
and Image Authentication Colloq., London, U.K., pp. 4/1– index”, IEEE Signal Processing Letters, XX, 2002.
4/21, 2000.
[27] http://en.wikipedia.org/wiki/Kerckhoffs'_principle
[16] J. J. K. Ó. Ruanaidh, W. J. Dowling, and F. M. Boland,
“Watermarking digital images for copyright protection”, IEE
Rotary Transducer
Ilie Borcosi1, Nicolae Antonie2, Alina Dinca3
“Constantin Brancusi” University
Calea Eroilor, no.30, Targu-Jiu, Romania
1
[email protected], [email protected], [email protected]
Abstract – Using the speed voltage generators to measure the formed of the following components: emitters (LED),
rotary of a driving system is an advantageous solution for the receivers (photodetectors), rotating codewheel, stationary
analogical structures of the command circuit. For the numerical mask and the electronic circuit for the handling of the signals.
systems we will need to use an analogical-digital converter and The two wheels (rotating and stationary) are divided in the
for this reason these will be more useful and economical using the
numerical transducer. Trough this paper we propose a numerical
opaque and transparent zones. When the transparent zones of
rotary transducer which offers advantageous price/ both wheels are superposed, the receivers get the light flux
performances report. from the emitters and transmit an impulse to the output. It is a
The purpose of this application is to diagnose a process relative transducer that gives a number of specified impulses
automatically, with no human analysis involved. for one rotation of its ax, which depends on the number of the
transparent zones. This impulse rotation number will lead to
I. INTRODUCTION the transducer resolution that can be considered like an
analogical-digital converter of the position.
The control of a driving system of physical process
For a number of N rotation impulses the equivalent
presumes the obtaining of information about it, information
parameters of the position transducer will be:
that need to be worked on, memorized or communicated to a
-the variation beach of the output: N (bits);
higher level. The physical process is determined by physical
-the transducer resolution: [log2N] (bits);
values that can be transformed to electrical signals (usually
analogically by nature) by the transducer obtaining in this way -the precision of the transducer: 2π/N [rad/bit].
information about the process. We can work on this signals Because the encoder doesn’t give information about the
absolute position, we will use a special interface, basically a
using analogical technique or numerical technique case in
witch the analogical signals given by the transducers are numerical circuit, to count the impulses given by the encoder.
converted into numerical signals using special electronic At the same time, because using only one train of impulses
doesn’t allow the obtaining of the rotation sense, the encoder
circuits named data acquisition systems.
The speed transducer or rotary transducer will give give 2 trains of rectangular impulses at 90 degrees one against
information about the speed of the motor during an application the other. A sense detection circuit can be designed so that it
could, based on the notification of this difference phase
with the diving system. Usually we use a speed voltage
generator of direct current to give the information about the (advance or delay with 90 degrees of one signal against the
necessary speed. Because this transducer has an analogical other) to determine incrementing and decrementing of the
position info.
voltage at its output, proportional to the speed of the motor,
we have to use an analogical – digital converter if we want to Because this transducer is relative, which doesn’t point to
obtain the numerical value of this signal, used in a numerical the position modification facing the starting position of the
system. If we know the constant of the speed voltage generator, system, the encoders give another numerical signal, the also
kTG [V/rad/s], and the maximum value of the speed of the called zero signal of the encoder. This signal becomes active
giving a pulse by the length of an encoder step just one time
motor, Ωmax [rad/s], we can determine the maximum value of
on the rotation and it can be also used for the initialization
the tension at the output of the speed voltage generator which
after putting it under supply voltage so we can set an absolute
needs to be measured, VGmax = kG * Ωmax [V]. Usually, the position info of the system.
connecting scheme of the output of the speed voltage In the figure 1 we can see the shape of the signals given by
generator to the entrance of the analogical – digital converter
the encoder (A and B the two measurement canals; Z the
has to be adjusted (by using a simple resistive divisor) used to initial position canal) [2]. It is possible the using of each one
adapt the variation beach of this signal to the accepted values of the positive and negative edge of the signals A and B of the
beach from the entrance of the analogical-digital converter
encoder which allows the increasing by 4 times the precision
(usually +/- 10 V or +/- 5 V). The converter resolution in the of the encoder by implementing some adequate decoding
driving system applications is usually 10-12 bits. circuits.
The position transducer makes the measurement of the
driving system position. An advantageous solution is offered
by the utilization of the incremental optical encoder which
eliminated the analog digital converters [1]. The encoder is
373
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_65, © Springer Science+Business Media B.V. 2010
374 BORCOSI ET AL.
a)
b)
Fig. 4. The input signals (Vosen) and output (Vc) from the handling circuit.
Shaft
B. System needs
B A The determination of the sense and the value of rotation are
C made by implementing the logical diagram from figure 5.
Or c)
According to the logical diagram, first we attribute to the
rotation n and the variable v1 value zero, after that we check
Fig. 2. The element component of the transducer the A sensor. If A=0 we verify C, and if C=0 then we verify B.
In case that A=1 logical and C=1 logical we set the counter
We use infrared sensors so that the natural light doesn’t (which will start to implement the impulses with fi frequency)
influence the work. The three sensors A, B, C (figure 2.c) and the control variable v1 (which is used to approve the
transmit at the output Vosen=1 logical when they are evidence after a complete rotation). If follows B=1 and then
illuminated. The signal from the output of the sensors (Vosen) C=1 we made a complete rotation (v1=1) and because of these
is applied to a handling circuit (derivation) made with Schmitt succession of activating the sensors we can tell that the
Trigger inverters (figure 3) to eliminate the errors which transducer disk makes a rotation to the right, we calculate the
appear in the cases of slow fronts (at small rotations). In this value of the revolution n (we process the value to the Temp
way, at the illumination of every receiver, the handling circuit counter). After that we reset the contain of the counter and
generates a short impulse (the signal VC from the figure 3 prepare for a new rotation. If the succession of illumination of
which correspond to the C sensor). the sensors is C=1, A=1, B=1 and then C=1 it means that the
transducer makes a revolution to the left. When one of the
The A and B sensors are used to determine the rotation sense, successions of a rotation sense (left or right) is not kept, we
and the C sensor which obtains the value of the rotation which reset the counter and verify which the new succession is. To
sets up a counter which increments the tact impulses. verify the eventual stopping of the motor in the countering
The rotary will be determined using the formula: time, we inserted a procedure in the figure 6, where the current
n = (fi/ Tc)*60 = 1/(Ti*Tc)*60 [circ/min], value can be compared to the reference value T. This value (T)
where fi=1/Ti represents the frequency of the impulses tact represents the maximum time interval as long as the counter
applied to the counter and Tc the value of the counter (number can increment (Tcmax-preset=Tempmax-preset), between the activity
of increment impulses which are equivalent to a temporization of two sensors.
– Temp).
ROTARY TRANSDUCER 375
START
v1=0
n=0
a
1
*
0
c A
1 0
C * C 0
0 1 Reset counter
1 0 Set counter 1 B 0
v1 B
v1=1
Reset counter Reset counter
Set counter
v1=0 v1=0
v1=1 1
B
0
* d
Right rotation d 1 0
1 0
Processing Temp (n)
Reset counter
A C
*
Set counter 1 A 0
b a
0 1
1 0 v1
B * Left rotation
Set counter
v1=1 Processing Temp (n)
1 C 0 Reset counter
Set counter
Reset contor
v1=0
b 1 A 0 *
1 0
C
*
Reset counter
1 0 v1=0
d B
c
Fig. 5. The logical diagram.
*
It is chosen so that we can measure the rotation n nmin-preset, 1 0
where nmin-preset is better to have a small value, resulting for Temp <T
Tcmax-preset= Tempmax-preset a bigger value. We consider that the Reset
motor is stopped if Temp>T (n< nmin-preset) when the counter is counter
reset and it’s attributed the variable v1 and the rotation n has
zero value. *
Fig.6. The verification procedure for stopping the motor.
376 BORCOSI ET AL.
Microcontroller Display
P1.1
P1.2
P1.3
Date
VA
VB INT0 Address
VC
RX TX
Level Communication
Up Module
Fig.7. The processing circuits with microcontroller.
C. Making the system -it can be used for monitoring rotation by keeping the dates
Handling the signals given by the circuit in fig. 3 (VA, VB, in a data base or by showing it;
VC) which shows the sensors’ state, it can be made with the -it does need an A/D converter, in the case of a numerical
help of a microcontroller or the calculator through control system, can be cuter the handling circuit with a
implementation of the logical diagram in fig.5, according to microcontroller, its function will be implemented in the
this example we used a microcontroller type AT89C2051 control and command system. Actually we acknowledge a
(according to the figure 7) [4]...[6]. trend to realize the numerical exclusive systems with alone
The signals VA, VB, VC are connected to the specific input to microcontroller witch provides the all control and command
the external interruption INT0, and at the same time, one functions;
separately to another inputs P1.1, P1.2 and P1.3 which -on the disk we can make a lot of orifices. The real rotation
determine if the signal is activated. The connecting to the will be equal to the calculated rotation multiplied to the
INT0 input should be done using a logical gate OR which adds orifices number. Is preferred that the number of orifices to be
the three signals, but, considering that the interruption a multiple of 2, and the distance between them must not be
becomes active on the negative edge then the signals must be bigger than the distance between the A and B sensors;
negated and applying the principals of Boolean algebra in the - we used this rotation transducer and we recommended it to
place of OR gate and using the NAND gate. The timer/counter be used in drive systems with the DC motor controlled by a
function is made by counting the impulses with a frequency computer (using the parallel port) which has a single external
equal to the oscillator frequency divided by 12. The interrupt input. To this input we connect more signals for real
temporization frequency choosing the oscillator frequency time control, for example, synchronization input (for control
fosc=12 MHz is fi=1MHz and the period is Ti = 1μs. The in phase) and speed transducer outputs. The proposed speed
timer/counter circuit is on 16 bits. We can temporize transducer will enable rarely the interrupt input (for each of its
maximum: 216*Ti=65536*1μs=65536μs 65ms. signals, once at a complete rotation), compared with an encoder,
It is preferred to obtain a temporization of 50 ms, other wise at which the activation depends on the number of pulses.
if we want a bigger temporization which is easier to multiply. REFERENCES
These value of temporization result if we load the [1] Fl.Grofu, L. Popescu, Digital system for optical incremental encoder
temporization circuit with the initial value 15536, this connection, U.P.B. Sci. Bull., Series c, vol. 69, no. 3, 2007.
[2] Rotary Encoder, http://hades.mech.northwestern.edu/wiki/index.php.
temporization until the value 65536 and each time are touched /Rotary_Encoder.
these value a counter is incremented. [3] O. Olaru, I. Borcosi, M.C. Popescu, O. Tulpan, N. Antonie, The
To detect the stop of the motor in the timer/counter time we revolution measurement of a d.c. machine,The XIIth Edition of the
considered nmin preset=0,1 rot/s, which means that if in 10 International Scientific Conference of the Engineering Faculty,
November 23rd-24th 2007, Tg.-Jiu.
seconds a full rotation is not made, it means that the motor is [4] L. Toma, Sisteme de achizitie si prelucrare numerica a datelor, Ed. de
stopped and the timer/counter circuit is cleared, resulting for T Vest, Timisoara, 1996.
the value of 10 s. The temporization value is 10 seconds and is [5] D. Marinescu, Microcontrolerul 80C32 Manual de utilizare, Editura
obtained by measuring 20 times 50 ms. Tehnica, 1998.
[6] AT89C2051 and AT89C52 datasheet, available at: http://www.atmel.com.
III. CONCLUSIONS [7] Fl. Grofu, L. Popescu, C. Cercel, Acquisition System for Monitoring
Vibration,, CISSE 07 - IETA 07, University of Bridgeport, USA,
The measurement system of rotation designed and made by December 03-12, 2007, Published by Springer Press.
us has the following advantages: [8] I. Borcoi, O. Olaru, M.C. Popescu, A. Dinca, N. Antonie, M. Ionescu,
-it's easy and simple to make and offers advantageous report Device with Analogical Circuits for Protection to the Lack of the Pulse
price/performances because it has the easier processing digital for the Three-Phase Rectifiers in Electrical Drive, INTERNATIONAL
JOURNAL NAUN, Issue 4, Volume 2, December 2008.
system (a single microcontroller) than optical incremental
encoder [1] ;
Regression Analysis about Humidity Elimination
from Diesel Fuel Via Bioorganic Compounds to
Increase Antifouling Action
C.G. Tsanaktsidis1, N. Sariannidis2 and S.G. Christidis1
1
Laboratory of Qualitative Fuel Control
1
Department of Pollution Control and Technologies/ 2Department of Financial Applications
Technological Education Institute of Western Macedonia,
Kila, Kozani –Greece 50100
Abstract: In this paper we try to export a regression equation The selection of aspartic acid as a raw material was supported
which describes the variation of humidity of Diesel fuel with by the variation of humidity degree that these particle groups
time. A hydrophilic polymer is used to eliminate humidity from show in a water environment.[15]
diesel fuel and the regression analysis shows that the maximum
decrease of humidity in an hour it is of about ~ 35%.
Keywords: Regression Analysis, Diesel Fuel, Humidity Elimination
Bioorganic Compounds, Hydrophilic Polymer Antifouling Action.
I. INTRODUCTION
377
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_66, © Springer Science+Business Media B.V. 2010
378 TSANAKTSIDIS ET AL.
20 + 0.1 30 0.0635
20 + 0.1 60 0.0489
It must be referred that 20 mL is the demanded quantity of 6 -0.074 -0.095 6.0702 6 -0.021 -0.151 3.9683
Diesel fuel for the determination of humidity according to 7 0.147 -0.006 7.1333 12 0.031 -0.096 4.0153
method ASTM D-1744, while the quantity 0.1 g of polymer is 8 0.074 -0.158 7.6720 24 -0.005 -0.172 4.0176
defined from the limits of chemical additives in liquid fuels
which is 1-4 Kg/tn. Table II. LB Test for Residuals and Squared Residuals from the Regression Estimation.
Note: LB(n) are the n-lag Ljung-Box statistics for the residual series. LB(n) follows chi-square
C. Statistical Empirical Findings variable with n degree of freedom; the series of residual contains 9 observations .
The analysis of the regression estimation showed that: the [5] Jarque C. M, Bera A. K. (1987) A test of normality of observations and
regression residuals. International Statistical Review 55: 163-172
decrease of humidity is changed periodically with time and
comes to its maximum value at the first hour. [6] Engle R. F. Autoregressive conditional (1982) Heteroscedasticity with
The use of equation can predict the development of humidity estimates of the variance of United Kingdom inflation. Econometrica 50
values for important time periods and gives the moment for (4): 1987-1007.
the filtering of fuel so that the humidity is removed in the [7] Durbin J, Watson G.S. (1950) Testing for serial correlation in least
biggest degree. The elimination of humidity makes the fuel square regression. Biometrika 37: 409-428
used in the car machines and gives combustion with less
pollution for the environment. [8] Ljung G. M, Box G.E.P. (1978) On a measure of lack of fit in time series
models. Biometrika 65: 297-303
It is necessary to study the description via equations of the
phenomenon associates with volume of fuel, mass of polymer [9] The Presidential Green Chemistry Challenge Awards Program:
in order to export a total equation. Summary of 1996 Award Entries and Recipients; EPA744-K-96-001;
The result of elimination of humidity from fuel gives the U.S. Government Printing Office: Washington, DC, 1996
possibility to be rendered as antipollution at his use, so the [10] J. Clark and D. Macquarrie Handbook of GREEN CHEMISTRY AND
analysis of economic cost of the whole process is required TECHNOLOGY 2002
when it is applied in the industrial production.
Collaboration of sectors of science, organic chemistry, [11] Wheeler, A.P, Koskan L.P “Large scale thermally synthesized
polyaspartate as a substitute in polymer applications” Mat. Res. Soc
technology of fuels, statistical analysis for the export of Symp. Proc, 1993, 292,277.
conclusions in the qualitative control of fuels.
[12] J. Mosig, C.H Gooding, A.P Wheeler, “Kinetic and thermal
characterization on the hydrolysis of polysuccinimide” Ind. Enf. Chem.
REFERENCES Res. 1997. 36. 2163-2170
[1] Chuvieco E.; Riaño D.; Aguado I.; Cocero D., Estimation of fuel moisture [13] ASTM D1744-92 Standard Test Method for Determination of Water in
content from multitemporal analysis of Landsat Thematic Mapper Liquid Petroleum Products by Karl Fischer Reagent (Withdrawn 2000)
reflectance data: applications in fire danger assessment, publisher
Taylor and Francis Ltd, International Journal of Remote Sensing, Vol. [14] ASTM D2276 - 06 Standard Test Method for Particulate Contaminant
23, No 11, 2002 , pp. 2145-2162(18). In Aviation Fuel by Line Sampling ,2006
[2] MINGHOU XU*, YAOGUO FAN, JIANWEI YUAN, CHANGDONG [15] .. roganis, C.Tsanaktsidis and I.P. Gerothanassis, 14N NMR
SHENG AND HONG YAO, A SIMPLIFIED FUEL-NOx MODEL relaxation times of several protein amino acids in aqueous solution-
BASED ON REGRESSION ANALYSIS, INTERNATIONAL JOURNAL OF comparison with 17O NMR data and estimation of the relative hydration
ENERGY RESEARCH, Int. J. Energy Res., 23, 157Ð168 (1999) numbers in the cationic and zwitterionic forms, J. Magn.Reson.,
2003,164, 294
[3] David Carter, Daniel A. Rogers, Betty J. Simkins, Does Fuel Hedging
Make Economic Sense? The Case of the US Airline Industry, AFA 2004 [16] S.G. Christidis, A.I Maroulis, K.Chatziantoniou -Marouli ,
San diego Meetings, September 16, 2002. A.Tsiomlektsis, Aristotle University of Thessaloniki. «The experiment
in teaching of Chemistry. Green Approach. Composition and
Applications of Thermal Polyasparte Anion.». Proceedings of 20th Pan-
[4] Gerfried Cebrat, Avraam Karagiannidis and Agis Papadopoulos, Hellenic Conference of Chemistry. Ioannina September 2005.
Proposing intelligent alternative propulsion concepts
contributing to higher CO2 savings with first generation
biofuels, Management of Environmental Quality: An International
Journal, Vol. 19 No. 6, 2008, pp. 740-749
Optimal Threshold for Locating Targets Within
a Surveillance Region Using a
Binary Sensor Network
Zhenxing Luo and Thomas C. Jannett
Department of Electrical and Computer Engineering
The University of Alabama at Birmingham
Birmingham, AL 35294
Abstract - This paper considers the design of the optimal thre- optimal threshold for binary sensors detecting targets having a
shold for locating targets within a specified surveillance range range of powers and locations within a specified surveillance
using a network of binary sensors. The threshold was determined region. The paper is organized as follows. Since this paper
by minimizing a cost function representing the summation of the
variances of estimation errors for the x and y coordinates of the extends the work in [1], Section II closely follows the presen-
target. The cost function was evaluated using multiple integra- tation and notation in [1]. Section II presents the sensor mod-
tion over the specified ranges for target power and target loca- el, exemplary network layout, and maximum likelihood esti-
tion. A geometrical interpretation of the optimal thresholds for mator for quantized data. Section II also reviews a method for
targets with known power and location is presented to offer in- designing optimal quantization thresholds for targets with
sight into the problem. The optimal threshold was validated in known power and location. Section III provides a geometrical
Monte Carlo simulations using a field of sensors having a uni-
form grid layout. In simulations, the summation of the variances interpretation of the optimal thresholds for targets with known
of the estimation errors achieved using the optimal threshold power and location. Section IV develops the optimal threshold
approached the minimum of the cost function. for binary sensors locating targets having a range of powers
and locations within a specified surveillance region. Methods
I. INTRODUCTION for computing the optimal threshold are presented and demon-
T ARGET localization is an important and challenging prob- strated. The optimal threshold is validated and evaluated in
lem in sensor networks [1]-[4]. In a network or network Monte Carlo simulations. Discussion and conclusions are pre-
region, sensors transmit data to a central processing node or sented in Sections V and VI.
fusion center. For the target localization problem, the
processing node receives data from each sensor and estimates II. PROBLEM FORMULATION
the target location. Some methods are based on time differ- We consider an exemplary field of N = 441 sensors un-
ence of arrival (TDOA) [5], [6]. However these methods re- iformly deployed with sensors placed 9 m apart in each of the
quire accurate timing, which is not feasible with inexpensive x and y directions (Fig. 1). Many ideas presented in this paper
sensors. Energy-based methods, which use signal strength apply for other sensor deployments, including fields in which
information, are more practical in real situations [7]-[10]. sensor are placed at random locations.
Much recent work has been focused on maximum likelihood The target emits a signal, with the signal intensity decreasing
target localization in networks of inexpensive sensors that as the distance from the target increases. Following the pres-
have limited energy and communications resources [1], [9]. entation in [1], we consider a signal power model
When resources are limited, it is desirable for the sensors to
transmit quantized data to the processing node [1], [11]. A Gi P0'
maximum likelihood estimator that uses quantized data and ai2 = (1)
( di d0 )
n
381
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_67, © Springer Science+Business Media B.V. 2010
382 LUO AND JANNETT
100
0 −∞ < si < ηi1
° 1 ηi1 < si < ηi 2
80
°°
Di = ® . (6)
60
°L − 2 ηi ( L − 2) < si < ηi ( L −1)
°
40
°̄ L − 1 ηi ( L −1) < si < ∞
Y-coordinate (m)
20
0
The probability that Di takes specific value l is
-20
)& η − ai η −a
pil (ηi , θ ) = Q ( il ) − Q ( i ( l +1) i ) (0 ≤ l ≤ L − 1) (7)
-40 σ σ
-60
where Q (⋅) is the complementary distribution function of the
-80 standard Gaussian distribution
-100 2
-100 -80 -60 -40 -20 0 20 40 60 80 100 ∞ 1 − t2
X-coordinate (m) Q ( x) = ³ e dt . (8)
Fig. 1. Sensor field layout showing 441 sensors placed 9 m apart.
x
2π
where (xi , yi ) and (xt , yt ) are the coordinates of sensor i and After collecting the data, D , the processing node estimates
the target, respectively, and n is the power decay exponent. the parameter vector θ = [ P0 xt yt ]T by maximizing the log-
We assume that Gi = G for i = 1,...,N , such that P0 = GP0' . In likelihood function of D
this paper, we use n = 2 and we assume that the target is at N L-1
&
least d 0 = 1 meters away from any sensor at all times, result- lnp(D ) = ¦¦ (Di - l)ln [ pil (i ,)] (9)
ing the simplified signal power model
i=1 l=0
where
P0
ai2 = . (3) 1, x = 0
d i2 (x) = ® . (10)
¯0, x = 0
The measured signal at the ith sensor is modeled as
In summary, the maximum likelihood estimate of
si = ai + wi (4) θ = [ P0 xt yt ]T is determined through the optimization prob-
lem:
for i = 1,...,N where wi is a Gaussian noise
max lnp ( D ) . (11)
wi ~ N ( 0, 2
). (5)
The maximum likelihood estimate (11) and the corresponding
Fisher Information matrix (FIM) and CRLB for the estimation
At each sensor, the measured signal si is quantized and
error variance are derived in [1], which also addresses the
transmitted to the processing node. design of sensor decision thresholds based on the CRLB using
As described in [1], the processing node receives quantized both optimization and heuristic methods.
multi-bit data from each sensor. The received data are In this paper, we considered binary decisions made at each
D = {Di : i = 1,...,N } , where Di can take on any discrete val- sensor that yield binary data
ue from 0 to 2 M - 1 . Using the definition L = 2 M , the set of
quantization thresholds for the ith sensor is 0 −∞ < si < η
& Di = ®
i = [i0 ,i1 ,...,iL ] , where ηi 0 = −∞ and ηiL = ∞ . The ith sen- ¯1 η < si < ∞
. (12)
sor supplies quantized data
OPTIMAL THRESHOLD FOR LOCATING TARGETS 383
A. Optimal Quantization Thresholds for Fixed Power and in a surveillance region and then examine the patterns of sen-
Fixed Target Location sors that would fire if the optimal thresholds are applied.
The CRLB for the ML estimate (11) is a function of the For P0 =10000 and binary decisions made at each sensor, op-
quantization thresholds and was used in [1] as the basis for the timal thresholds were determined over a subsection of the
cost function field (Fig. 1) represented by a 46 by 46 grid of target locations
with x = 0.0, 0.2,…, 9.0 and y = 0.0, 0.2,…, 9.0. For this grid
& & & & &
&
V( ) =
() ()
J 11 [J 33 + J 22 (& ) ] - J ( ) - J ( )
2
13
2
12
(13)
of target locations, the optimal threshold (14) varies symmetr-
ically with target location from a minimum of 6.89 to a maxi-
J () mum of 11.24 (Fig. 2). For a target located far away from any
sensor, the threshold is lower than for a target located near a
which represents the summation of the variances of the loca- sensor. The optimal threshold sometimes changes abruptly
tion estimation errors in the x and y directions. In (13), J ij is from one target location to another.
the element on the ith row and jth column of the FIM, J .
Expressions for J ij are presented in [1] and will not be repro-
duced here. The cost (13) is an obvious choice for designing a
threshold that provides high accuracy in the estimation of the
target location, although the resulting estimate of the power,
P0 , may not be accurate. For a known P0 the optimal thre-
Threshold
10
-5 -5
X-coordinate (m)
For a specified target location and power, this method gives Y-coordinate (m)
the threshold that is optimal for use at all sensors in the field.
a) Optimal thresholds over a field subsection having
However, the method does not address the practical problem vertices [-9, -9], [9, -9], [-9, 9], and [9, 9].
of choosing the threshold needed to locate targets having un-
known power and unknown positions within a surveillance
region. Since the method for threshold optimization (14) de-
pends on knowledge of the parameters θ = [ P0 xt yt ]T to be
estimated, different thresholds would be required for different
target locations within a surveillance range of interest, and the
thresholds are optimal only for a specific P0 . However, the
Threshold
nary decisions are at relatively low, medium, and high levels. range of target locations, then small changes in target location
We demonstrate the impact of the different thresholds by will not produce different firing patterns, resulting in poor
showing the patterns of sensors that fire (detect a target) for identifiability (uniqueness of solution) and possibly, biased
these thresholds at P0 =10000 (arbitrary units) in the noise estimates. If the power is significantly lower than the power
free case (σ = 0). For a target located in the center of a grid of for which the threshold is optimal (power mismatch), few or
four sensors, the target is located as far away as possible from no sensors may fire, resulting in large target location estima-
any of the four sensors, the threshold is low, and four sensors tion errors [1].
fire (Fig. 3a). For a target located an equal distance between Finally, we examine the consequences of applying subop-
two sensors, the optimal threshold is at an intermediate level timal thresholds. If the applied threshold is higher than the
and six sensors fire (Fig. 3b). For a target located close to a optimal threshold, few or no sensors may fire, resulting in
sensor, the threshold is relatively high and two sensors fire large estimation errors. If the applied threshold is lower than
(Fig. 3c). the optimal threshold, more sensors may fire than with the
optimal threshold, perhaps with little consequence. For target
30
locations near the border of the sensor field, a low threshold
Sensor Field Y-coordinate (m)
25
that results in a large pattern of sensors firing may produce an
20 asymmetric firing pattern that introduces estimation errors.
15 Further study of the log-likelihood function (9) and FIM or
10
CRLB may provide insight into the relationship between per-
formance, thresholds, and sensor firing patterns.
5
0
0 10 20 30
IV. OPTIMAL THRESHOLDS FOR A RANGE OF POWERS
Sensor Field X-coordinate (m) OVER A SURVEILLANCE REGION
a) Target location (13.5,13.5) and optimal threshold 6.89.
The minimization in (14) provides a threshold that is optim-
30
al only for a known P0 and specified target location. However,
Sensor Field Y-coordinate (m)
(η,θ )
t t 0
J
Fig. 3. Demonstrative target locations (star) and corresponding patterns of
fired sensors (solid circles) that occur for the optimal thresholds found using (15)
(14) if there is no measurement noise. For a), b) and c) the lower left sensor is computed by integrating over .
corresponds to location (0,0) in the field of Fig. 1. (P0=10000, binary deci- The optimal threshold (14) designed for a specific power
sions, σ = 0).
performs poorly for mismatched P0 values [1]. When the ac-
Inspection of the sensor firing patterns for the optimal thre- tual power is lower than the value of P0 used to calculate the
sholds (Fig. 3) facilitates assessment of the impacts of noise optimal threshold, few or no sensors fire, and the location
and power mismatch. Noise may make more or fewer sensors &
fire, resulting in asymmetric sensor firing patterns that pro- ()
variance V computed using (13) becomes very high.
duce estimation errors. If the thresholds are the same over a Therefore, we expect that the cost (15) will be dominated by
OPTIMAL THRESHOLD FOR LOCATING TARGETS 385
the contributions of the low target powers. In a uniform sensor veillance region. The cost (16) was evaluated over a P0 set
field, for a fixed power, the optimal thresholds (Fig. 2) found [8000 10000 12000] by exploiting symmetry to approximately
using (14) for specific target locations and the corresponding evaluate (16) for the surveillance region based on the grid of
&
()
values of V follow a symmetric pattern within the field. 81 target positions described earlier (Table I). The minimiza-
tion (17) was carried out using exhaustive search. In consider-
An understanding of the sensitivity of the optimal threshold to ation of Table II in [1], the cost was expected to be dominated
mismatched low P0 values and recognition of the symmetry in by the contributions from the low powers within the P0 inter-
&
()
V suggest that approximating the cost (15) as val. The cost was evaluated for P0 within the interval from
8000 to 12000 using a P0 set having only three elements be-
& & & & & cause using more elements would be expected to increase the
&
V ( ,) = ¦¦¦
( ) ( )
J 11 , [J 33 , + J 22 ( ,
&
) ] - J ( , ) - J ( , )
2
13
2
12
(16)
computational load without much improvement in the solu-
xt yt p0 J ( , ) tion. In addition, for P0 within the interval from 9000 to
11000, the cost (16) was minimized using the same method
for a P0 set [9000 10000 11000] (Table I).
may give useful results and avoid a prohibitive computational
load in finding the optimal thresholds that are the solutions to B. Evaluation of Thresholds in Monte Carlo Simulations
& The optimal thresholds were evaluated in Monte Carlo si-
( )
& V , θ .
min (17) mulations (1000 runs, binary decisions, n=2, σ = 1) in which
targets were placed at random locations uniformly distributed
TABLE I Random
OPTIMAL THRESHOLD AND COST VERSUS THE SET OF POWERS USED FOR - 7.481 - 4.92
(8000-12000)
OPTIMIZATION OVER A SURVEILLANCE REGION
& 10000 7.91 6.892 4.05 4.73
Power, P0 Optimal Threshold, Cost, V ( ,)
1
8000 7.11 4.38 7.48 is the optimal threshold found for P0 set [8000 10000 12000] in (16).
2
6.89 is the lowest of the optimal threshold values found for any target loca-
10000 7.91 4.05 tion within the surveillance region (Fig. 2a).
387
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_68, © Springer Science+Business Media B.V. 2010
388 BRIONES ET AL.
2 § AU ·
x1 = ¨ F1 ( x1e − x1 ) − i i Δx ¸¸ I/P TC
M 1 ¨© Cp1 ¹ (1)
hot fluid
2 § AU ·
x 2 = ¨¨ F2 ( x 2 e − x 2 ) + o o Δx ¸¸
M2 © Cp 2 ¹ controlled flow
TT
dxˆ 2F
The L operator is declared by [3]: 1= 1 ( x1e − xˆ1 ) +
dt M
1
n ∂h( x )
f ( x); L h ( x ) = L §¨ L h( x ) ·¸( x )
k k −1
L h( x) = ¦ f© f M ( 2 F k ( x − xˆ ) + M k ( x − xˆ ))
f
i = 1 ∂xi
i f ¹ 1 2 1 1 1 2 2 2 2
n ∂ § n ∂h( x )
¨ ¦
· 2 ( F M − F (t ) M )
L L h( x) = ¦ g ( x ) ¸ f ( x) 2 1 1 2
f g ¨
j = 1 ∂x j © i = 1 ∂xi
i ¸ i
¹ (5)
dxˆ 2F
2= 2 ( x2e − x2 ) +
Thus the relative degree of the system (1) is equal to 2, and dt M
due to its coincidence with the number of states in the 2
system (1), then the system (1) has a complete relative M ( 2 F k ( x − xˆ ) + M k ( x − xˆ ))
2 1 1 1 1 1 2 2 2
degree and this guaranties that y can be used to describe the
uncertain term and that there are no unmodeled dynamics. 2( F M − F M )
1 2 2 1
The consideration of a complete relative degree for the V. ANALYSIS OF THE DYNAMIC SYSTEM.
system (1) guaranties the existence of a stable nonlinear
observer, that will converge in its dynamic behavior with the The analysis consists of the comparison of the
system (1) when time has an infinite limit [4]: numerical simulation between the dynamic behavior of the
systems (1) and (5) in concurrent and countercurrent fluid
flows.
dxˆ −1
= f ( xˆ ) + g ( xˆ )u + [Q ( xˆ )] K [ y − h( xˆ )]
dt (4) In the system (5), the observation gains locate the
y = h( x) characteristic polynomial poles in -1, where this
characteristic polynomial is defined by the error system
formed by the difference between the systems (5) and (1),
Where the superscript ˆ denotes the observed states and
and it allows for the dynamic behavior of the system (5) to
K=[k1,k2]T∈ℜ2 is the set of observer gains with magnitude tend to the dynamic behavior of system (1), when the time
and sign such as the states from system (4) tend to the states tends to infinity.
from the system (1) when time tends to infinity and the
Jacobian matrix Q(x) is defined by: Apart from what has been mentioned before, it is necessary
to consider only known nominal values of the parameters
h( x) involved in the systems (1) and (5), and they are calculated
L h( x) at the heat exchange initial conditions in the equipment.
f These values are defined in Table 1.
Q( x) =
TABLE I
Ln − 1 ( x ) NOMINAL VALUES USED IN THE NUMERCIAL SIMULATIONS IN
f
CONCURRENT AND COUNTERCURRENT HEAT TRANSFER
Parameters Heat exchanger regime
However, with the use of the vector maps f(x) and g(x) given Concurrent Countercurrent
by the system (3) onto the system (4), besides the Jacobian Ui [=] kcal/m2 ºC hr 115.033 100.244
matrix above defined in terms of the Lie algebra and Uo[=] kcal/ m2 ºC hr 103.286 89.653
Ai[=]m2 0.2736
executing the rules associated to this kind of algebra, the Ao[=]m2 0.3040
nonlinear observer is: F1[=]kg/hr 11.279
M1[=]kg 0.767
M2[=]kg 3.258
Cp1[=]kcal/kg ºC 0.999
Cp2[=]kcal/kg ºC 0.997
x1e[=]ºC 90
x2e[=]ºC 29
390 BRIONES ET AL.
300 (a
90 (a
Nominal System 240
80 mean temperature difference
Observer System 180
70 calculated by the observer in the inner tube
120
60 60
Temperature (°C)
50 0.0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1.0
Temperature (°C)
40
300
30
240
(b
0.0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1.0 180 mean temperature difference
50 calculated by the observer in the annular space
120
(b
45 60
0.0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1.0
40
Nominal System 60
35 (c
Observer System 50
30 40 mean temperature
30 difference calculated by formula
25
20
0.0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1.0
Time (hr) 0.0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1.0
Time (hr)
Fig. 2. Numerical simulation of heat exchange in countercurrent regime. a) Fig. 5. Numerical estimation of mean temperature difference. a) reference
Hot fluid temperature. b) Cold fluid temperature. by observation in the inner tube. b) reference by observation in the annular
space. c) definition of logarithmic mean temperature difference.
180 (a
150 mean temperature difference
120 calculated by the observer in the inner tube Figure 2 and Figure 4 show that system (1) and (5) have an
90
60 equal dynamic behavior in terms of temperature evolution
0.0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1.0 over time, despite the fact that system (5) uses only mass
and flow data from the equipment, meanwhile system (1)
Temperature (°C)
180 (b
150 mean temperature difference needs more information, particularly the mean effective
120
90
calculated by the observer in the annular space
temperature difference and the overall heat transfer
60
coefficient.
0.0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1.0
60
(c However, the nonlinear observer (5) has a better description
50
40 mean temperature of hot temperature than cold temperature regardless of the
30
20
difference calculated by formula
heat exchange regime, because the difference between the
0.0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1.0
real and the estimated hot temperature is practically zero,
Time (hr) meanwhile the difference between the real and estimated
Fig. 3. Numerical estimation of mean temperature difference. a) reference cold temperature is 1°C.
by observation in the inner tube. b) reference by observation in the annular
space. c) definition of logarithmic mean temperature difference.
Thus the nonlinear observer (5) is a promissory dynamic
90
Nominal System
(a model in terms of describing the hot temperature in the
80
70
Observer System double pipe heat exchanger using a minimum of
60 information, and in physical applications, the hot
50 temperature has more importance than the cold temperature,
40
because the first one is commonly the variable that is
Temperature (°C)
30
intended to control in a process.
0.0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1.0
50
(b
45 Figure 3 and Figure 5 show the dynamic evolution of the
40 mean temperature difference, this way, the observer system
35
Nominal System describes the mean temperature difference in the equipment,
Observer System
30
regardless of heat exchanger regime, and this information is
25
complementary to the measure of the maximum quantity of
0.0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1.0 heat that the fluids can exchange.
Time (hr)
Fig. 4. Numerical simulation of heat exchange in concurrent regime. a) Hot This is an extra value, because this kind of nonlinear
fluid temperature. b) Cold fluid temperature. observer allows user to isolate the principal uncertainties in
the heat transfer process (the mean effective temperature
difference and the overall heat transfer coefficient), and
eventually describes the properties of the equipment.
A PROPOSAL OF A NONLINEAR OBSERVER 391
So, even if there is a logarithmic mean effective temperature Even more, this nonlinear observer allows the user to
difference (Figure 5c), this parameter only describes the estimate the temperature inside the equipment and uses this
capacity of temperature that can be transferred between estimation for other applications besides control process
fluids (approximately 20°C), but the estimation of this (e.g. improve de the design of double pipe heat exchanger),
parameter obtained from the nonlinear observer, describes and thus, it has with wide possibilities of being physically
the internal temperature of the equipment (about 80°C), and implemented.
this information is more useful if one of the process
objectives is controlling the inner temperature in the REFERENCES
equipment, to avoid the formation of fouling in the inner
tube and the annular space and this measure is very [1] Yunus A. Çengel, Heat and Mass Transfer – A Practical
Approach. Singapore, McGraw-Hill, 2006, pp.619-626.
important, because it is much too difficult to install a set of [2] Y.S.N. Malleswararao and M. Chidambaram, “Nonlinear
temperature sensors inside the equipment to obtain this controllers for a heat exchanger”, J. Process Contr.Vol. 1, No.2,
information, which can be estimated by the proposed pp. 17–21, 1992.
nonlinear observer. [3] Alberto Isidori, Nonlinear control systems: An introduction,
Germany, Springer, 1989, pp. 145-150
[4] H. Murata et. al., “Control of inverted pendulum-wheeled cart
VI. CONCLUSIONS system using nonlinear observers”, SICE’95 Proceedings of the
34th SICE Annual Conference. International Session Papers, Vol.
The proposed nonlinear observer is a promissory dynamic 26-28, pp. 1133-1136, July 1995.
[5] A. Zavala-Rio et. al., “Bounded positive control for double-pipe
model to be used in the synthesis of advance control heat exchangers”, Control Engineering Practice, Vol. 17, No. 1,
strategies, in which the only variables known are flows and pp.136-145, January 2009.
masses in the equipment. [6] C.-M. Astorga-Zaragoza et. al., “Performance monitoring of heat
exchangers via adaptive observers”, Measurement, Vol. 40, No.
4, pp. 392-405, May 2007.
While there are other observers previously proposed [7] C.-M. Astorga-Zaragoza et. al., “Observer-based monitoring of
([5,6,7]), they use two critical parameters (the mean heat exchangers”, ISA Transactions, Vol. 47, No. 1, pp. 15-24,
effective temperature difference and the overall heat transfer January 2008.
coefficient), that can be estimated from experimental [8] Prabhata K. Swamee et. al. “Optimum design of double pipe heat
exchanger”, International Journal of Heat and Mass Transfer ,
correlations with some margin of error. The nonlinear Vol. 51, pp. 2260–2266, 2008.
observer hereby proposed does not use them and it even [9] T.W: Fraser Russell, et. al., Mass and Heat Transfer – Analysis
estimates them with a minimal margin of error than the of mass contactors and heat exchangers, United States of
experimental correlations estimate. America, Cambridge University Press, 2008, pp. 90-102.
Vibration Signal Processing
by Deconvolution Method
Popescu Ion Marian
University Constantin Brâncu i Târgu-Jiu, Romania, [email protected]
Runceanu Adrian
University Constantin Brâncu i Târgu-Jiu, Romania, [email protected]
I. INTRODUCTION
393
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_69, © Springer Science+Business Media B.V. 2010
394 MARIAN AND ADRIAN
III. PROCESSING THE ACQUIRED VIBRATION SIGNAL As you can see, most of the vibration signal energy is at
frequencies above 1 kHz or even 10kHz.
Using data acquisition system [1] described above and The power spectrum for the “good” bearing is 20 times
generate vibration platform were obtained the following signals lower than the power spectrum for the “spent” bearing. As the
acquired for a bearing "good" and a bearing “spent" as in Fig.8. motor that rotates the mechanism of the 2 bearings that are
fitted vibration sensors have 1500rpm, i.e. around 25Hz, it can
be concluded that in this signal is included vibration and
effective response vibration sensor which has a important
weighting.
³
x r (t ) = x1 (τ )h(t − τ )dτ (7)
0
Returning to equation (2), we have:
.. . ..
m x r (t ) + c x r (t ) + kx r (t ) = −m x1 (t ) (8)
Starting from the premise that mechanical vibration of the
structure is small displacement around a position of
Fig.9. Vibration signals for the 2 bearings consist equilibrium which is also a stationary point for the system,
1000 samples approximating the vibration transducer as a linear system is
Processing was realized using the development environment valid.
LabWindows/CVI 8.1 and processing was done for the So, we can apply the Laplace transform and we have zero
acquired signal from the charge amplifier, that is connected initial conditions:
accelerometer. The signal is a signal of acceleration (vibration) ms 2 X r ( s ) + csX r ( s ) + kX r ( s ) = −ms 2 X 1 ( s ) (9)
of the mechanical system.
Mathematical model (transfer function) of the vibration
We made for these 2 signals the power spectrum (using
transducer is:
Fourier transform) and obtain the processing of Fig.10.
VIBRATION SIGNAL PROCESSING BY DECONVOLUTION METHOD 397
X r ( s) − ms 2
H ( s) = = 2 (10) After processing relationship (12), we have:
X 1 ( s ) ms + cs + k
X ( z)
Using the Tustin substitution: H ( z) = r = (13)
X1( z)
H ( z) = H ( s) 2 z −1 (11) − 4m + 8mz −1 − 4mz − 2
=
s=
T z +1 (4m + 2Tc + kT ) (
2
) (
+ − 8m + 2kT 2 z −1 + 4m − 2Tc + kT 2 z − 2)
4 (z − 1)2 Further we note this:
−m b1=-4m; b2=8m; b3=-4m;
T 2 (z + 1)2
H ( z) = (12) a1=4m+2Tc+kT2; a2=-8m+2kT2; a3=4m-2Tc+kT2
4 (z − 1)2 2 (z − 1) We obtain discrete transfer function:
m +c +k
T 2 (z + 1)2 T (z + 1)
b + b z −1 + b z −2 X ( z)
H ( z ) = 1 2 −1 3 − 2 = r (14)
a1 + a 2 z + a3 z X 1( z)
Equation (14) can be implementated by considering the
delay operator z-1:
a1 x rk + a 2 x rk −1 + a3 x rk − 2 = b1 x1k + b2 x1k −1 + b3 x1k − 2 (15)
where:
xrk- is the signal of the output vibration transducer
(transducer response)
x1k- is the signal of the input vibration transducer
(real vibration of bearing)
Real vibration signal for the mechanical system after this
processing is:
b b a a a
x1k = − 2 x1k −1 − 3 x1k _ 2 + 1 x rk + 2 x rk −1 + 3 x rk − 2 (16)
b1 b1 b1 b1 b1
Because the transducer transfer function from relation (10) is
a proper function, can be implemented numerically as
relationship (16), we obtain the signal of input (real vibration
of mechanical structure) of the transducer, knowing output
signal of transducer and the mathematical model of the
transducer. In this way, is made a digital deconvolution, which
is valid given that the transfer function is proper.
In fig.11. first 2 graphs are the response of the transducer to
impulse mechanically generated and the next 2 graphs
represent the response of the mathematical model simulated.
Power spectrum obtained for the new signal x1 is presented
in fig.12. for the 2 bearings. As you can see, most of the signal
energy is in the low frequencies.
Although the 2 vibration signals are acquired on different
channels from acquisition board and from different bearings,
can be observed a correlation of spectral components
(highlighted by green dotted lines) in relation to frequency and
amplitude compared with the notes that the bearing used (2nd
graph) has some component of 200 times the bearing
something better (first graph).
This method, with digital deconvolution, introduces large
errors primarily due to initial conditions is not identical with
real conditions.
In the graphs presented, after the processing by the relation
Fig.11. Rectangular impulse response for the real system and (16) vibration signal, has eliminated a total of 100,000 samples
simulated system from 400,000 before taking a Fourier transform. By this was
removed from the calculated signal variation due to the initial
conditions.
398 MARIAN AND ADRIAN
Another error that affects processing is not known with structure "good", and this process taking as reference for
precision mathematical model of the vibration transducer and comparison to other similar structures in operation.
in particular the coefficients c and k which will have
REFERENCES
corresponding n i . In this case it was considered
m=0.002; =0.1; n =45000;
[1] Grofu F., Popescu L. and Cercel C. (2008)
“Acquisition System for Monitoring Vibrations”,
International Joint Conferences on Computer,
Information, and Systems Sciences, and Engineering
(CISSE 2007) Conference, Conference Proceedings
book, December 3-12, 2007, University of Brigeport,
USA, publish in Novel Algorithms and Techniques In
Telecommunications, Automation and Industrial
Electronics, ISBN 978-1-4020-8736-3 (Print) 978-1-
4020-8737-0 (Online), pg. 89-92, DOI 10.1007/978-1-
4020-8737-0_16.
[2] Shin K., Hammond J., “Fundamentals of Signal
Processing for Sound and Vibration Engeneers”, John
Wiley&Soons Ltd., 2008, ISBN 978-0-470-51188-6
[3] R.Keyth Mobley,“An Introduction to Predictive
Maintenance”, 2nd Edition, 2002, Elsevier Ltd.,
ISBN: 978-0-7506-7531-4
[4] Mateescu A.,Dumitriu N., Stanciu L. “Semnale si
sisteme-Aplicatii in filtrarea semnalelor”, Ed.Teora,
2001, ISBN:973-20-0666-8
[5] Buzdugan G., Fetcu L., Rades M. “Vibratiile
sistemelor mecanice”, Ed. Academiei Romane
Bucuresti, 1975.
Fig.12. Power spectrum obtained after deconvolution [6] Moheimani, S.O. Reza, Fleming, Andrew J.,
“Piezoelectric Transducers for Vibration Control and
Damping”, Series: Advances in Industrial Control,
IV.CONCLUSIONS 2006, Springer-Verlag, ISBN: 978-1-84628-331-4
[7] Harris C.M.; Piersol, A.G. “Harris' Shock and
In an automation system, digital signal processing of Vibration Handbook”, (5th Edition), 2002, McGraw-
vibration is a very difficult problem which must take into Hill, ISBN 0-07-137081-1.
account the following aspects: [8] Douglas Thorby “Structural Dynamics and Vibration
> Acquisition system must meet the powerful features of real in Practice”, 2008, Elsevier Ltd., ISBN: 978-0-7506-
time, because the vibration signal is a very dynamic signal. 8002-8.
> Another problem is strongly nonlinear nature of a vibration [9] Dan Necsulescu, “Advanced mechatronics:
signal, which complicates its processing. Most mechanical monitoring and control of spatially distributed
systems that generate vibrations have imperfections in their systems”, 2008, World Scientific, ISBN 9812771816,
structures, to which adds that the vibration in a mechanical 9789812771810
structure have more degrees of freedom and transmission [10] www.leonovabyspm.com, Leonova Infinity-
through the structure. Instruction Manual, 2008
> Vibration transducers have limited responses and introduce [11] http://www.ni.com/soundandvibration/, Sound and
distortions of amplitude and phase signal provided. vibration systems, 2009.
> Any acquisition system (including transducer) and signal [12] Honeywell Sensotec, Accelerometers Frequently
processing of vibration require laborious calibration platforms Asked Questions, www.sensotec.com, 2003
specially designed for a specific type of application. [13] Maxim-IC, App Note 604: Fast Memory Transfers
> Structures analysis and on-line processing of vibration with the Ultra High-Speed Microcontroller -
signals are very difficult to achieve because in a mechanical DS89C420,
structure, the point where vibration measurements is made, is [14] I.J.Taylor. ”Identification of Bearing Defects by
far from processing system, and networks cable introduce Spectral Analysis”. Journal of Mechanical Design,
additional distortions Transaction of ASME, vol. 120, 1980, pp 199-204.
As a conclusion, the most applications for acquisition and [15] Vibration Measurement and Analysis
processing of vibration in purpose of predictive maintenance of Bruel & Kjaer Handbook BA 7676-12, 1998
a mechanical structures, processing the signal for a mechanical
An IT Change Management Tool to Optimize
Scheduling and Planning
Denílson Cursino Oliveira, Raimir Holanda Filho
University of Fortaleza- UNIFOR
[email protected], [email protected]
Abstract – The IT governance has become strategic to dependencies and the rules to build an optimized schedule
business objectives. The main reason is the increase of of changes based in one change window. The main
importance that IT has in business decision. To achieve this feature of this model is the possibility of building a
importance, the IT management must ensure a Just-in-time parallel implementation of changes using a simpler
feedback, electing the change management as main point of IT
to help the business goal through the use of tools with optimized
heuristic with same complex set of changes if compared
procedures to guarantee a lower financial loss during the with the state-of-art.
implementation of changes. This work presents a tool to In [4] was added some resources to change
minimize the impact of the changes on the business goals, management tool: more dependency treatment as the
respecting the dependencies between the changes and featuring principle cross dependency; the allocation of changes in
innovations as dynamic change window determination. The tool different change windows; the automated determination
presented in the work builds an optimized determination of of time implementation of each family of changes,
change windows, treatment of dependencies and allocation of enabling the determination of the financial loss.
changes, helping the business to minimize the impact. In [5] is presented an optimization of IT changes,
showing mathematical concepts to determine the financial
I. INTRODUCTION AND MOTIVATION loss, cost of changes implementation over a set of
The actual IT governance aligns the real goals of changes attached to a change window. One change
the business with an automated response information management conceptual model of a decision support tool
technology to achieve strategic goals with more is presented. The implementation of changes is performed
efficiency and less time. To formalize the IT governance sequentially. On the other hand, our model does this in
globally, some set of procedures and good practice have parallel.
been developed. Control Objectives for Information and The CHAMPS [6] focuses on optimization of a
related Technology (COBIT) [1] is one of them that scheduling of changes to satisfy external RFC´s.
promotes a management and control of information with CHAMPS develops a complex dependence hierarchy that
use of measures, performance indicators and maturity is used to build a final scheduling, where it reflects a least
models. Other example is the IT information library financial impact during the changes. Our tool has a
(ITIL) [2], which consists of a set of best practices that simpler but powerful heuristic for the construction of the
determine how the IT infrastructure of a company is dependences between changes and their treatments. Also
organized. The library ITIL is an IT service management it builds an optimized scheduling based on financial or
framework that promotes an efficient management of IT timeline variable.
changes in order to minimize any impact upon IT The results obtained in previous work [3][4]
services. showed that formalize some treatments over the change
Some automated IT tools were developed, using dependence and a simpler heuristic to build an optimized
the good practice and procedures of ITIL and COBIT, allocation of changes, now we present more features to
aligning IT and business strategy, measuring the IT the tool and a conceptual model.
performance and building IT feedbacks to strategic
decisions. We propose a change management tool to II. THE CHANGE MANAGEMENT TOOL
select the needed changes, determine in an automated As commented in previous work, as important as
form the best time intervals to become change window doing a planned change management is the selection of
and build an optimized schedule of changes. changes that should be implemented. We present a model
The first results of this study were published in divided into three main steps: change filtering, refinement
[3], where we presented: the division of the changes in of changes and scheduling.
family of changes, where each one represents an IT
service; the concepts of direct and indirect dependencies
between the changes; the treatment realized over the
399
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_70, © Springer Science+Business Media B.V. 2010
400 OLIVEIRA AND FILHO
B. Refinement of Changes
A. Change Filtering In this step, all the changes have already been
The first step, we perform a filtering of all possible released. Our model requires treatments of these changes.
changes, represented by RFC documents. This document These changes are linked to one or more configuration
is a formal proposal to implement a change, referencing item (CI), which are part of a CMDB. The CMDB has a
details as implementation cost, implementation and not mapping of all CI, their dependencies and characteristics.
implementation risks, proposed date to implement, etc. Each change could be linked to new CIs (configuration
IT investments adds details to RFC documents and item), which are out of CMDB. If this occurs, is
the change advisor board (CAB) [2] analyze inherent to necessary to map it to determine the dependencies
the proposed changes, considering the financial reserve to between this new CI and the other that are part of CMDB.
IT investments, return of investments (ROI) and other After the mapping of dependencies between the
financial analysis. We want to propose a more detailed changes, formation of the family of changes, our model
analysis with respect to ROI linked to this tool. applies treatments over the dependencies. The model
The back-out plan is formed by all actions to return provides two cases: cross dependency and multiple
the proposed change to your previous state in case of indirect dependencies from one family of changes to
failure. another [3] [4].
The CAB is formed by stakeholders from each area The final object of phase is the set of changes with
affected by the proposed changes. This group is all CIs and the mapping and treatments of your
responsible for receiving, analyzing, allowing or denying dependencies.
the changes. The set of changes allowed is described in
forward schedule of changes (RFC). C. Scheduling
In [4], we described that each change is The schedule of changes centers the action
represented by C(i, j) , where i={1, 2, 3,…, n} and j={1, 2, implemented by the model to build an optimized
3, …, m}. The values of i and j will determine the allocation of the changes, considering a planning to get a
position of C(i, j) in the scheduling of changes. The set of minor implementation time and/or less financial loss.
changes is represented by SC={C(1, 1),…,C(n, m)}. The SC is The roll-back plan contains the actions to be
formed by one or more families of changes. A family of implemented and gets the previous state of one change in
changes is formed by changes C(i, j) with the same i and case of failure during your implementation. The roll-back
different value of j. In the Schedule, each family plan is different, bigger and more complex than back-out
represents a column. The indirect dependence (ID) plan, because it includes adding, modifying or removing
represents the dependence among changes with different some of them. It modifies the action, takes into
i. We have a direct dependence (DD) when the consideration two main points to do: The treatments of
implementation of C(i,j) depends on of C(i,j-1). dependencies in the previous step and new mapping of
AN IT CHANGE MANAGEMENT TOOL TO OPTIMIZE SCHEDULING AND PLANNING 401
dependencies by the addition of new CIs. Our studies Analyzing the table 1, we have the change
showed that some modifications in the change actions are windows shown in table 2.
necessary to form the roll-back plan, although it inherits
some of them from back-out plan. Table 2. Change windows.
Intervals
IT Service 0h to 3h 3h to 6h ... 21h to 24h
1 $ 1.200 $ 1.400 ... $ 2.100
2 $ 1.400 $ 1.300 ... $ 1.700 Figure 3: scene to implement the changes.
402 OLIVEIRA AND FILHO
The branch office depends on the main one, existence of cross dependencies and multiple ID. After
because of this point, the exchange of core routers could that, the rules of allocation of changes on array [4] are
be performed in parallel form. When the change of core- applied. In figure 5 shows the last state of schedule of
router of main office begins, the IT services of branch changes.
office break at the same time. The changes of core-switch
were divided into three phases: change each unit;
configure the operational system; perform security
settings.
The change of switches could be performed just
after the first and second phases of the core-router
changes.
The changes of servers can start after the second
phase of core-routers and distribution-switches. This is
necessary because the change of servers to work with the
authentication service depends on the changes of the
routers and switches.
The model brings together the changes linked to an
IT service to a column: the family of changes. In this
case, to illustrate it, we consider the changes of each
office as two different IT services. The table 3 shows the
identification of IT service as family of changes.
Table 3: Identifying the IT services.
Abstract — A Content-Addressable Memory (CAM) is a order to choose the appropriate route. A CAM is a good
memory that implements the lookup-table function in a single choice for implementing this lookup operation due to its fast
clock cycle using dedicated comparison circuitry. CAMs are search capability.
especially popular in network routers for packet forwarding Fig. 1 [2] illustrates how a CAM accomplishes address
and packet classification, but they are also beneficial in a
variety of other applications that require high-speed table
lookup by implementing the routing table shown in the table.
lookup. The main CAM design challenge is to reduce power As shown in Fig. 1, the packet destination-address of 01101
consumption associated with the large amount of parallel is the input to the CAM. According to the table, two
active circuitry, without sacrificing speed or memory density. locations match, with the (priority) encoder choosing the
In this project, a low power content-addressable memory using upper entry and generating the match location 01, which
pipelined hierarchical search scheme is implemented. The corresponds to the most-direct route. This match location is
search operation is pipelined and the match-lines are broken to the input address to a RAM that contains a list of output
several segments. Whenever the stored words fail to match in ports, as depicted in Fig. 1. A RAM read operation outputs
their initial segments, the search operation is discontinued for the port designation, port B, to which the incoming packet is
subsequent segments, hence the power is saved. The proposed
scheme has been implemented in a 64×64 bit ternary CMOS
forwarded. We can treat the match location output of the
CAM. The schematics of both the pipelined and non-pipelined CAM as a pointer that retrieves the associated word from
CAMs are designed. PSPICE power simulation is used to the RAM. In the particular case of packet forwarding, the
extract the power consumption of both memory designs. The associated word is the designation of the output port. This
simulation results demonstrate an effective overall power CAM/RAM system is a complete implementation of an
reduction in the pipelined CAM compared to non-pipelined address-lookup engine for packet forwarding.
architecture for the given example memory pattern.
I. INTRODUCTION
405
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_71, © Springer Science+Business Media B.V. 2010
406 SONG ET AL.
the low power pipelined CAM memory design for the given are also becoming popular in solving other problems such as
example memory pattern. sorting and range searching.
D10
D11
D12
D13
D14
D15
D16
D17
D18
D19
D20
D21
D22
D23
D24
D25
D26
D27
D28
D29
D30
D31
D32
D10
D11
D12
D13
D14
D15
D16
D17
D18
D19
D20
D21
D22
D23
D24
D25
D26
D27
D28
D29
D30
D31
D32
D10
D11
D12
D13
D14
D15
D16
D17
D18
D19
D20
D21
D22
D23
D24
D25
D26
D27
D28
D29
D30
D31
D32
D1
D2
D3
D4
D5
D6
D7
D8
D9
D1
D2
D3
D4
D5
D6
D7
D8
D9
D1
D2
D3
D4
D5
D6
D7
D8
D9
period T is:
stage12 stage22 stage32 stage42
R1 L1 R1 L1 R1 L1 R1
u1
u2
u3
u4
u5
u6
u7
u8
u9
u1
u2
u3
u4
u5
u6
u7
u8
u9
u1
u2
u3
u4
u5
u6
u7
u8
u9
u1
u2
u3
u4
u5
u6
u7
u8
u9
u10
u11
u12
u13
u14
u15
u16
u17
u18
u19
u20
u21
u22
u23
u24
u25
u26
u27
u28
u29
u30
u31
u32
u10
u11
u12
u13
u14
u15
u16
u17
u18
u19
u20
u21
u22
u23
u24
u25
u26
u27
u28
u29
u30
u31
u32
u10
u11
u12
u13
u14
u15
u16
u17
u18
u19
u20
u21
u22
u23
u24
u25
u26
u27
u28
u29
u30
u31
u32
u10
u11
u12
u13
u14
u15
u16
u17
u18
u19
u20
u21
u22
u23
u24
u25
u26
u27
u28
u29
u30
u31
u32
R2 L2 R2 L2 R2 L2 R2
R3 L3 R3 L3 R3 L3 R3
R4 L4 R4 L4 R4 L4 R4
R5 L5 R5 L5 R5 L5 R5
R6 L6 R6 L6 R6 L6 R6
R7 L7 R7 L7 R7 L7 R7
E (t ) 1
R8 L8 R8 L8 R8 L8 R8
R9 L9 R9 L9 R9 L9 R9
R10 L10 R10 L10 R10 L10 R10
T Vdd T
Pavg (T ) = = ³ P (t )dt = ³ idd (t ) dt (1)
R11 L11 R11 L11 R11 L11 R11
R12 L12 R12 L12 R12 L12 R12
R13 L13 R13 L13 R13 L13 R13
R14 L14 R14 L14 R14 L14 R14
D10
D11
D12
D13
D14
D15
D16
D17
D18
D19
D20
D21
D22
D23
D24
D25
D26
D27
D28
D29
D30
D31
D32
D10
D11
D12
D13
D14
D15
D16
D17
D18
D19
D20
D21
D22
D23
D24
D25
D26
D27
D28
D29
D30
D31
D32
D10
D11
D12
D13
D14
D15
D16
D17
D18
D19
D20
D21
D22
D23
D24
D25
D26
D27
D28
D29
D30
D31
D32
D10
D11
D12
D13
D14
D15
D16
D17
D18
D19
D20
D21
D22
D23
D24
D25
D26
D27
D28
D29
D30
D31
D32
D1
D2
D3
D4
D5
D6
D7
D8
D9
D1
D2
D3
D4
D5
D6
D7
D8
D9
D1
D2
D3
D4
D5
D6
D7
D8
D9
0 0
R16 L16 R16 L16 R16 L16 R16
stage12
stage13
stage22
stage23
stage32
stage33
stage42
stage43
T T T
R1 L1 R1 L1 R1 L1 R1
u1
u2
u3
u4
u5
u6
u7
u8
u9
u1
u2
u3
u4
u5
u6
u7
u8
u9
u1
u2
u3
u4
u5
u6
u7
u8
u9
u1
u2
u3
u4
u5
u6
u7
u8
u9
u10
u11
u12
u13
u14
u15
u16
u17
u18
u19
u20
u21
u22
u23
u24
u25
u26
u27
u28
u29
u30
u31
u32
u10
u11
u12
u13
u14
u15
u16
u17
u18
u19
u20
u21
u22
u23
u24
u25
u26
u27
u28
u29
u30
u31
u32
u10
u11
u12
u13
u14
u15
u16
u17
u18
u19
u20
u21
u22
u23
u24
u25
u26
u27
u28
u29
u30
u31
u32
u10
u11
u12
u13
u14
u15
u16
u17
u18
u19
u20
u21
u22
u23
u24
u25
u26
u27
u28
u29
u30
u31
u32
R2 L2 R2 L2 R2 L2 R2
dQc d (CVC ) dV (T )
R12 L12 R12 L12 R12 L12 R12
k T
R13 L13 R13 L13 R13 L13 R13
R14 L14 R14 L14 R14 L14 R14
D10
D11
D12
D13
D14
D15
D17
D16
D18
D19
D20
D21
D22
D23
D24
D25
D26
D27
D28
D29
D30
D31
D32
D10
D11
D12
D13
D14
D15
D16
D17
D18
D19
D20
D21
D22
D23
D24
D25
D26
D27
D28
D29
D30
D31
D32
D10
D11
D12
D13
D14
D15
D16
D17
D18
D19
D20
D21
D22
D23
D24
D25
D26
D27
D28
D29
D30
D31
D32
k ⋅ idd (t ) = = =C C
D1
D2
D3
D4
D5
D6
D7
D8
D9
D1
D2
D3
D4
D5
D6
D7
D8
D9
D1
D2
D3
D4
D5
D6
D7
D8
D9
D1
D2
D3
D4
D5
D6
D7
D8
D9
stage14
R1
stage24
L1 R1
stage34
L1 R1
stage44
L1 R1 dt dt dt C 0
u1
u2
u3
u4
u5
u6
u7
u8
u9
u1
u2
u3
u4
u5
u6
u7
u8
u9
u1
u2
u3
u4
u5
u6
u7
u8
u9
u1
u2
u3
u4
u5
u6
u7
u8
u9
u10
u11
u12
u13
u14
u15
u16
u17
u18
u19
u20
u21
u22
u23
u24
u25
u26
u27
u28
u29
u30
u31
u32
u10
u11
u12
u13
u14
u15
u16
u17
u18
u19
u20
u21
u22
u23
u24
u25
u26
u27
u28
u29
u30
u31
u32
L10
u11
u12
u13
u14
u15
u16
u17
u18
u19
u20
u21
u22
u23
u24
u25
u26
u27
u28
u29
u30
u31
u32
u10
u11
u12
u13
u14
u15
u16
u17
u18
u19
u20
u21
u22
u23
u24
u25
u26
u27
u28
u29
u30
u31
u32
R2 L2 R2 L2 R2 L2 R2
R3 L3 R3 L3 R3 L3 R3
R4 L4 R4 L4 R4 L4 R4
R5 L5 R5 L5 R5 L5 R5
That is, the voltage in capacitor C is numerically equal to line bits, the last stage is not turned on. As a result, power is
the average power of the CUT. Thus by measuring the saved.
voltage VC(T) in PSPICE simulation, we know the value of
Pavg(T). The power consumption of a circuit can be
extracted by PSPICE simulation.
the pipelined CAM with the second stage mismatch during amount of parallel active circuitry, without sacrificing speed
time period of T1=21ns is found to be Pavg=561.288μW. or memory density. In this paper, a low power 4-stage
Compared with the non-pipelined scheme, a power saving pipelined 64×64 CAM memory design [3] is implemented.
of 74% is observed. This power saving is slightly lower than The power savings of the pipelined match-lines is due to the
the case when the first stage is mismatched. This is fact that only a small portion of the ML segments will be
reasonable because the first stage is still turned on for activated in general searching operation, and the rest of the
comparison when the mismatch occurs in the second stage. segments will be deactivated to save power. The schematic
Similarly, the average power consumption of the 4-stage of the pipelined CAM memory is designed in PSPICE.
pipelined CAM during time period T1=21ns when the Simulation results demonstrate the correct function of the
mismatch occurs in the third stage is found to be design. Furthermore, PSPICE simulations are used to
Pavg=1.0417mW in PSPICE power simulation. This extract the average power consumptions of both pipelined
translates to a power saving of 52.65% compared to the non- and non-pipelined CAM memories. Simulation results show
pipelined CAM design with the same input pattern. that if all the four stages mismatch, a significant power
Further, the average power consumption of the 4-stage saving of 77.1% is observed for the pipelined CAM
pipelined CAM during time period T1=21ns when the compared to the non-pipelined design. The power savings
mismatch occurs in the 4th stage is found to be also depends on where the first mismatch occurs in the four
Pavg=1.45mW in PSPICE power simulation. Compared to stages. If the mismatch occurs in the early pipeline stage,
the original non-pipelined CAM design, a power saving of more power saving will be achieved due to the factor that
34% is observed. This leads to least power saving compared the rest of the stages can be turned off to save power.
to the previous cases because three out of four stages in the Similarly, power can also be saved if the hierarchical
pipelined CAM is activated for comparison in this case. search-lines activate only a small portion of the local
Compare the energy consumption of the propose searchlines (LSLs). In the future work, we will study how to
architecture with the non-pipelined architecture, we can see save power in search-lines by using hierarchical search
that the addition of pipelined match-lines reduces the energy scheme and combine it with the pipelined scheme.
consumption, consisting of a ML energy consumption and
no change in SL energy. Since most ML segments miss in REFERENCES
the first 16-bit stage (rarely activating the subsequent 48
bits), one expects the ML energy consumption to be reduced [1] A. S. Tanenbaum, “Computer Networks,” Prentice Hall,
by 74%, compared to the non-pipelined architecture. The Upper Saddle River, NJ, 2003.
simulation results measurement, however, the total 56% ML [2] J. M. Rabaey, A. Chandrakasan, and B. Nikoliæ, Digital
energy is reduction in the 64×64 pipelined content Integrated Circuits: A Design Perspective, 2nd ed.
addressable memory designs. Englewood Cliffs, NJ, Prentice-Hall, 2003.
From above simulations, we can see that generally the [3] K. Pagiamtzis and A. Sheikholeslami, “Pipelined match-
pipelined CAM design leads to significant power savings lines and hierarchical search-lines for low-power content-
compared to the non-pipelined CAM design. Especially if addressable memories,” Proc. IEEE Custom Integrated
the first stage is found to be a mismatch with the searching Circuits Conf., 2003, pp. 383–386.
bits, all the four stages will be turned off and a significant [4] R. Sangireddy, and A. K. Somani, “High-speed IP
power saving of 77.1% is observed for the given patterns. If routing with binary decision diagrams based hardware
the mismatch occurs in a later stage, less power saving will address lookup engine,” IEEE Journal on Selected Areas in
be resulted because more stages will be activated for Communications, vol. 21, no. 4, pp. 513-521, May 2003.
comparison. If no mismatch exists, i.e. all the four stages [5] K. Etzel, “Answering IPv6 lookup challenges,”
match the searching bits, then all the four stages will be Technical Article, Cypress Semiconductor Corporation, Oct.
activated and no power saving can be achieved. Generally, 27, 2004.
most of the patterns lead to mismatch in one of the four [6] A. J. McAuley, and P. Francis, “Fast routing table
stages. Hence effective power savings can be achieved for lookup using CAMs,” Proceedings INFOCOM’93, San
the pipelined CAM design. Francisco, CA, vol. 3, pp. 1382-1391, 1993.
[7] T.-B. Pei and C. Zukowski, “Putting routing tables in
V. Conclusions and Future Work silicon,” IEEE Network Mag., vol. 6, pp. 42–50, Jan. 1992.
[8] L. Chisvin and R. J. Duckworth, “Content-addressable
A CAM is a memory that implements the lookup-table and associative memory: Alternatives to the ubiquitous
function in a single clock cycle using dedicated comparison RAM,” IEEE Computer, vol. 22, pp. 51–64, July 1989.
circuitry. CAMs are especially popular in network routers [9] H. Miyatake, M. Tanaka, and Y. Mori, “A design for
for packet forwarding and packet classification, but they are high-speed lowpower CMOS fully parallel content-
also beneficial in a variety of other applications that require addressable memory macros,” IEEE J. Solid-State Circuits,
high-speed table lookup. The main CAM-design challenge vol. 36, pp. 956–968, June 2001.
is to reduce power consumption associated with the large
Design and Optimization of Piezoelectric
Dual-Mode Micro-Mirror
Jichao Zhong, Xingguo Xiong, Zheng Yao, Junling Hu*, Prabir Patra*
Department of Electrical and Computer Engineering,
*Department of Mechanical Engineering
University of Bridgeport, Bridgeport, CT 06604, USA
Abstract—MEMS micro-mirrors have been widely used According to the working modes, MEMS micro-mirrors can
in optical communication, projection display, and be divided into two categories: piston mirrors and torsional
microscanner, etc. Various MEMS micro-mirror devices mirrors. Piston micro-mirrors move perpendicular to the
have been reported. In this paper, the design of a bulk- device plan, while torsional micro-mirrors rotate along a
micromachined MEMS piezoelectric dual-mode micro- certain axis. Various piston and torsional micro-mirrors
mirror is proposed. The working principle of the dual- have been reported [1]-[4]. The working principles and
mode micro-mirror device is analyzed. Based on the design considerations of MEMS micromirrors have also
analysis, a set of optimized design parameters of the been analyzed in details [5]-[6]. Furthermore, dual-mode
micro-mirror is suggested. Theoretical analysis shows micro-mirrors which can work in both piston and torsional
that the proposed micro-mirror can deflect for a modes have been reported [7]-[8]. Dual-mode micro-mirrors
maximum displacement of 2.5m in its piston mode, and offers more flexibility in light modulation. Users can
tilt for a maximum angle of 4.88º in its torsional mode. reconfigure them into either piston or torsional modes
The fabrication flow of the MEMS micro-mirror is also according to their individual needs.dd
suggested. The proposed micro-mirror has the MEMS micro-mirrors can be activated using various
advantages of large displacement and large tilt angle due mechanisms, such as electrostatic driving, electromagnetic
to piezoelectric activation. The PZT actuated micro- driving, thermal driving, piezoelectric driving, as well as
mirror (PAM) also leads to improved linearity in tilting shape-memory alloy driving, etc. Among them, electrostatic
angle and piston displacement. Further, fast response of driving is popular due to their easy in implementation. But
the PZT unimorph on the applied actuation voltage the maximum displacement is limited by snap-down effect
ensures wider operation bandwidth of the PAM in in perpendicular driving. Electromagnetic driving has larger
contrast to thermal activated micro-mirrors. activation force and displacement, but the fabrication is
more complex. Thermal driving is relatively slow in
Keywords: MEMS (Microelectromechanical System), dual- response, and the energy efficiency is not high due to
mode micro-mirror, piezoelectric technology (PZT), piston thermal dissipation. Piezoelectric driving utilizes
micro-mirror, torsional micro-mirror. piezoelectric effect to activate MEMS micro-mirrors. It can
achieve large displacement or rotation angle with high
I. Introduction energy efficiency. Various piezoelectric micro-mirrors have
been reported [9]-[11].
Microelectromechanical Systems (MEMS) deal with In this paper, a bulk-micromachined piezoelectric-driven,
devices or systems in the size range of 1μm~1mm, which dual-mode MEMS micro-mirror device is proposed. The
integrates both mechanical and electrical components into a dual-mode micro-mirror can be configured to work in either
single chip. MEMS merges with nanotechnology and piston or torsional mode according to the need. The working
creates the field of Nanoelectromechanical Systems principle of the dual-mode micro-mirror device is analyzed.
(NEMS). Optical MEMS is a unique application where Based on the analysis, a set of optimized design parameters
MEMS are used to direct, guide, filter, and, in some of the micro-mirror is suggested. Theoretical analysis shows
instances, amplify light. Well known applications of optical that the proposed micro-mirror can deflect for a maximum
MEMS include optical switching and digital light displacement of 2.5m in its piston mode, and tilt for a
projection. In optical switching applications, micro-mirrors maximum angle of 4.88º in its torsional mode. The
are used to steer light from an incoming fiber optic input to fabrication flow of the MEMS micro-mirror is also
a fiber optic output. In digital light projection applications, suggested. The proposed micro-mirror can achieve large
micro-mirrors, such as digital micro-mirror device (DMD) displacement and large tilt angle due to piezoelectric
by Texas Instruments Inc., incident light is reflected by an activation. The PZT actuated micro-mirror (PAM) also has
array of torsional micro-mirrors so that each mirror pixel improved linearity in tilting angle and piston displacement.
can control a corresponding spot on the screen to be bright
or dark. In this way, vivid pictures can be projected onto the
screen.
411
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_72, © Springer Science+Business Media B.V. 2010
412 ZHONG ET AL.
This equation can be expanded to a full matrix form the PZT actuators will expand or shrink simultaneously,
[12]: which in turn activates the frame (and the mirror) to move
perpendicular to the device plane. Hence, the micro-mirror
works in piston mode. By applying electrical driving
voltages to the PZT actuators on either the piston or
torsional beams, the micro-mirror can work in either piston
or torsional mode. Hence this is a piezoelectric activated
dual-mode micro-mirror design.
tilting
axis
Note that, for any given piezoelectric material, the dij
components connecting the strain and the applied field in
the inverse effect are identical to the dij connecting the
polarization and the stress in the direct effect. Figure 2. Schematic diagram of the dual-mode piezoelectric
The electromechanical coupling coefficient k is a micro-mirror
measure of how much energy is transferred from electrical
to mechanical energy, or vice versa, during the actuation
process [12]:
energy _ converted
k2 =
input _ energy
this relation holds true for both mechanical-to-electrical and
electrical-to-mechanical energy conversion. The magnitude
of k is a function of not only the material, but also the
geometries of the sample and its oscillation mode.
activation. The PZT actuated micro-mirror (PAM) also has [6] X. M. Zhang, F. S. Chau, C. Quan, Y. L. Lam, A. Q.
improved linearity in tilting angle and piston displacement. Liu, “A study of the static characteristics of a torsional
In the future work, we will further improve the structure micro-mirror”, Sensors and Actuators A: Physical, Vol. 90,
design, so that the driving voltage can be reduced from 2001, pp. 73–81.
current 10V in piston mode to around 5V. In this way, it can [7] U. Krishnamoorth, K. Li, K. Yu, D. Lee, J. P. Heritage,
be compatible with the power supply of on-chip CMOS O. Solgaard, “Dual-mode micro-mirrors for optical phased
circuitry. By reducing the required driving voltage, the array applications”, Sensors and Actuators A: Physical, Vol.
power consumption of the micro-mirror can also be further 97-98, 2002, pp. 21-26.
reduced. [8] X. Xiong, T. Dallas, S. Gangopadhyay, J. Berg, T.
References Henryk, “Design and Simulation of Aluminum Bi-
functional Spatial Light Modulator", Proc. of the 48th IEEE
[1] G. Zhou, V.J. Logeeswaran, F.E.H. Tay, F.S. Chau, International Midwest Symposium on Circuits & Systems
"Digital deflection micro-mirrors for adaptive optics (MWSCAS’05), Cincinnati, Ohio, USA, Aug 7-10, 2005, pp.
applications", 12th International Conference on Solid-State 159-162.
Sensors, Actuators and Microsystems (Transducers'03), [9] Y. Yee, J. U Bu, M. Ha, J. Choi, H. Oh, S. Lee, and H.
Vol. 1, June 8-12, 2003, pp. a591-a594. Nm, “Fabrication and characterization of a PZT actuated
[2] K. Jia, S. Pal, H. Xie, "An electrothermal tip–tilt–piston micro-mirror with two-axis rotational motion for free space
micromirror based on folded dual S-shaped bimorphs", optics”, Proc. of IEEE MEMS Workshop 2001, pp. 317–320.
Journal of Microelectromechanical Systems, Vol. 18, Issue [10] H. Nam, Y. Kim, S. Cho, Y. Yee and J. Bu, "Low
5, Oct. 2009, pp. 1004-1015. voltage PZT actuated tilting micro-mirror with hinge
[3] G. Vdovin, S. Middelhoek, L. Sarro, "Deformable mirror structure", Conference Digest of 2002 IEEE/LEOS
display with continuous reflecting surface micromachined in International Conference on Optical MEMs, 2002, pp. 89-90.
silicon", IEEE Proceedings of Micro Electro Mechanical [11] S. J. Kim, Y. H. Cho, H. J. Nam, J. U. BU,
Systems (MEMS'95), Jan. 29 - Feb. 2, 1995, pp. 61–65. "Piezoelectrically pushed rotational micro-mirrors for wide-
[4] K.B. Lee, L. Lin, "A vertically-supported two-axial angle optical switch applications", IEEE The Sixteenth
torsional micro-mirror", Proceedings of 17th IEEE Annual International Conference on Micro Electro
International Conference on Micro Electro Mechanical Mechanical Systems (MEMS'03), Kyoto, Japan, Jan. 19-23,
Systems (MEMS'04), 2004, pp. 41-44. 2003, pp. 263-266.
[5] O. Degani, Y. Nemirovsky, “Design considerations of [12] C. Liu, "Foundation of MEMS", Chapter 7:
rectangular electrostatic torsion actuation based on new “Piezoelectric Sensing and Actuation”, Prentice Hall, Sept.
analytical pull-in expression”, Journal of 5, 2005, ISBN: 0131472860.
Microelectromechanical Systems, Vol. 11, pp. 20–26, 2002.
Device for Measurement of Clamped Joints
Friction Torque
Ing. Eva Batešková, PhD., Ing. Jozef Maščenik,PhD., Ing. Jozef Haľko, PhD., Ing. Štefan Gašpár
Faculty of manufacturing Technologies, TU Košice with seat in Prešov, Štúrova 31,
08001 Prešov, Slovak republic
[email protected], [email protected], [email protected], [email protected]
Abstract - This contribution deals with a proposal of II. PROPOSAL OF 3D DEVICE FOR FRICTION JOINT
device for measurement of clamped joints. In the second MEASUREMENT
chapter, we propose a model of clamped joint. In the
experimental part we measured values of the friction torque Fig. 1 shows the proposed device solution. The
depending on twisting moment, and in the conclusion we device consists of frame 1, to which bearing shell sliding
measured shaft surface abrasivity. surface 14 is screwed. Holder 8 is welded to the frame and
there are pegs 11 screwed in it. Via screw 4, through
INTRODUCTION collar 13, dynamometer 5, peg holder and peg, the
clamping power is transferred on claws 3. Weight 12
One of the basic construction elements for slides axially along lever 6. Shaft 2 is from both sides
construction production either in engineering or building placed in bearings 9, which are in shell 10. Screw 7 serves
and construction industries are clamped joint claws. From to ensure the shaft position.
the point of view of working life and criteria on the
clamped joint jaws, the main importance belongs to Measuring procedure:
stating the correct clamping power, as well as stating the By the screw rotation, clamping of the shaft by means of
correct friction torque in the joint. If there is a small claws is generated. From the dynamometer, the clamping
clamping power on this joint, it can cause an insufficient power is read. By means of the weight, it slides axially in
friction torque and subsequently damage the clamped the direction from the shaft axis along the lever until the
joint and thus damage the whole mechanism. On the clamped joint is broken. From the weight mass and the
contrary, high clamping power can cause a big distance from the shaft, twisting moment of the clamped
deformation either on the hub or clamped joint claws, joint breach is calculated. The friction torque is
which can lead to cracks in the claws or disabling the determined from the condition for transfer of the twisting
mutual re-assembling and disassembling of the hub and moment. The disadvantage of this device is measurement
claws [1]. scope restriction by the lever length and relatively
complicated construction, which, however, eliminates the
I. BASIC REQUIREMENTS ON JOINTS shortcomings from the two previous proposals.
417
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_73, © Springer Science+Business Media B.V. 2010
418 BATEŠKOVÁ ET AL.
holes for small claws. The holder with the sliding surface
and collar were inserted into the tubes with the internal
diameter of φ 12 mm on the construction frame. The claws
were put on the pegs according to the peg distribution.
The shaft was forced into the bearings and subsequently
ensured against drop-out by snap rings 40. The lever was
screwed by a screw with a washer. Afterwards, a locking
element against lever fall was screwed into the frame. The
weight was put on the lever. The dynamometer was Fig. 3 Device for measurement of surface abrasiveness Mitutoyo SJ 301
screwed into the collar on the peg holder and on the
opposite side of the dynamometer the collar with bearing Measurement of equal pressure distribution on hub
were screwed. The frame was supplemented by screw perimeter:
M24 x 2, which presses on the dynamometer collar and For this measurement, we used big claws of the
serves for educing the axial power. Finally, the entire clamped joint, which were put on the pegs and
construction was fixed to the work table. subsequently tightened by the screw to 2000 N. The
measured values of the clamped joint breach, as well as
the calculated clamping power is in chart 1. Conversion of
the number of dynamometer divisions to power F was
done according to the equation (1)
TABLE I
MEASUREMENT ON CLAWS WITH EXTERNAL DIAMETER OF
Ø 90 MM
Number of lever
divisions
33 31 29 27 25
Fig. 2 Complete clamped joint device
Number of
dynamometer 155 145 131 122 120
divisions
III. EXPERIMENTAL MEASUREMENTS
F [N] 561,5 525,3 474,6 441,9 434,7
The sample for this measurement is a shaft with the Number of lever
diametre of φ 40 mm and one type of claws : divisions
19 17 15 13 11
Number of
• For equal pressure distribution on the claw hub dynamometer 71 56 43 40 32
perimeter with external diametre of φ 90 mm divisions
Measuring method: F [N] 257,2 202,9 155,8 144,9 115,9
Every measurement began by clamping the Number of lever
divisions
13 15 17 19 21
clamped joint by power of 2 kN by a screw. On the lever,
Number of
the weight was set on 33 divisions, whereas every division dynamometer 55 55 78 80 81
corresponds with a different twisting moment and divisions
afterwards the screw was slowly loosened until the F [N] 199,2 199,2 282,6 289,8 293,4
clamped joint was broken. The breach occurs if the lever Number of lever
27 29 31 33
moves. When the lever moves, we read the number of divisions
divisions from the dynamometer. Gradually, the weight Number of
moves towards the rotation axis by two division up to 9 dynamometer 101 105 109 110
divisions
divisions. F [N] 365,9 380,4 394,9 398,5
Subsequently, the measurement was performed
from 9 divisions to 33 divisions. After these
measurements, we carried out measurement of the surface
abrasiveness by device Mitutoyo SJ 301, and The calculation of the friction torque for equal pressure
consequently soft regrinding of claws by abrasive-coated distribution was done according to equation (2)
paper and re-measurement.
π
M T = FN . f ´.d = FN . . f .d (2)
2
TABLE II
CHART OF CALCULATED FRICTION TORQUES FOR CLAWS OF
Ø 90 MM
F [N] 561,5 525,3 474,6 441,9 434,7
MT
5,2920 4,9508 4,4730 4,1648 4,0970
[N.m]
F [N] 257,2 202,9 155,8 144,9 115,9
MT
2,4241 1,9123 1,4684 1,3657 1,0923
[N.m]
F [N] 199,2 199,2 282,6 289,8 293,4
MT
1,8774 1,8774 2,6634 2,7313 2,7652
[N.m]
F [N] 365,9 380,4 394,9 398,5
MT
3,4485 3,5852 3,7218 3,7558
[N.m]
CONCLUSION
The calculated values of the friction torque are
distorted. They depend on many factors, among which
there are e.g.: precision of clamped joint make-out;
surface abrasivity; clamping power; conditions of
measurement; number of measurement etc. This causes
the fact that when more measurements are carried out, the
value of friction torque varies, which causes the
dispersion in measurements. Finally, for demanding
applications, it is suitable to recommend modern
Fig. 4 Friction torque dependence on twisting moment hydraulic tightening machines for their higher precision,
and precise digital torque wrenches for clamped joint
Surface abrasivity measurement: breach. This contribution forms a part of the solution of
Fig. 5 shows the recording of measuring abrasivity of the the grant task VEGA 1/4156/07 and KEGA 3/6279/08.
surface of shaft with diameter of 50 mm.
REFERENCES
[1] MEDVECKÝ, Š.: Základy konštruovania, EDIS, Źilina, 1999.
[2] PAVLENKO, S. – HAĽKO, J.: Časti strojov I. FVT Prešov, 2006,
ISBN 80-8073-100-4.
[3] ROVŇÁKOVÁ, Sylvia - LÍŠKA, Ondrej: Predchádzanie kolíziám
pri automatizovaných pracoviskách využitím modelovania a
simulačných programov, Prešov, TU Košice, FVT, 2008. s. 339-
342. ISBN 978-80-553-0044-3.
Self-Optimization for Dynamic Scheduling in
Manufacturing Systems
Ana Madureira, Ivo Pereira
Institute of Engineering – Polytechnic of Porto, Portugal
421
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_74, © Springer Science+Business Media B.V. 2010
422 MADUREIRA AND PEREIRA
global optimum solution, but to continuously track the communities to look for inspiration in diverse fields, e.g.
optimum in dynamic environments, or to find a robust robotics, artificial intelligence or biology, to find new ways of
solution that operates optimally in the presence of designing and managing systems. Hybridization of different
perturbations [2][9]. In spite of all the previous trials, the approaches seems to be a promising research field of
scheduling problem is still known to be NP-complete, even computational intelligence focusing on the development of
for static environments. This fact poses serious challenges to the next generation of intelligent systems.
conventional algorithms and incites researchers to explore
new directions [9-11] and Multi-Agent technology has been V. AUTODYNAGENTS ARCHITECTURE
considered an important approach for developing industrial Distributed environment approaches are important in order
distributed systems. to improve scheduling systems flexibility and capacity to react
to unpredictable events. It is accepted that new generations of
III. MULTI-AGENT SYSTEMS
manufacturing facilities, with increasing specialization and
Multi-agent paradigm is emerging for the development of
integration, add more problematic challenges to scheduling
solutions to very hard distributed computational problems.
systems. For that reason, issues like robustness, regeneration
This paradigm is based either on the activity of "intelligent"
capacities and efficiency are currently critical elements in the
agents which perform complex functionalities or on the
design of manufacturing scheduling system and encouraged
exploitation of a large number of simple agents that can
the development of new architectures and solutions,
produce an overall intelligent behavior leading to the solution
leveraging the MAS research results.
of alleged almost intractable problems.
The work reported in this paper is concerned with the
Considering the complexity inherent to the manufacturing
resolution of realistic scheduling problems, called here
systems, dynamic scheduling is considered an excellent
Extended Job-Shop Scheduling Problems (EJSSP) [9].
candidate for the application of agent-based technology. In
Moreover, it is concerned with integrated scheduling of jobs
many implementations of MAS systems for manufacturing
which are products composed by several parts or components
scheduling, the agents model the resources of the system and
that may be submitted to a number of manufacturing and
the tasks scheduling are done in a distributed way by means
multi-level assembly operations.
of cooperation and coordination amongst agents [9-10]. When
AUTODYNAGENTS is an Autonomic Scheduling System
responding to disturbances, the distributed nature of multi-
in which communities of agents model a real manufacturing
agent systems can also be a benefit to the rescheduling
system subject to perturbations. Agents must be able to learn
algorithm by involving only the agents directly affected,
and manage their internal behavior and their relationships
without disturbing the rest of the community which can
with other autonomic agents, by cooperative negotiation in
continue with their work.
accordance with business policies defined by user manager.
IV. NATURE INSPIRED OPTIMIZATION TECHNIQUES It is a Multi-Agent system where each agent represents a
resource (Machine Agents) in a Manufacturing System. Each
Biological and natural processes have been a source of Machine Agent must be able: to find an optimal or near
inspiration for computer science and information technology. optimal local solution through Genetic Algorithms, Tabu
The interest of the Nature Inspired Optimization Search or other NIT; to deal with system dynamism (new jobs
Techniques, also named Meta-heuristics, is that they arriving, cancelled jobs, changing jobs attributes, etc.); to
converge, in general, to satisfactory solutions in an effective change/adapt the parameters of the basic algorithm according
and efficient way (computing time and implementation to the current situation; to switch from one Meta-Heuristic
effort). NIT have often been shown to be effective for algorithm to another and to cooperate with other agents.
difficult combinatorial optimization problems appearing in Scheduling approach followed by AUTODYNAGENTS
several industrial, economical, and scientific domains system is rather different from the ones found in the
[3][9][11]. Prominent examples are evolutionary algorithms, literature; as we try to implement a system where each agent
simulated annealing, tabu search, scatter search, memetic (Resource Agent) is responsible for optimizing the scheduling
algorithms, ant colony systems and particle swarm of operations for one machine through a NIT. This considers
optimization. a specific kind of social interaction that is cooperative
When considering and understanding solutions followed by problem solving (CPS), where the group of agents work
nature it is possible to use this acquired knowledge on the together to achieve a good solution for the problem.
resolution of complex problems on different domains. From The original Scheduling problem defined in [7][9], is
this knowledge application, on a creative way, arise new decomposed into a series of Single Machine Scheduling
computing science areas - Bionic Computing. Problems (SMSP)[9]. The Machine Agents (which has an
The complexity of current computer systems has led the NIT associated) obtain local solutions and later cooperate in
software engineering, distributed systems and management
SELF-OPTIMIZATION FOR DYNAMIC SCHEDULING IN MANUFACTURING SYSTEMS 423
order to overcome inter-agent constraints and achieve a detect changes occurred in the schedule, allowing the system
global schedule. to a dynamic adaptation. With this agent, the system will be
Two possible approaches, to deal with this problem, could prepared to automatically handle dynamism by adapting the
be used. In the first, the AUTODYNAGENTS system waits solutions to external perturbations. While, on one hand,
for the solutions obtained by the machine agents and then partial events only require a redefinition of job’s attributes
apply a repair mechanism to shift some operations in the and re-evaluation of the objective function, on other hand,
generated schedules till a feasible solution is obtained (Repair total events require changes on the solution’s structure and
Approach). In the second, a coordination mechanism is size, carried out by insertion or deletion of operations, and
established between related agents in the process, in order to also re-evaluation of the objective function. Therefore, under
interact with each other to pursuit common objective through total events, the modification of the current solution is
cooperation. These coordination mechanism must be prepared imperative, through job arrival integration mechanisms
to accept agents subjected to dynamism (new jobs arriving, (when a new job arrives to be processed), job elimination
cancelled jobs, changing jobs attributes). mechanisms (when a job is cancelled) and regeneration
mechanisms in order to ensure a dynamic adaptation of
population/neighborhood.
The Self-Optimizing Agent is responsible for the
automatically tuning of the meta-heuristics’ parameters,
according to the problem. This agent receives the initial
problem, or the changes detected by Self-Configuring Agent,
and automatically choose the meta-heuristic to use, and
makes its self-parameterization. If some dynamism occurs,
parameters may change in run-time. This tuning of
parameters is made through learning and experience, since it
uses a Case-based Reasoning (CBR) module. Each time a new
problem (case) appears, the CBR uses past experience in
order to specify the meta-heuristic and respective parameters
for that case. When the new case is solved, it is stored for
later use.
Fig. 1. AutoDynAgents Architecture
Finally, the Self-Healing Agent gives the capacity to the
system for diagnosing deviations from normal conditions and
The AUTODYNAGENTS architecture (Figure 1) is based
proactively takes actions to normalize them and avoid service
on six different types of agents. In order to allow a seamless
disruptions. This agent monitors other agents in order to
communication with the user, a User Interface Agent is
provide overall self-healing capabilities. Since agents may
implemented. This agent, apart from being responsible for the
crash for some reason, self-healing provides one or more
user interface, will generate the necessary Task Agents
agents backup registries in order to grant storage for the
dynamically according to the number of tasks that comprise
reactivation of lost or stuck scheduling agents with
the scheduling problem and assign each task to the respective
meaningful results, thus enabling the system to restart from a
Task Agent.
previous checkpoint as opposed to a complete reset. With this
The Task Agent will process the necessary information
agent, the system becomes stable, even if some deadlocks or
about the job. That is to say that this agent will be responsible
crashes occur.
for the generation of the earliest and latest processing times,
Rescheduling is necessary due to two classes of events:
the verification of feasible schedules and identification of
Partial events imply variability in jobs/operations attributes
constraint conflicts on each job and the decision on which
such as processing times, due dates or release times; and
Machine Agent is responsible for solving a specific conflict.
Total events imply variability in neighborhood/population
The Machine Agent is responsible for the scheduling of the
structure, resulting from new job arrivals, job cancellations,
operations that require processing in the machine supervised
machines breakdown, etc.
by the agent. This agent will implement meta-heuristic and
local search procedures in order to find best possible VI. SELF-OPTIMIZING MODULE
operation schedules and will communicate those solutions to The objective is that each machine agent adopt and
the Task Agent for later feasibility check. provides self-parameterization of the solving method in
Respectively to the Self-* Agents, the Self-Configuring accordance with the problem being solved (parameters can
Agent is responsible for monitoring the system in order to change in run-time).
424 MADUREIRA AND PEREIRA
Each machine agent must be able to define which NIT will the use of a CBR system.
be used and define initial parameters of NIT according to the The CBR is an artificial intelligence methodology which
current situation or even to commutate from one algorithm to aims to solve new problems by using information about the
the other according to current state and previous learning. solutions to previous similar problems [5]. It operates under
Agents self-optimize through learning and experience. the premise that similar problems require similar solutions.
Nature Inspired Optimization Techniques like Tabu The CBR consists in a cycle (Figure 2), usually described as
Search, Genetic Algorithms, Simulated Annealing, etc., are the ‘4 Rs’ [1]:
very useful to obtain good solutions in feasible execution 1. Retrieve the most similar case or cases;
times. Sometimes they can even obtain the optimal solution. 2. Reuse the information and knowledge retrieved to
But to be possible to obtain optimal or near-optimal solutions solve the problem;
it is required the correctly tuning of the parameters from the 3. Revise the proposed solution;
different meta-heuristics. It was proved to be a very hard task 4. Retain the revised solution to be useful for future
since it needs some expertise knowledge from the meta- problem solving.
heuristic in use and from the concerned problem. Sometimes There are two approaches to use the CBR in our system.
it needs the trial-error method and the difficulties increase The first one consists in retrieving the most similar case or
when there are more than one meta-heuristic to be used, cases to the new problem, regardless the meta-heuristic to
because it requires an a priori choice of the meta-heuristic use. Therefore, it is retrieved the case(s) containing the meta-
and then the tuning of its parameters. heuristic and its parameters to use. The second approach
consists in choose the meta-heuristic a priori and then
retrieve the most similar cases, containing the parameters to
use for that meta-heuristic. The first approach is more
appropriate to the problem in question since that is important
for the system to decide which meta-heuristic to use and the
respective parameters, because not all meta-heuristics are
suitable to every types of problems. Consequently, it will be
possible to know, for example, which meta-heuristics are
more suitable to a particular type of problem.
A. CBR Architecture
The architecture of the proposed CBR system is presented
in Figure 2. Initially, every new problem leads to a new case
for the system. In the first phase of the ‘4REs’ cycle, the
previous most similar cases are retrieved from the database.
As it is possible to see in Figure 3, the retrieved cases are
compared with the new case and two things can happen. If
Fig. 2. CBR system architecture. there is a very low similarity between them, the retrieved case
will consist in a set of pre-defined parameters, tuned from
To solve this problem, we present a module, described in expertise knowledge. If not, the most similar case is retrieved.
this section, representing the third part from Autonomic
Computing Self-CHOP, named Self-Optimizing. Kephart and
Chess [4] defined it as the capacity of “monitor, experiment
with, and tune their own parameters and learn to make
appropriate choices about keeping functions or outsourcing
them”. With this definition we propose an embedded module
capable of monitoring the system and tuning the parameters
of the different meta-heuristics, with respect to each new
single problem emerging on the system.
This module needs to know how to tune up the different
parameters of every meta-heuristic. It is impossible to predict Fig. 3. CBR Retrieve phase
every single problem to treat, so the system must be capable
of learning about its experience during lifetime, e.g., as After being recovered, the case can be reused becoming in
humans do. To perform this learning mechanism, we propose a suggested solution. After this, and in the Revise phase, the
SELF-OPTIMIZATION FOR DYNAMIC SCHEDULING IN MANUFACTURING SYSTEMS 425
CBR system executes the problem in the MAS, using the the parameters of the cases for tuning the respective NIT.
given solution. To escape from local optimal solutions and The CBR table, as referred before, saves the attributes of
stagnation, we propose the use of some controlled every single case in the CBR system. The fields are described
randomness in the parameters of the proposed solution. After in Table I.
the MAS execution end, the case are confirmed or not as a
good solution, and if so, it is retained on the database as a Table II – TS table fields description
new learned case, for future use. Field Data type Description
This represents the lifecycle of the proposed CBR system in NeighGen Double Neighborhood generation percentage
the Self-Optimizing module, as each new problem will be SubNeigh Double Sub-neighborhood percentage
TabuListLen Integer Tabu list length
evaluated and become a case of the casebase. We believe that StopCrit Integer Stopping criteria, number of iterations
the system will be capable of learning and evolving in the
parameterization of the meta-heuristics, adapting to every Table III – GA table fields description
new problem that may appear and improving the existing Field Data type Description
problems. InitPopGen Double Initial population generation
percentage
PopSize Double Population size percentage
NumGen Integer Number of generations
CrossRate Double Crossover rate
MutRate Double Mutation rate
Abstract-The analysis of stochastic systems is most difficult stages are obtaining the set of all the possible
notoriously hard, especially when their performance is states of a system and transition matrix between them. In
described by multidimensional Markov chains. the paper is used a method for automatic construction of
Approximation of general distribution functions by phase-type numerical models for systems described by Markov chains
distributions rise to large Markov chains. If the model is too with a countable space of states and continuous time.
large to analyze in its entirety, it is divided into To construct a model we need to describe the
subsystems. Each subsystem is analyzed separately and performance of a stochastic system in the event language
global solution is constructed from the partial solutions. [3]. It allows automating some stages of the model. The
Numerical algorithm to solve for the steady-state probabilities created software in C++ generates the set of possible states,
of these Markov chains from a system of linear equations by
decomposition method is presented in the paper. Based on
the matrix of transitions among states, constructs and
these probabilities, we can compute a wide range of relevant solves equilibrium equations to find steady state
performance characteristics, such as average number of probabilities.
customers of a certain type in the system and expected
postponement time for each customer class.
Keywords: Queuing system, Markov chain, decomposition II. APPROXIMATION OF QUEUING SYSTEM MODELS
method, steady state probabilities.
Continuous-time Markovian processes with countable
space of states are related with exponential distributions. It
I. INTRODUCTION means that transition time from one state to another must be
exponentially distributed. However, the mentioned
Complex queuing systems can arise in various practical condition often does not hold in the real life systems, so
applications, for example, telecommunication and computer additional difficulties are expected to arise in the
systems, production, logistics etc. Queuing models are investigation and modeling of such systems. To solve this
important tools for studying the performance of complex problem, an approximation of the stochastic systems by
systems, but despite the substantial queuing theory Markovian processes is used, when fictitious phase-type
literature, it is often necessary to use approximations in the method is applied [4-6]. The diagram of approximation of a
case the system is non-markovian. Use of phase-type (PH) general distribution G by phase-type distribution is shown
distributions is a common means of obtaining tractable in Fig. 1.
queuing models [1-2], but it rises to large Markov chains. The examples of approximation schemes of queuing
Markov modelling is a common approach to analyzing the systems are presented in figures 2 and 3.
performance of various stochastic systems. However, the
performance analysis of complex queuing systems has a
common source of difficulty: The Markov chain that
models the system behaviour has a state space that grows
infinitely in multiple dimensions. A method of
decomposition to analyze Markov chains of large size is
given in next sections.
It is known that creation of analytical models requires
large efforts. Use of numerical methods permits to create
models for a wider class of systems. The process of creating Fig.1. Approximation of general distribution G using phase-type
numerical models for systems described by Markov chains distribution.
consists of the following stages: 1) definition of the state of
a system; 2) creating equations describing Markov chain; 3)
computation of stationary probabilities of Markov chain; 4)
computation characteristics of the system performance. The
427
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_75, © Springer Science+Business Media B.V. 2010
428 VALAKEVICIUS ET AL.
number of class indicates the priority rank (class 1 has e5 –a medium priority customer was served in the first
highest priority and class N has the lowest priority).
Priority rule is nonpreemptive. Class i customers arrives in server with intensity μ m ;
the system according to a Poisson process with rate λi . The e6 – a low priority customer was served in the first server
service times of class i customers are exponentially
with intensity μ l ;
distributed with mean 1 / μ i . Within each class, service
discipline of customers is a First Come First Served e7 – a high priority customer was served in the second
(FCFS). Queue length of each class customers can have server with intensity μ h ;
various limitations.
e8 – a medium priority customer was served in the second
IV. NUMERICAL MODEL OF THE SYSTEM server with intensity μ m ;
Consider a queuing system with 3 priority classes – high, e9 – a low priority customer was served in the second
medium and low, and 2 servers, which we call first and server with intensity μ l ;
second. In addition, we consider limitation L ( L ∈ N ) for
summary waiting space of each customer class. The The set of transition rates between the states is the
system’s performance is described in the event language, following:
using methods from [1].
The set of system states is the following: {(
INTENS = λh , λm , λl , μ h , μ m , μ l )}
N = {(n1 , n2 , n3 , n4 , n5 )}, n1 + n2 + n3 ≤ L; n4 ≤ 3; n5 ≤ 3 , The created software generates the possible set of states
and matrix of transition probabilities.
n1 , n 2 , n3 , n 4 , n5 ∈ Z + , Calculation of steady state probabilities can require a
where large amount of calculations and computer resources. For
n1 – the number of high priority customers in the queue; example, if the set of states is described as
N = {(n1 , n 2 , n3 , n 4 , n5 )}, n1 + n 2 + n3 ≤ L; n 4 ≤ 3; n5 ≤ 3 ,
n 2 – the number of medium priority customers in the it is easy to prove that the total number of states equal to
queue; ( L + 1)(L + 2 )(L + 3)
⋅ 4 ⋅ 4 = ( L + 1)( L + 2)( L + 3) .
8
N =
n 3 – the number of low priority customers in the queue; 3! 3
For example, if L = 10 then system has 4576² states!
n 4 – indicates the state of the first server (0 – if server is
empty; 1 – if high priority customer is being served; 2 - if
A DECOMPOSITION METHOD IN MODELING QUEUING SYSTEMS 429
TABLE 1.
1
0,1
RESIDUALS OFLARGE G/M/1 MODELS
11
13
15
17
19
21
23
25
27
1
9
0,01
0,001 Matrix size Average residual
0,0001 n KMS Takahashi
Residual
1E-05 Blocks1
1E-06 Blocks2 123 1,4055E-12 1,6567E-12
1E-07
1E-08
243 3,5846E-011 9,6571E-012
1E-09 483 6,2866E-011 7,9955E-011
1E-10
1E-11
Iteration
REFERENCES
Fig. 9. Comparison of convergence in G/M/1 model
[1] G. Mickeviius, E. Valakeviius , “Modelling of non-
Markovian queuing systems”, Technological and economic
Applying the decomposition method with the optimal development of economy, vol . XII, No 4, pp. 295-300, 2006.
strategy of the partition into the block structure, the [2] T. Osogami , M. Harchol-Balter., “Necessary and Sufficient
approximate stationary probabilities of large G/M/1 models Conditions for Representing General Distributions by Coxians”,
School of Computer Science, Carnegie Mellon University,
were computed. The results require a lot space in the paper, 2003.
so Table 1 shows average residuals. The accuracy (from [3] E. Valakevicius, H. Pranevicius, “An algorithm for creating
Markovian Models of Complex System”, In Proceedings of the
10 −12 to 10 −10 ) remained as in models, where the size of 12th World Multi-Conference on Systemics, Cybernetics and
the transition probability matrix was around 15. The Informatics, Orlando, USA, June-July 2008, pp. 258-262.
number of iterations taken to reach this accuracy varies [4] M.A. Johnson, “An empirical study of queuing
depending on the size of the matrix, but there were no cases approximations based on phase-type distributions”, Commun.
Statist.-Stochastic Models, vol. 9, No. 4, pp. 531-561, 1993.
when algorithms exceeded the number of 30 iterations. [5] J.A. Luhman, M.J. Johnson, “Behaviour of queuing
approximations based on sample moments”, Applied Stochastic
Models and Data analysis, vol. 10, pp. 233-246, 1994.
VI. SUMMARY [6] M. J. Faddy, “Examples of fitting structured phase-type
distribution”, Applied Stochastic Models and Data Analysis,
vol. 10, pp. 247-255, 1994.
At the beginning of this paper we mentioned not only [7] S.W. Fuhrmann, R.B. Cooper, “ Stochastic decompositions in
G/M/1, but also M/G/1 models (accordingly G is Coxian the M/G/1 queue with generalized vacations”, Operations
distribution or the mixture of Erlang distribution). But the Research, vol. 33(5), pp.1117-1129,1985.
[8] A. Horváth,, G. Horváth, and M. Telek., “A traffic based
analysis of these models showed that the different structure decomposition of two-class queueing networks with priority
of the transition probability matrix doesn’t let partitioning service”, Computer Networks, vol. 53, pp.1235-1248, 2009.
into sub matrices. After applying the procedure of the [9] W. J. Stewart, Introduction to the Numerical Solution of
partition (even with simple condition – each row and Markov Chains”,.Princeton University Press, Princeton, New
Jersey, 1994.
column must have at least one nonzero element), resulting
matrix is the same as primary transition probability matrix.
Queuing systems requires more investigation in the
future.
Modular Design and Structure for a Mobile Sensory
Platform
Ayssam Elkady, Vipul Babariya, Jovin Joy and Tarek Sobh
School of Engineering
University of Bridgeport
I. I NTRODUCTION
A mobile manipulator offers the dual advantage of mo-
bility offered by the platform, and dexterity offered by the
manipulator. The mobile platform extends the workspace of
the manipulator. We are developing and constructing a mobile
manipulation platform called RISCbot II. The RISCbot II mo-
bile manipulator platform is shown in figure 1. Mobile manip-
ulators are potentially useful in dangerous and unpredictable
environments such as construction sites, space, underwater,
service environments, and in nuclear power stations.
Sensor fusion has been an active area of research in the field
of computer vision and mobile robotics. Sensor fusion is the Fig. 1. The RISCbot II mobile manipulator platform.
combination of sensory data from different sources, resulting
in better and more reliable information of the environment than
data derived from any individual sensor. Sensor fusion algo-
rithms are useful in low-cost mobile robot applications where
results can be merged by fusing the data from multiple sensors,
acceptable performance and reliability is desired, given a lim-
and presenting the result in a way that enables the tele-operator
ited set of inexpensive sensors like ultrasonic and infrared sen-
to perceive the current situation quickly. Sensor fusion is used
sors. Depending on the modalities of the sensors, sensor fusion
to reduce the workload for the operator to enable him or her
can be categorized into two classes (as described in [1]): sensor
to concentrate on the task itself. Sensor fusion is commonly
fusion using complimentary sensors, and sensor fusion using
used to reduce uncertainty in localization, obstacle avoidance,
competing sensors. Complementary sensors consist of sensors
and map building. Furthermore, sensor fusion may be used
with different modalities, such as a combination of a laser
to improve teleoperation by creating user interfaces which
sensor and a digital camera. In contrast, competing sensors are
efficiently facilitate understanding of remote environments and
composed of sensors suit which have the same modality, such
improve situational awareness.
as two digital cameras which provide photographic images of
the same building from two different viewpoints. Sensor fusion In this paper we describe our mobile manipulation platform,
has some critical problems, including the synchronization of and introduce the tasks that the RISCbot II is performing them
sensors. Different sensors have different resolutions and frame in, specifically teleoperation, navigation, obstacle avoidance,
rates, so the sensors need to be synchronized before their manipulation, face detection, and face recognition.
433
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_76, © Springer Science+Business Media B.V. 2010
434 ELKADY ET AL.
Fig. 9. Cylindrical Sensor Mount with rectangular plates on perimeter. Fig. 12. Structure Meshing
III. A PPLICATIONS
A. Face Detection
Efficiency of any detection process can be improved dra-
matically if the detection feature encodes some information
Fig. 11. Top Plywood sheet.
about the class that is to be detected. Our primary method
MODULAR DESIGN AND STRUCTURE FOR A MOBILE SENSORY PLATFORM 437
B. Face Recognition
Face Recognition is a task that identifies or verifies a
Fig. 14. Equivalent elastic Stain simulation result.
person from a digital image or a video source. The primary
method of face detection uses a overlapping energy histogram
as described in Face Recognition by Overlapping Energy
Histograms [3]. For the sake of efficiency, OpenCV library
was used extensively while developing the code.
1) General Approach: Face Recognition algorithm is a
two step process with a training stage and search stage. The
general approach is shown in figure 16. The training stage
is done initially to create a database which is not computed
again as it is a time consuming process. The search phase
uses the database created from the previous stage to look up
unknown images. Overlapping energy Histogram comprises of
taking DCT coefficients of every 8x8 pixel values with a 75
percent overlap as shown in figure 17 (i.e Taking DCT for
every 8x8 block after every 2 pixels). A sample of which is
shown in Figure 18. Once the image has been transformed
into a histogram Euclidean Distances as show in Eqn 1
is computed, where Ω is the feature vectors calculated via
the histogram. Further Euclidean Distance has been used in
Fig. 15. Total deformation.
many face recognition techniques [3]. This distance is used
to calcualte the Database Threshold(θ), which is used for
identifying and rejecting images at the search stage
of detection is based on Haar-Link Features (as described in
[2]) and OpenCV computer vision library. Haar encodes the 2
n = Ω − Ωn (1)
existence of oriented contrasts between regions in images. A
set of features can be used to encode the contrast exhibited by
human faces and their special relationships. C. Feature Extraction
Our detection program is derived primarily from OpenCV Once the image with a face is transformed using overlapping
and uses classifiers (cascade of boosted classifiers working DCT procedure values of Feature Set F2,
with haar-like features) that are trained with a few hundred
sample views of human faces. Classifiers used in our project F2 = [DC; AC01 ; AC10 ; AC11 ]
were obtained from OpenCV library. The classifier is built
using positive and negative examples of the same size. After a ,is put into a histogram with a bin size of 60. The real
classifier is trained it can be used to search across an image at challenge has been in selecting the appropriate appropriate
locations. The output of the classifier is 1 if it finds an object bin sizes for calculating threshold values. A general deviation
(face) in a region. The classifier is resized, rather than the for the algorithm was using a predetermined bin size of 60
image, to detect objects (faces) of different sizes. Therefore, which proved to be optimally good, both computationally and
to find faces of unknown sizes, the scan procedure is done for detection. This was determined by running a series of test
several times at different scales. on known images at the time of development.
438 ELKADY ET AL.
Fig. 19. Face Detection Module Detecting Multiple People in a Single Frame
Abstract — In this paper we present wradvs, a free open source graphical user interface, assists users during the configuration
application developed under the GNU General Public License process with tooltips, and validates all the submitted
that provides a comprehensive solution for IPv6 stateless information. Additionally, wradvs provides an Event Log
autoconfiguration. The application works as a service and Viewer with detailed information of the messages that are sent
implements RFCs 2460, 4861, 5006, 4191, 3775, and 3963. It also and received in the autoconfiguration process. Even if the main
has an event log viewer that records detailed information of all goal of the application is to support the teaching and learning
events in real time, allowing users to troubleshoot network process, it can be very useful for a network administrator that
problems without the need of additional tools. The main goal of uses IPv6 servers based on Windows. The rest of the paper is
wradvs is to be used as a didactic application in network
organized as follows: the IPv6 Router Advertisement message
advanced courses at Central University of Venezuela. Thus, it has
a friendly graphical user interface and presents a lot of valuable
is presented in Section II; IPv6 stateless autoconfiguration is
information to users while they configure all the parameters and discussed in Section III; related works are viewed in Section
options of the Router Advertisement messages. IV; wradvs is presented and justified in Section V; and finally
Section VI concludes the paper.
Keywords: ipv6; stateless autoconfiguration; router
advertisement; windows; service; daemon; didactic application. II. ROUTER ADVERTISEMENT
A Router Advertisement (RA) message is an Internet
I. INTRODUCTION Control Message Protocol version 6 (ICMPv6) message
Due to the imminent IPv4 address space depletion, defined by the Neighbor Discovery (ND) protocol [2].
LACNIC (Latin American and Caribbean Internet Addresses IPv6 routers send unsolicited RA messages pseudo-
Registry) is encouraging the adoption of IPv6 [5][13] as soon periodically and solicited RA messages in response to the
as possible. However, this adoption has been slow in receipt of a Router Solicitation (RS) message. The interval
Venezuela and in many other countries. One of the reasons for between unsolicited advertisements is randomized to reduce
the delay is the lack of IPv6 network specialists. Therefore, the synchronization issues when there are multiple advertising
training of IPv6 specialists has become an important issue. In routers on a link. RA messages contain the information
the undergraduate program of Computer Science at Central required by hosts to determine default gateways, the link
University of Venezuela (in Spanish: Universidad Central de prefixes, the link MTU, specific routes, home agent
Venezuela), some courses have been upgraded or added to the information, recursive DNS servers, whether or not to use
curriculum to face the problem. For example, Advanced stateful autoconfiguration, and the duration for which addresses
Network Protocols (in Spanish: Protocolos Avanzados de created through stateless address autoconfiguration are valid
Redes) is a new course that was introduced to the curriculum of and preferred [5].
the undergraduate program of Computer Science in 2005. Its
objectives include the understanding of IPv6 standards such as Fig. 1 shows the structure of the RA message which
the stateless autoconfiguration process described in [1]. contains the following fields:
Some daemons or servers have been developed, by the • Type: the value of this field must be 134 and identify
community or manufacturers, to support the stateless the type of ICMPv6 message (in this case, a RA).
autoconfiguration process. However, most of these daemons do
not support all the parameters and options of the stateless • Code: the value of this field must be 0.
autoconfiguration process, and sometimes, configuration can • Checksum: stores the checksum of the ICMPv6
be very difficult and the debugging system can be very poor. message.
These deficiencies can make the teaching and learning activity
process very laborious. For these reasons, we developed • Current Hop Limit: indicates the default value of the
wradvs (Windows Router Advertisement Server). wradvs is not Hop Limit field to be put in the IPv6 header for packets
only one of the most complete daemon created for IPv6 sent by hosts that receive this RA message. A value of
stateless autoconfiguration, but it also provides a friendly 0 indicates that the default value of the Hop Limit field
is not specified by this router.
443
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_77, © Springer Science+Business Media B.V. 2010
444 ABADI ET AL.
• Managed Address Configuration Flag (M): when set, The options that can be present in a RA message are the
it indicates that hosts receiving this RA message following:
must use a stateful configuration protocol, such as
DHCPv6 (Dynamic Host Configuration Protocol for • Source Link-Layer Address option: contains the
IPv6), to obtain addresses in addition to the link-layer address of the interface from which the
addresses that might be generated from stateless RA message was sent.
address autoconfiguration.
• MTU option: contains the recommended MTU for
• Other Stateful Configuration Flag (O): when set, it the link. Should be sent on links that have a variable
indicates that hosts receiving the RA message must MTU.
use a stateful configuration protocol, such as
DHCPv6, to obtain other configuration information. • Prefix Information option: contains the prefixes that
are on-link and/or used for stateless address
• Home Agent Flag (H): when set, it indicates that the autoconfiguration.
router sending the RA message is also functioning as
a Mobile IPv6 home agent on this link [3]. • Route Information option: contains more-specific
routes that improve the ability of hosts to pick an
• Default Router Preference (Prf): indicates whether to appropriate next hop for an off-link destination [4].
prefer this router over other default routers according
to the level of preference. Valid values in binary are: • Recursive DNS Server option: contains one or more
01 (high), 00 (medium) and 11 (low) [4]. IPv6 addresses of recursive DNS servers [6].
• Reserved: this field is reserved for future use and • Home Agent Information option: contains
must be set to 0. information specific to this router’s functionality as a
home agent [3].
• Router Lifetime: indicates the lifetime (in seconds)
of this router as a default router. A value of 0 • Advertisement Interval option: contains the interval
indicates that this router is not a default router; at which this router sends unsolicited multicast RA
however, all other information contained in the RA messages [3].
message is still valid.
III. IPV6 STATELESS ADDRESS AUTOCONFIGURATION
• Reachable Time: indicates the amount of time (in
milliseconds) that a host should assume a neighbor is The IPv6 stateless address autoconfiguration defines the
reachable after receiving a reachability confirmation. mechanism that allows a host to generate its own addresses
A value of 0 indicates that this router does not make and others network parameters using local information and
any recommendation about the reachable time. information advertised by routers in RA messages. This
mechanism requires no manual configuration of hosts. The
• Retransmission Timer: indicates the amount of time address autoconfiguration process [1] for a physical interface
(in milliseconds) that a host should wait before the of an IPv6 node is as follows:
retransmission of a Neighbor Solicitation message.
A value of 0 indicates that this router does not When an interface becomes enabled, a tentative link-local
specify the retransmission timer. address is generated based on the link-local prefix
(FE80::/64) and an interface identifier, which is commonly
derived from the link-layer address. The uniqueness of the
tentative link-local address is verified using the duplicate
address detection (DAD) algorithm. If the address is in use
by another host, the autoconfiguration stops and manual
configuration must be performed on the host.
WRADVS: A DIDACTIC SERVER FOR IPV6 STATELESS AUTOCONFIGURATION 445
Otherwise, the host sends a RS message to request an • Event log: refers to the existence of event registry
immediate RA message, rather than waiting for the next and how clear the events registered are.
scheduled RA message.
The studied daemons were: radvd, rtadvd, in.ndpd, the
The information advertised in the RA message (Current Cisco IOS, and netsh.
Hop Limit, Reachable Time, Retransmission Timer) is set.
For each Prefix Information option sent in the RA message, Table I shows a summary of the study for all the Router
the following actions occur: Advertisement daemons. For each criterion, a grade was
given between 0 and 5 stars, where 0 stars indicate a very
1. If the On-Link flag is set to 1, the prefix is added to poor or a lack of support, and 5 stars represent an optimum
the prefix list. or efficient implementation.
2. If the Autonomous flag is set to 1, the prefix and an radvd is one of the most popular Router Advertisement
appropriate interface identifier are used to generate daemons. Most of the stateless autoconfiguration parameters
a tentative address. and options can be specified with radvd. Configuration is
done through a text file; there is no graphical user interface.
3. Duplicate address detection is used to verify the
The syntax of the configuration file is easy to read and learn.
uniqueness of the tentative address.
radvd provides an efficient and informative error handler; the
4. If the tentative address is in use, it is not assigned to depuration process shows information about the daemon’s
the interface. events. However, this information is not specific enough.
The depuration information can be saved in a log file for
A RA message could also include other options, such as: posterior analysis. We tested radvd on Debian GNU/Linux 4.
Source Link-Layer Address, MTU, Route Information,
Recursive DNS Servers, Home Agent Information, and rtadvd is another Router Advertisement daemon mostly
Advertisement Interval. used on FreeBSD and Mac OS. It supports most of the
stateless autoconfiguration parameters and options. The
IV. RELATED WORKS syntax of the configuration file (text file) is hard to read and
learn, the depuration information does not provides valuable
A Router Advertisement daemon listens to RS messages information (e.g., value of the Router Advertisement’s fields)
and sends RA messages on demand and periodically as and it does not allow a custom path for the log event. We
described in [2]. These advertisements allow any listening tested rtadvd on FreeBSD 7.0 and Mac OS X Leopard.
host to configure its addresses and some other parameters
automatically through IPv6 stateless autoconfiguration. in.ndpd was developed by SUN Microsystems for
Solaris. It is compliant with RFC 2461 which was deprecated
We studied several Router Advertisement daemons of by RFC 4861. in.ndpd just supports an incomplete set of
different operating systems to make a comparative study. stateless autoconfiguration parameters and options. The
The criteria for evaluating these daemons were: debugging process displays important and well formatted
• Installation: refers to the easiness of the installation information. It was tested on Solaris 10.
process. Some versions of the Cisco IOS (Internetwork Operating
• Configuration: refers to the easiness of the System) include a Router Advertisement daemon. This
configuration of parameters and options. daemon supports just an incomplete set of stateless
autoconfiguration parameters and options; however, it offers
• Supported parameters: takes in consideration the a very good debugging system. Since the IOS is a
amount of fully supported parameters and options. commercial product, no source code is available.
• Graphical user interface: refers to the existence of a netsh is a command-line scripting utility that allows
graphical user interface and its friendliness. (among other things) sending RA messages in a Windows
system. It does not have a graphical user interface. It only
• Error handling: refers to how specific and clear the permits to specify a few parameters and options of the
error messages are when there are troubles in the stateless autoconfiguration process. The error messages are
configuration. very poor. The source code is not available.
For teaching and learning purposes, support of all software development [7]. Scrum method is divided in
parameters and options of the stateless autoconfiguration Sprints. A Sprint is a period of time, where every phase of
process is an important issue. Another important issue is the the development occurs. The four main modules were
availability of the source code so students can improve it and developed in the first four Sprints, starting with the wradvs
adapt it to specific needs. Additionally, a friendly graphical Server that is installed as a service, and finishing with the
user interface can be very useful for non-expert students. construction of the detail of the events that occur in real time,
Since none of the existing daemons offers all the required showing a hexadecimal dump of the messages that are sent
features, we decided to develop our own server. and received. Then, there was a Sprint for the creation of the
Tooltips that gives all the information of the parameters,
V. WINDOWS ROUTER ADVERTISEMENT SERVER options and protocols that can be configured, making the
application an excellent teaching tool. The following Sprint
wradvs (Windows Router Advertisement Server) is a was introduced to improve the feedback to users and make
Router Advertisement server for Windows developed under the application more intuitive. Finally, there was a Sprint for
the GNU General Public License. It supports RFCs 2460, testing and debugging all the software.
4861, 5006, 4191, 3775, and 3963. It has a friendly graphical
user interface, checks entered values for consistency, offers a The application uses two external projects that were
good debugging system for troubleshooting purposes, and necessary to accomplish some functionality:
shows tooltips to guide users during the configuration
• WinPcap: an industry-standard tool for link-layer
process. At the present time, wradvs has an interface for
network access in Windows environments. It allows
English and Spanish users. Switching from one language to
applications to capture and transmit network packets
another can be done with just a few clicks.
bypassing the protocol stack, and it has additional
useful features; including kernel-level packet
A. Methodology filtering and a network statistics engine [9].
We wrote the application with Microsoft Visual C#
Express Edition [8]. wradvs was divided in four main • SharpPcap: a packet capture framework for the .NET
modules: the wradvs Server, the Server Configuration Tool, environment based on the famous WinPcap libraries.
the Log Viewer and the IPv6 Address Configuration Tool. It provides an API for capturing, injecting, analyzing
Fig. 2 shows a simplified class diagram of the application. and building packets using any .NET language such
as C# and VB.NET [10].
To develop wradvs, we used Scrum, an iterative
incremental framework for managing complex work (such as
new product development) commonly used with agile
Abstract - As wireless devices become ubiquitous in society, involves the movement of goods from large centralized ports to
proper security methodologies must be deployed that are other ports by utilizing cargo ships that traverse the oceans and
cognizant of the physical limitations inherent in the device design other waterways. The maritime shipping industry relies on
in order to ensure that the information produced in these systems shipping freight that is containerized, meaning that the cargo is
is not abused. One such area of abuse involves an attacker
sealed by the owner in a standardized metal container and is
locating and disabling a wireless device that is transmitting a
relatively large amount of data compared to the other devices in not palletized or packaged by the maritime shipping provider.
the same physical area; this large amount of transmitted data This containerization has allowed the shipping industry to
may indicate that the wireless device is important or valuable to achieve great efficiencies as the whole system has become
the owner. standardized.
In this work, we propose and test a novel design that mitigates There are only a few standard sizes of containers that are
this threat as related to P2P wireless communication systems that used throughout the industry (e.g., 8’x8’x20’, 8’x8’x40’, etc).
have battery life as a limiting constraint. Our solution uses The container shipping industry measures freight in terms
specialized honeytoken devices that transmit realistic noise in twenty foot equivalent units (TEUs), which represents the
order to mask the presence or absence of communication from the
capacity of a single 20’x8’x8’6” container. Estimates for 2008
actual wireless devices. This will allow the wireless devices to
only transmit when required, conserving battery life and adding indicate that over 18 million containers make over 200 million
to the security of the system by making it impossible to locate a trips per year, which account for approximately 90% of non-
highly active device. The honeytoken noise is easily removed by bulk cargo shipped worldwide [6]. Currently, the largest
trusted insiders; however it can not easily be removed by container transport ship is the Emma Maersk which can
attackers. The idea of deploying honeytoken devices to protect transport over 11,000 20-foot containers [5]. The ports are
P2P devices with battery life constraints is a new approach, and impressive locations, for example, the Port of Los Angeles
our key contribution is the proposal, design, and testing of this encompasses 7500 acres, 43 miles of waterfront and features
system. 27 cargo terminals.
Container shipping via large cargo ships is an industry that
I. INTRODUCTION transports millions of tons of cargo on hundreds of cargo ships
In the following sections we identify the maritime system in to destinations all over the world. In a typical journey, a
which we are deploying wireless peer to peer sensors in order shipper loads goods into a container and then seals the
to better track and secure intermodal shipping containers. We container. Next, the container is transported via truck to a port
also define several key properties that all honeytokens possess, where it is loaded onto a container ship for overseas transport.
regardless of the type of object they are protecting. We also Once the ship arrives at its destination port, the container is
present the motivation for using honeytokens to better protect unloaded and again loaded onto a truck for transport to its final
our wireless P2P system. destination where it is unsealed and unloaded. Due to the value
of the cargo and the potential for terroristic actions such as
A. Maritime Shipping Industry smuggling contraband, poisoning food cargo or stealing
The shipping industry is a complex system, composed of contents, the container must remain secure throughout its entire
many subsidiaries that allow and facilitate the transport of journey. Also, the container must remain at a safe temperature
goods from manufacturers to retail outlets or other entities in for refrigerated goods and the containers should not experience
the supply chain. In this work, we are focused on one a high level of acceleration for fragile goods. A system that
particular subsystem, the maritime shipping industry, which can monitor the position of the container at a port or on a ship
449
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_78, © Springer Science+Business Media B.V. 2010
450 WHITE ET AL.
and whether or not the container has been breached will battery constraints, the devices will only transmit when they
provide shipping personnel with additional information to have information to communicate or are passing on
ensure the secure transport of the containers and the goods information received from another node. This allows for fault
within the container. Our system provides part of this security tolerance and long battery life, which are two requirements of
and is described in section 3. the system.
However, the proposed system exposes a security risk.
B. Honeytokens While an attacker may not be able to understand what the
Honeytokens are an information security tool that operates devices are communicating, they can easily locate the devices
on the theory of deception. Honeytokens are related to that are communicating the most as most devices will be in a
honeypots, which are systems whose value lies in attracting, sleep state the majority of the time. For example, when a
distracting, and potentially tracing attackers by presenting a container is breached it will raise a signal that will propagate
seemingly worthwhile, but unprotected, computer. through the system. As this is a rare event, if an attacker has
Honeytokens are different from honeypots in that they aren’t been monitoring the wireless communications with a high
an actual device. Instead, they are seemingly worthwhile and powered receiver and suddenly notices a spike in the received
valid pieces of digital information. While organizations have traffic, they can be assured that something important has
sophisticated firewalls and intrusion detection systems in place happened. Also, if a device is consistently being queried for
to detect physical attacks on their networks or data systems, the information by the owner, this device will also communicate
loss of information posses a much greater threat, especially in relatively more data than the other containers in the P2P
wireless systems like ours [1]. network, making it stand out from the rest. An intelligent and
The term ‘honeytoken’ was coined by Augusto Paes de motivated attacker could potentially target this container for
Barros in 2003 in a discussion on honeypots. Honeytokens are malicious actions, such as stealing the contents, as it appears to
manufactured by security engineers to look attractive to an be valuable to its owners.
attacker [1], hence the name. Examples would be a customer We are going to use honeytokens to mitigate this risk. The
database of personal information, a list of passwords, honeytokens will be deployed throughout the ship and they
proprietary design documents, a list of credit card numbers, or will produce realistic looking noise; the communication from
a seemingly worthwhile wireless signal passing through a P2P the honeytokens will be very similar to the actual
network. communications produced by the sensors. Each honeytoken
Honeytokens have common properties, regardless of the device will be given a unique identification value so that they
digital object they mimic. First, the honeytoken must appear can be separated from the actual data produced by the P2P
valuable to an adversary [2]. If the honeytoken does not sensors. The honeytoken identification values will be known
appear worthwhile, it will never be interacted with, stolen, or by the controlling system, and the controlling system will
examined. As wireless systems become ubiquitous, the notify the sensors to not pass the noise produced by the
information that they produce will become valuable as it will honeytokens through a learning phase when the ship is loaded.
be used in many different areas to communicate important The noise will appear at random times and at random
data. durations, effectively masking the actual signal
Second, the honeytoken must be designed in such a manner communications that the sensor nodes produce. Because there
so as to not interfere with any real data. The implementers of will always be a high level of realistic noise in the RF
the honeytokens must be able to tell what is real and what is environment, an attacker will no longer have the ability to
not, and thus the honeytoken must have unique, identifiable single out an individual container, adding to the security of the
properties. If the honeytokens can not be separated from the system. The use of honeytokens also allows the actual P2P
real data, the whole system suffers and the accuracy of the sensors to transmit only when they have actual information to
system is compromised. send, conserving battery life.
Third, honeytokens are implemented so that they appear The rest of this paper is organized as follows. Section 2
real to an attacker. If the honeytokens are obviously an details some previous work in the area of P2P security and
imitation, the attacker would have no reason to use or examine honeytokens. Section 3 describes the current P2P wireless
them. If the honeytoken is easily detected, it might also be system and our proposal for using honeytokens. Section 4
easily removed by the attacker while still accessing the actual shows how we simulated the system and took comparison
data that the honeytoken was designed to protect. Secrecy is of measurements on a system that uses honeytokens and with a
the utmost importance because if an attacker is able to find out system that does not, and section 5 presents these results.
what objects are honeytokens, they can easily ignore them [3]. Section 6 concludes the work.
honeytokens and also other honeypot methods that have been equipment that requires frequent maintenance will not perform
used to secure P2P systems. well within the economic constraints of the shipping industry.
Honeytokens have been used in personally identifiable Solid state electronic components are generally rugged enough
information (PII) databases to detect unauthorized access [1]. to withstand the application as long as they are mounted to the
In many industries, privacy laws prohibit individuals from container in a position that is not vulnerable to damage by the
accessing personal information that they don’t have a valid use equipment used for loading and unloading. However, the
for. Detecting when unauthorized access has occurred is power requirements of the instrumentation must be addressed.
difficult, as the data requirements for applications and users Typically, containers do not have access to a direct, always on,
vary greatly. power source. Therefore, each wireless node must rely on
Honeytokens are known as ‘copyright traps’ when used in battery power, which must last several months to prevent
the cartographic and dictionary industries [3]. Mapmakers are excessive maintenance requirements.
known to insert fake streets, misspelled objects, and non- Each container node periodically broadcasts its status
existent bends in rivers to catch illicit copying. Dictionaries (breached or secure) to all the container nodes that are
have also inserted fake words into their dictionaries in order to listening. For example, the sensor nodes may broadcast a
catch copyright violators. Both of these uses of honeytokens system secure message every hour on the hour; also the sensors
have been shown to work [3]. will wake every ten minutes to listen for breaches or other
Honeypot systems have been used to analyze P2P botnets messages from the system. If a breach message is received, the
[7]. The honeypot attracts the bots and the actions that are message is passed on throughout the P2P network until it
performed are analyzed. If the bot interacts with the honeypot reaches the intended recipient. This is a basic overview of the
for an extensive period of time, many of its abilities can be P2P environment in which we are operating in. While the
identified and analyzed. Honeytoken email addresses have focus of this work is not the implementation of the network
been used to monitor distributed P2P spam networks [8]. The itself, it is necessary to have an appreciation of the physical
addresses appear valid, but they are present just to attract and constraints on the devices in order to understand the security
waste attacker’s time and resources. Many other approaches risks inherent in this design.
have been used to secure P2P communications [9], [11], but
little work has been done on using honeytokens to add noise to B. Honeytoken Deployment
the system and to distract attackers from the actual The honeytokens will be deployed throughout the RF
communications. environment on the boat. They must be deployed in sufficient
amounts in order to effectively cover the actual
III. PROPOSED SYSTEM communications that the sensors are transmitting. Also, as the
In the following sections, we describe the P2P wireless sensors are synchronized to transmit and/or receive only at
maritime system. As the sensors are going to be deployed onto certain predetermined discrete intervals, the honeytokens must
shipping containers, there are several physical constraints that also be attuned to this. The honeytokens must produce noise
must be accounted for. Also, we describe in detail our that appears realistic to an outsider when compared to the
proposal for deploying the honeytoken devices throughout the actual transmissions, however since the communications are all
RF environment on the ship. encrypted, this will be a straightforward process.
The honeytoken devices themselves will be simpler than the
A. P2P Wireless Maritime System P2P devices. The honeytokens will have a transmission only
The container network is created by instrumenting each radio, a battery, and a small processor that can store the
container with intrusion detection hardware and a wireless device’s assigned identification number and perform the
network device. The device will have a battery, radio, necessary calculations in order to produce random noise at a
temperature sensor, light sensor and accelerometer in order to specified time slot. The honeytokens will run on a separate
perform the necessary physical security requirements. Due to battery based power supply from the P2P devices so as to not
the limited range of the wireless devices, many containers will have a negative effect on the battery of the actual devices.
not be able to communicate directly to the ship or port There will be no sensors or other physical interfaces as the
network. However, the containers are packed tightly together honeytokens are present to only produce noise.
and each container will be in range of several neighbor The honeytoken devices can be placed on physical
containers. Typically, containers are stacked several units high containers or throughout other areas of the ship. When the
in rows on the ship or port. honeytokens are placed throughout the ship, the controlling
Additionally, some of the containers will be in range of the computer records the identification number of the honeytoken
ship or port network. Therefore, all co-located containers, and notifies collocated devices about their presence. If an
either on a ship or a port, will form an ad-hoc peer-to-peer actual P2P device retransmits the noise, it will eventually reach
(P2P) network. Part of the cost-effectiveness of the the ship’s computer. The computer will have access to a list of
containerization shipping process is the fact that individual the honeytoken identification values and it will be able to
containers are designed to be rugged and require very little determine that this message is in fact noise. The shipboard
maintenance. Therefore, the addition of sensitive electronic computer will then send a message to the device that
452 WHITE ET AL.
retransmitted the noise alerting it to the fact that this device is communications. We compared these four scenarios against
in fact a honeytoken. In this way, all trusted devices can learn the system with no honeytokens.
what devices are actual sensors and which devices are present The attacker only has a limited range in which they can
just to produce noise. This method of learning what devices receive signals from the sensors successfully. For our tests, we
are honeytokens will require extra communication during the assumed that the attacker was at a point in the middle of the
setup phase, but once the honeytokens are learned, the P2P ship and that the attacker’s receive antenna had a range of 50
devices can ignore the noise while still being able to retransmit meters [10]. In reality, an attacker could also be scanning from
valid messages. a port at the side of the boat, and we will take this into account
As mentioned previously, honeytokens must appear in a future work. In order for the attacker to receive a
valuable, realistic, and must not interfere with the valid data. transmission from a sensor or honeytoken, the receive antenna
The honeytokens on our system are valuable and realistic has to be less than 50 meters from the transmission source, or:
because they produce the same type of signals that the actual
P2P devices transmit. An attacker does not have the ability to
remove this noise, and they can no longer locate a signal (S x − I x )2 + (S y − I y )2 + (S z − I z )2 < 50 m
device that is transmitting a large amount of data. As each of
the honeytokens has a unique identification value that is only where (Sx, Sy, Sz) represents the spatial coordinates of the
known to the trusted devices, each device can correctly attacker’s antenna, and (Ix, Iy, Iz) represents the spatial
identify what is and what is not noise, keeping the devices coordinates of a particular sensor. The above equation gives
from interfering with each other. the distance the object is from the antenna if a line is drawn
from the sensor to the antenna in three dimensional space.
IV. SYSTEM SIMULATION For all of the simulations, we ran the same series of
In the following section, we detail the steps that were taken simulated messages. For the simulations, the P2P sensors
to test our hypothesis that honeytokens could be used to mask broadcast a system secure message every 60 minutes, unless
the presence of devices that were transmitting a lot of data the container had been breached, in which case a message is
relative to other devices in the P2P network. First, we detail sent stating that fact. Also, the sensors wake every 10 minutes
the physical setup of the simulation with several thousand for 30 seconds in order to listen for breach signals or other
shipping containers, each with an affixed P2P device. We then transmissions from the system. We simulated the probability
describe the placement and amount of deployed honeytokens. of each container needing to transmit a message at a given time
We perform several experimental simulations using various slot as 5%, which is what we expected the probability to be in
amounts of honeytokens. We conclude by describing the actual usage on the ship as the breach and other events are rare
message set that was used to compare the systems. occurrences. We then simulated 2.5% of the sensors as “hot”
sensors. These sensors communicated at a much higher rate,
A. System Setup even though it was still a rare event for the system overall; for
For our experiments, we simulated one ship with several our experiments we used a Poisson distribution with a lambda
thousand containers, each with a P2P sensor attached. As value of one. Again, this relatively small amount of “hot”
mentioned previously, the Emma Maersk is the largest sensors is what was expected in actual usage. As the
container ship currently sailing [5]. It is capable of holding honeytokens only produced noise, we allowed them to all be
approximately 11,264 twenty foot equivalent containers. We considered “hot” sensors, with a Poisson distribution that was
located the dimensions of the ship and placed each of the skewed towards a higher transmission probability. If we
11,264 containers at a specific place on the ship. The ship is allowed the honeytokens to transmit all the time, this would
397 meters in length, and 56 meters wide. Based upon the easily be removed by the attacker as it would simply raise the
dimensions of the ship and other technical data that was noise floor threshold; instead, we allowed the honeytokens to
obtained, the simulated containers were stacked 8 tall by 22 act like the devices that were transmitting the most, which
wide by 64 containers in length. We then simulated placing a made the noise harder to remove as it looked like the real
sensor on the outside of each container, each having a specific thing. We ran each simulation for a total of 3 hours, so as to
location in three dimensional space. collect a sufficient amount of data. At each ten minute interval
when the sensors woke up, measurements were taken on the
B. Test Setup amount of simulated transmissions that the attacker would
Next, we simulated various amounts of honeytokens intercept given the constraints identified above. As we get
distributed throughout the ship. The honeytokens were placed more real world usage data, we will adjust these simulation
randomly throughout the ship on the exterior of the containers properties to accurately reflect the environment.
and at the side of the ship’s deck. We ran simulations with 50,
500, 1000, and 5000 honeytokens. We anticipated that if only V. RESULTS
a small amount of honeytokens were used then the produced We first ran the simulation with no honeytokens present to
noise would not be enough to effectively mask the actual get a baseline comparison of how the system operates. As
expected, there were certain peaks in the transmissions that
SECURING P2P WIRELESS COMMUNICATIONS BY DEPLOYING HONEYTOKENS 453
Number of Transmissions
500
of communication is over twice what is normally transmitted.
400
Time frames one, seven, thirteen, and nineteen are the highest
300
communications where all sensors transmit a signal to indicate
that they are still operational. As this is a system wide event, it 200
0
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19
Transm issions w ith no Honeytokens
Time
600
Figure 3: Simulation results with 500 honeytokens.
500
400
The results with the usage of 500 honeytokens are
presented above in Figure 3. The communications are
300 beginning to appear much more random; however the largest
200 amounts of communications can still be seen at the fourth,
twelfth, and sixteenth time frame. However, the other time
100
frames have a much more similar distribution, and the peaks
0 are beginning to even out.
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19
T ime ( 10 minut e int er val s)
1000 Honeytokens
Figure 1: Intercepted transmission with no honeytokens.
600
Number of Transmissions
500
As mentioned previously, our goal of using the
honeytokens is to obfuscate the actual communications. Our 400
still producing noise at the same rate as the most active 100
Abstract-Due to globalization, stronger competition, in- sharpened, the IT governance requirements and thereby data
creased complexity, information explosion, interconnection and integrity requirements are increased and based on the main
extensive use of technology, information management is a main role of data, information and knowledge the requirements for
performance driver and key differentiator for sustainable or- confidentiality and availability are increased.
ganization success. Data, information, knowledge and asset are
exposed to most different threats. Therefore more than 5800
In this respect information and system availability, confiden-
organizations worldwide are implementing an information se- tiality and data integrity, as well as data protection and ful-
curity management system in accordance with ISO/IEC 27001. fillment of legal and regulatory requirements is central for
Many organizations of different sizes are implementing also organizations’ success.
standard based management systems, such as ISO9001 or oth- One million organizations of different sizes and scopes are
ers. On the basis of several practical applications in most dif- implementing, already since several years, quality manage-
ferent organizations are described applied methods and expe- ment systems, such as ISO9001, or other management sys-
riences by implementing suitably holistic management systems tems based on international standards (e.g. environment
for risk prevention and quality improvement. This promotes ef- ISO14001, IT service management ISO22000). All these
fectiveness, efficiency, legal conformity, information security
awareness and organization development including improved
management systems require common principles: organiza-
information security for sustainable organization success. tion objectives and strategies, business processes, resource
management and continuously optimization of the organiza-
tion [4]. The established management system must be docu-
Keywords: information security, quality management, holis-
mented, communicated, implemented and continuously im-
tic enterprise management, data protection, risk management
proved. These systems are implemented more frequently ho-
listic, whereby are integrated according with the organiza-
tional purpose and objectives different aspects, like quality,
environment, hygiene, occupational health and safety, as
I. INTRODUCTION well as personnel development, resource management, IT -
management, communication management, controlling and
A Starting Situation also knowledge management.
Due to globalization, ever stronger competition, information
explosion, increasing interconnection, extensive use of IT- B Purpose of the article
systems and the growing complexity data and information
management for knowledge generation and continual im- The increased requirements for information security includ-
provement are main performance driver and key differentia- ing availability, confidentiality and integrity, international
tors for competitive advantages and sustainable organization standards such as IEC/ISO 27001 [5], IEC/ISO 27002 [6],
success. Thus data, information, knowledge, IT and produc- ISO/IEC 20000-1, Sarbanes-Oxley or other legal or regulato-
tion systems become a completely new role. But data, in- ry requirements, the improved risk awareness promote a ho-
formation and knowledge are exposed to most different listic information security approach. There are just now more
threats, such as physical and environmental threats, technical than 5800 IEC/ISO 27001 information security systems im-
threats, organizational and human related threats. The Com- plemented and certificated [7]. The information security
puter Security Institute (CSI) and the FBI describe that virus management systems are often implemented parallel to other
attacks are the source of the greatest financial losses, fol- standard based management systems. The trend to holistic
lowed by unauthorized access to networks, lost or stolen lap- management systems and the need of developing an inte-
tops or mobile hardware, and theft of proprietary information grated solution to support optimally all requirements of an
are the next and all together was reported a lost account for organization and to promote a consistent and integrated im-
more than $39,370,717 [1]. The kes Microsoft study and the plementation and operation influenced us to integrate an in-
report about the IT security situation in Germany explain formation security management system with an ISO 9001
that collaborators error and negligence are the main cause for quality management system. This will enrich management
damage followed by software and hardware problems [2], systems by information security aspects, use common syner-
[3]. In the last years the data protection law requirements are gies and resources, and it will contribute efficiently, re-
455
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_79, © Springer Science+Business Media B.V. 2010
456 STOLL
sources carefully, fast and successfully to support informa- Thus information security and thereby availability of data,
tion security for sustainable organizational development and documents, asset and infrastructures, confidentiality of in-
success. formation, documents and data, as well as data integrity
should be optimized and the organization will be promoted
C Research Approach in achieving its objectives including information security ob-
jectives and reducing risks.
We developed a holistic information security management
model by integrating the ISO 9001 standard for quality man-
agement systems or other international standards for man-
agement systems and the information security management III. MAIN STANDARD REQUIREMENTS
standard in accordance to IEC/ISO 27001 and IEC/ISO
27002. This model was implemented after in different small A Main Requirements of Standard Based Management Systems
and medium sized organizations. The case study results were
collected by established process and system measurement The ISO 9001 quality management standard and other stan-
methods and interviewing the managers and collaborators. dards for management systems require common principles
[4]:
D Structure of the Article • Organizations’ objectives and strategies must be estab-
lished regarding stakeholder requirements [see top of
Firstly we describe based on the starting situation the project Fig.1].
objectives [II]. After we explain the requirements for our ap-
• All business processes including the management process,
proach based on the main requirements of quality manage-
support processes, resource processes and optimization
ment systems (ISO 9001) and other international standards
processes must be defined and promote the optimized ful-
for management systems [III A]. Also the main requirements
filment of the organizations’ objectives under the focus of
for information security management in accordance to
the respective standard [see the graphic under the top of
IEC/ISO 27001 and IEC/ISO 27002 [III B] are illustrated.
Fig. 1].
After we present our holistic information security manage-
ment model [IV] and the implementation of that model [V] • Process oriented resource management must be promoted
with the development of policy and objectives [V A], the including human resource development, IT – management
risk assessment, risk treatment and business continuity plan and other infrastructures, tools and instruments [see bot-
[V B], the process analysis and process improvement [V C], tom of Fig. 1].
the resource management [V D], the continually improve- • The organization, their objectives and strategies, services/
ment of the whole system and the system documentation [E]. products and processes must be continually optimized ac-
Finally we report about our project experiences and results cording to established processes in sense of a PDCA cycle
of the implementation in different small and medium sized (plan, do, check, act) [see the circle around in Fig.1].
organizations [VI] with the achievement of the project prin-
ciples [VI A] and we reflect about success factors [VI B]. At The established management system must be structured and
the end we give an outlook [VII] and our conclusions [VIII]. systematically documented and communicated within the
organization and the collaborators must be continually moti-
vated for implementing the system and for recommending
improvements.
II. PROJECT OBJECTIVES These standard based management systems are implemented
more frequently in a holistic way. In accordance with the or-
How can be integrated the requirements of information secu- ganizational purpose and objectives are integrated into a
rity management in accordance to IEC/ISO 27001 and management system eventually environmental, hygienic, oc-
IEC/ISO 27002 and the requirements ISO 9001 quality man- cupational health and safety aspects, as well as human re-
agement systems or other standards based management sys- source development, resource management, knowledge
tems in order to contribute to organizations’ success and to management, IT - management or controlling.
be considered, used and improved as base for organization
development?
By integrating information security management in accor- interested parties
dance to IEC/ISO 27001 and IEC/ISO 27002 into a, the or- vision
policy
ganization best adapted, holistic information security man- objective, priority, strategy
agement model we expect to foster:
service/product realization
• effectiveness, resource management
• efficiency and cost reduction,
• shorter initial training periods for new collaborators, and
• improved information security. Fig. 1. Main requirements of standard based management systems
INTEGRATING INFORMATION SECURITY INTO QUALITY MANAGEMENT SYSTEMS 457
§
We present the requirements of ISO/IEC 27001 [5] and Interested Parties
Vision
ISO/IEC 27002 [6]: Policy + IS
Objectives, Strategies + IS
• An information security policy must be defined and ap- P o te
Im pac
o f n ti a l A m
Le v D a m a o un t
e ls g ,
o f R ise
ks
ts
P o te n t i a l
T h re a ts
L i k e l ih o o d
A sset
P r e v e n t iv e
A c t io n
I n te g r it y
R i s ik s
A c c e p ta b le ?
Re
R iss id u
ks a l
Services,
• A risk assessment must be conducted to establish a risk Production + IS
treatment plan to reduce risks to acceptable levels of risk. Ressource Management
+ IS
For the identified remaining risks a business continuity
plan must be developed, implemented, maintained, tested
and updated regularly.
• The needed resources must be determined and provided. Fig. 2. The Holistic Information Security Management Model
security management system scope and boundaries. We de- ing, generating, changing, handling and the accountable for
duce after from the policy the objectives and strategies inclu- it, as well as the functions, which will collaborate and/or be
sive information security objectives, such as availability, informed. Also external information sources or receivers
confidentiality and integrity. Also all legal, regulatory and (document logistic), the data protection class in accordance
corporate information security requirements are analyzed to the data protection law and further the required archiving
and adequate information security strategies deduced. Thus methods (for current, intermediate and historical archives)
the entire organization is focused on the fulfillment of stake- are identified. Thereby we establish necessary and licit
holder requirements including information security objec- access rights, archiving procedures, information security re-
tives and requirements. quirements regarding confidentiality, availability and integri-
ty, necessary data encryption with encryption procedures, as
B Risk Assessment, Risk Treatment and Business Continuity Plan well as signature rights and signature procedures as basis for
the implementation of adequate signature regulations. All
A risk assessment is conducted to establish a risk treatment treated data are examined for there necessity and lawfulness.
plan to reduce risks to an acceptable level of risk. Therefore Thus all information flows, documents, data and document
we identify the assets, define the required levels of confiden- handling and all processes are analyzed regarding informa-
tiality, integrity and availability for the assets, identify the tion security and optimized.
potential threats to those assets, define the realistic likelih- Function profiles and required competences are deduced
ood of security failures, the impacts that losses of confiden- from the role of the different functions at the single process
tiality, integrity and availability may have on the assets, the steps.
controls currently implemented, estimate the levels of risks
regarding the implemented controls and elaborate for the
higher risks adequate risk treatment plans. The management D Resource Management
must approve whether the risks are acceptable or further risk
treatments must be implemented. For the risk treatments ap- The organization must determine and provide due to stan-
propriate risk control objectives and controls are selected dard requirements necessary resources, tools and instruments
and approved. The risk treatment plan is integrated into the to obtain the established objectives and to continually im-
operational processes. Thereby the risk treatment and the ex- prove the organization. Thereby an optimal infrastructure for
isting standard based management system are integrated. For information security is promoted. Training and competence
the identified remaining risks a business continuity plan is objectives are planed and implemented according to defined
developed, integrated into the operational processes. Thus it processes and their effectiveness are evaluated. Thus also the
is implemented, maintained, tested and updated regularly to competences and awareness of all collaborators and persons
ensure that it is effective to maintain or restore operations to whom are assigned information security responsibilities
and ensure availability at the required level and in the re- are promoted systematically and structured, their effective-
quired time scale following interruption to, or failure of, crit- ness evaluated and possibly necessary corrective actions are
ical business processes. taken.
Introducing information security into holistic standard based organizations’ success, information security management
management systems requires a, the organization best should be enhanced by integrating it into all existing man-
adapted, system. This certainly can not be achieved by using agement systems.
a purchased general manual, which does not correspond with Implementing the holistic information security management
the lived processes and corporate objectives. Further the model promotes information security and thereby in accor-
management system must be constantly and effectively im- dance to organizations’ objectives availability, confidentiali-
plemented. ty and integrity, effectiveness, efficiency, legal and regulato-
This holistic information security model requires from the ry conformity, information security awareness and organiza-
system or information security manager, a part from standard tional development including information security improve-
management system skills, also information security skills, ment to guarantee stakeholder orientation and sustainable
the acquisition of necessary change management, organiza- organizations’ success.
tional learning, and controlling knowledge, the knowledge
about all relevant information security and data protection
laws, standards and regulations, as well as experience in in-
formation technology, knowledge management and informa- REFERENCES
tion management.
[1] Gordon, M. Loeb, W. Lucyshyn: CSI/FBI Computer Crime and Secu-
Sufficient IT-infrastructure and IT-support are also very im- rity Survey, i.cmpnet.com/gocsi/db_area/pdfs/fbi/FBI2006, 2006
portant for the project success. Only an appropriate IT sys- [2] <kes>-Microsoft® Sicherheitsstudie 2006, kes, vol. 4-6, Sonderdruck,
tem, technical equipment and infrastructure in accordance 2006
with organizational objectives promote continuously optimi-
[3] Federal Office for Information Security: Report about the IT security
zation of information security and sustainable organizations’ situation in Germany, bsi.de/english/publications/securitysituation/-
success. Lagebericht_2007_englisch.pdf
[4] ISO, ISO 9001:2008, Quality Managemnt Systems – requirements,
ISO, 1.12.2008.
[5] ISO, ISO/IEC 27001:2005, Information technology – Security tech-
VII. OUTLOOK niques – Information security management systems – Requirements,
5.10.2005.
Due to these excellent project experiences in several organi-
[6] ISO, ISO/IEC 27002:2005, Information technology – Security tech-
zations with different management systems there should be niques – Code of practice for information security management,
introduced enhanced information security management into 15.6.2005.
standard based management systems regarding all success [7] http://www.iso27001certificates.com/
factors [VI B].
[8] Y. Akao, Quality Function Deployment, integrating customer require-
Standard management system trainings should inform about ments into product design, Productivity Press, Portland, 1990.
the principles of information security management systems [9] M. Stoll Managementsysteme und Prozessorientiertes Wissensma-
and vice versa. nagement in Proc. 4th Conference on Professional Knowledge Man-
The management systems standards should emphasize the agement – Experiences and Visions, Gito Verlag, Berlin (2007), vo-
lume 1, pp. 433-434.
importance of information security and require the integra-
tion of information security into management systems. [10] T. Davenport: Thinking for a living, how to get better performance and
results from knowledge workers, Harvard Business School Press, Bos-
ton, 2005.
[11] P. Senge: Taking personal change seriously, The impact of Organiza-
tional Learning on management practice, Academy of Management
VIII. CONCLUSION Executive, vol. 17, p 47-50, May2003.
Integrating information security management into a process [12] M. Rosenberg: E-Learning, Strategies for Delivering Knowledge in the
Digital Age, McGraw-Hill, New York, 2001
oriented standard based management system enriches it.
Thereby information security can be introduced efficiently, [13] G. Reinmann-Rothmeier Mediendidaktik und Wissensmanagement. in
MedienPädagogik 10, www.medienpaed.com/02-2/reinmann1.pdf,
resources carefully, fast and successfully using many com- 2002.
mon synergies.
[14] M. Stoll, Workplace Process Integrated Learning and Knowledge Or-
Especially in organizations, where data, information and ganization, in H. Maurer and K. Tochtermann Eds. Proc. I-Know 07,
knowledge or availability of services, asset and/or infrastruc- 7th International Conference on Knowledge Management, J.UCS
ture or confidentiality or integrity has a high importance for Journal of Universal Computer Science, Graz, 2007.
Evaluation of Asynchronous Event Mechanisms for
Browser-based Real-time Communication Integration
Keith Griffin Colin Flanagan
Cisco University of Limerick
Galway, Ireland Limerick, Ireland
[email protected] [email protected]
Abstract — Desktop based real-time mark on a green icon is used to display availability while
communication applications are commonly an icon representing the hands of a clock is overlaid on a
used for presence based instant messaging blue icon to represent an inactive state.
and telephony applications. Such applications
use installed desktop components to handle
real-time asynchronous events on the client
originating from the communication system.
When the client is based in a web browser
these installed components are not available
however browser-based communication
applications are required to handle the same
type of asynchronous events. Moreover,
browser-based clients which typically run
over HTTP are challenged by HTTP itself
which is designed to be a synchronous
request-response protocol. We contend that a
suitable mechanism can be found to deliver
asynchronous real-time events to browser-
based applications
I. INTRODUCTION
In real-time communication integration systems such
as those providing telephony, instant messaging and
presence integration asynchronous events are a core and
mandatory feature. Basic events like a presence state
change from busy to available or lifting a telephone
handset resulting in an offhook event being delivered to an Fig. 1 Presence enabled buddy list in a desktop
application are fundamental to traditional desktop based communication client
communication integration applications. An example of
such an application displaying presence state is shown in A user’s presence state can be affected by multiple
Fig. 1 [1]. Presence is defined as a person’s availability sources as described in Fig. 2. These sources result in a
and propensity to communicate [2]. A person’s availability mix of user and system driven presence state changes.
can be reflected by a communication state. This state is
often presented as a colour, icon or colour coded icon. A • User driven presence: A user manually sets their
red phone icon for a contact may represent that the contact presence or availability state using a user interface
is busy and unavailable for voice communication, with such as a drop down menu. Presence states are
amber representing that they have been idle for some time often restricted to include “available”,
and green representing that they are available for voice “unavailable” or “busy” and “idle”.
communication. In this case the red, amber and green • System driven presence: A user’s state is
scheme which is commonly used in road traffic signals is automatically set by a system event e.g. logging
being use to represent availability and a phone icon is off from an instant messaging client would
being used to represent telephony or voice communication.
automatically set the user to “unavailable”.
In Fig. 1 the scheme that is used is to overlay a graphical
indication along with a colour scheme. A positive tick
461
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_80, © Springer Science+Business Media B.V. 2010
462 GRIFFIN AND FLANAGAN
Presence Source Sample Presence States first place. It has been found that several alternatives exist
User Busy, Available, Idle, In a to continuous polling which will be explored in the
meeting following sections.
System Logged in, Logged out, III. OVERVIEW OF ALTERNATIVE APPROACHES
Idle
Finding a solution for asynchronous events to browser-
Calendar Available, Busy based applications is not a trivial task. Browser-based
Online Meeting System / In a meeting, Meeting applications typically run over HTTP which despite having
Web Share state persistent connections is designed as a synchronous,
Telephony One the phone, call state stateless and connectionless protocol. Furthermore
Instant Messaging Message State communication related services such as presence generate
Fig. 2 Sample input sources for presence asynchronous events which require real-time handling.
Take the example of a user who has a presence state of
Regardless of how a presence state change occurs the “BUSY - ON THE PHONE” as a result of an integration
change must be reflected in applications subscribing to that between a telephony and presence service. This busy state
user’s presence in real-time. Some applications can may be displayed on a number of other clients who are
integrate directly with a protocol while others use subscribers or followers of this user’s presence state. The
middleware based integration [3]. While it is clear that number of clients depends on the number of others who
presence functionality is available to middleware and subscribe to the user’s presence and whether those users
installed desktop applications, it is proposed that suitable have a running client capable of displaying the presence
mechanisms can be found to offer presence functionality to state. When the user finishes the phone call the telephony
browser-based applications. We will use presence as a system will update the presence system to reflect a state of
sample real-time communication integration service to “AVAILABLE”. Once this state change happens, all other
discuss various mechanisms to extend real-time clients subscribed to that user’s presence must be updated
asynchronous presence events to a browser-based in real-time. Without this real-time update the value of the
application. system would be significantly diminished. In the case of an
installed desktop client which typically includes a local
II. ASYNCHRONOUS EVENTS FOR BROWSER-BASED protocol handler, receiving an asynchronous update could
APPLICATIONS OVERVIEW be described as straightforward although not trivial as it is
still a complex task. It is straightforward as the locally
Browser-based communication applications typically
installed desktop component can include a protocol
rely on HTTP as the protocol used to communicate with a
handler acting as a listener and ready to receive and handle
web server. HTTP 1.0 [4] improved on previously defined
incoming asynchronous events over a persistent
versions of the protocol which provided for raw data
connection. In the case of a browser-based client it is not
transfer over the internet. Improvements included allowing
so straightforward. The application running in the browser
messages to be of MIME like format containing meta
can make a request for a particular user’s presence from a
information about the data being transferred. Amongst
presence web service and it will get a response with the
other things HTTP 1.0 did not sufficiently take into
presence state. If that user’s presence state changes at an
consideration persistent connections. However persistent
unknown time in the future how will the browser-based
connections are considered in HTTP 1.1 [5] in order to
application become aware of this change? In the following
reduce the number of TCP connections and reduce
sections we will explore mechanisms to extend
congestion. However this does not address the ability for a
asynchronous real-time events to browser-based presence
client to receive real-time asynchronous events from a
applications.
server to a browser-based application. An asynchronous
event effectively amounts to a response without an explicit A. Continuous Polling and Streaming
request. The browser-based application will have One straightforward way to keep the client aware of
implicitly solicited an asynchronous event as part of an presence updates on the server is to not use an event
earlier request but the timing for an availability of that mechanism at all but instead to continuously ask what the
event will be unknown and driven by other events in the state is. Using this approach, rather than waiting for the
communication system. A mechanism is required which server to push events to the browser-based application the
will allow the delivery of asynchronous events which application implements a continuous cycle of HTTP
occur in the communication integration system to browser- requests for a given user or list of user’s presence. For
based clients. However any such mechanism must respect example request a presence state every second, receive and
the behavior of web-based systems and not allow immediate response, close the connection and repeat.
unsolicited events from the server to the client which could Various approaches to this technique have been suggested
create a security vulnerability. and implemented from the brute force technique suggested
One way to handle asynchronous events would be to above to more elegant standards submissions such as WS-
use a continuous polling mechanism but polling would Polling [6]. An obvious disadvantage of continuous polling
impact both the performance and scalability of the system is that it results in the server handling all of the polling
which are among the benefits that HTTP presents in the requests and as a result can be very inefficient on the
EVALUATION OF ASYNCHRONOUS EVENT MECHANISMS 463
server and also on the client which has to handle Bayeux suggests that all requests originate from the
continuous page refreshes. An improvement to continuous client i.e. a server may not initiate a connection to a client
polling would be to use a technique commonly referred to unless a client has previously sent a request to a server
as streaming [7]. This involves making a request from the which supports the browser security model. However
browser-based application to the presence web service and asynchronous events intended for a given client are
then not closing the connection at the server after the supported once a client has made a request to the server.
response has been sent. Subsequent asynchronous presence
events for the initially requested user would then be The transport for server to client event delivery can
streamed over this connection. Unlike continuous polling terminate the HTTP response after all messages are sent as
the server is not required to handle continuous requests in the case of a polling transport or use a streaming
however it is required to keep a connection continuously technique that allows multiple event messages to be sent in
open which means that this connection cannot be used for the same HTTP response. Two transport types are
other purposes. This approach takes the connectionless, supported by Bayeux: Long-Polling and Callback-Polling.
synchronous request-response nature of HTTP and turns it Callback-polling is a recommended but not mandatory
into a connection oriented approach like an installed transport type as defined by Bayeux. Callback polling uses
desktop client might use with a persistent TCP socket HTTP GET to pass the name of the callback to be
connection. For a web based application providing real- triggered to the server whereas long polling uses a POST
time asynchronous events this is not thought to be a request and response.
scalable solution, as it locks up too many connections. In Connections between the client and server are
particular, browsers often place a limit on the number of negotiated with handshake messages that allow the content
simultaneous connections they will hold open to a server. type, authentication, protocol version and other parameters
A technique that can be combined with polling for a to be agreed. Comet is a term often used to describe the
more efficient approach is Asynchronous Javascript And behavior of the Bayeux protocol. In fact the Bayeux
Xml (AJAX). Ajax consists of a number of technologies protocol definition was led by the same person that coined
that combine as a mechanism which can be used to poll for the term “Comet”. Cometd is an implementation of the
a specific update without updating the entire content of the Bayeux protocol by the Dojo Foundation [10].
page. Two key technologies used by Ajax are JavaScript Bi-directional Asynchronous Events with LongPolling
and XMLHttpRequest (XHR) [8]. XHR can be used to In order to support bi-directional communications as
request or poll a presence web service for updates. required for asynchronous presence updates, Bayeux
Following such an update the XHR object can use a call clients use two connections to a server to allow client to
back function to update the browser-based application server and server to client messaging occur simultaneously
using JavaScript. An advantage of Ajax is that it can be as shown in Fig. 3. Bayeux offers a transport option called
used to update a specific data set such a user’s presence long-polling which attempts to hold open a request until
state without requiring an entire page refresh. This offers a there are related events to deliver to the client. The intent
more efficient approach however it still relies on a brute is to always have a pending request to deliver events as
force polling technique to realize presence state changes they occur. The reconnect and interval advice fields are
on the server, as described in this example. used to maintain requests over time and address resource
B. Long Polling starvation and increased server load. In the worst case this
behavior can degenerate to that of traditional polling but
Long polling is a technique whereby a browser-based used efficiently it can minimize both latency in server to
client sends a request to a server where the server holds the client message delivery and the network resources required
request open until a response is available. This is similar to for the connection.
the streaming technique in that the request is held open but
different in that the connection can be closed either due to Bayeux Bayeux
a defined timeout or when a response is received. Once a Client Server
Req 0
connection is closed a new connection is initiated. As a
result a given client typically has a request open when an Req 1
asynchronous event such as a presence event becomes .
available. .
Wait
Resp 1 .
The Bayeux Protocol and Comet
Bayeux is a protocol used to transport messages Resp 0
between a web server and web browser using named
channels [9]. Messages can be delivered as follows:
Fig. 3 Bi-directional Asynchronous Operation
• Client to server
Bidirectional Streams Over Synchronous HTTP
• Server to client The XMPP Standards Foundation describes
• Client to client (via server) Bidirectional streams Over Synchronous HTTP (BOSH)
[11] as “a transport protocol that emulates the semantics
of a long-lived, bidirectional TCP connection between two
464 GRIFFIN AND FLANAGAN
entities (such as a client and a server) by efficiently using IV. ASYNCHRONOUS EVENTS FOR BROWSER-BASED
multiple synchronous HTTP request/response pairs COMMUNICATION INTEGRATION
without requiring the use of frequent polling or chunked
Two of the previously outlined mechanisms were used
responses”. As can be seen from this definition, BOSH is
to generate results for presence updates in a browser-based
a long polling technique and claims to be more efficient in
application, namely continuous polling and BOSH. Ajax
terms of bandwidth and responsiveness than Ajax. The
was used as a common technique to update only the
BOSH standard calls for the use of a HTTP Connection
presence data-set in the browser without requiring a full
Manager which is effectively a HTTP web server capable
page refresh.
of implementing the long polling technique as described in
the standard. BOSH works by having the HTTP A. Results for Continuous Polling
connection manager respond to a browser side request only Using Cisco’s Presence Web Service [18] a browser-
when a server side event is available. Once the client has based client was used to get and set presence states. In this
received this event it immediately makes another request case presence updates were polled using Ajax at the client.
to provide a vehicle for the next asynchronous event. The The resulting presence state change was observed in the
differences between the BOSH and Bayeux/Comet browser client as shown in Fig. 4. The web page shows a
techniques appear to be minimal and are related to client for user Colin Flanagan who is subscribed to
implementation differences when run over HTTP 1.0 and presence state changes for user Keith Griffin. The presence
1.1. state for Keith Griffin has just changed to Do Not Disturb
Interestingly, a standard has been defined for the use of (DND) at this point of the test.
the commonly used Extensible Messaging and Presence
Protocol (XMPP) [12] standard with BOSH. The XMPP
over BOSH standard [13] defines how XMPP stanzas can
be transported over HTTP. As defined in XMPP an XML
stanza is a discrete unit of structured information which
can be used to represent a user’s presence state. Clearly
this proven technique provides a mechanism for real-time
asynchronous presence events to be delivered to browser-
based applications for XMPP based presence and instant
messaging systems.
C. Web Sockets
Two key components of the HTML5 [14]
specification, as it relates to real-time asynchronous events
for browser-based applications, are the web sockets API
[15] and server sent events [16]. Web sockets attempt “to
enable Web applications to maintain bidirectional
communications with server-side processes”. The web
socket protocol [17] claims to provide an alternative to
HTTP polling for bi-directional communications between
a web page and server by using a single TCP connection to
manage traffic in both directions. It uses HTTP to achieve
a handshake between client and server by requesting a
HTTP “upgrade: websocket”. Following this the Fig. 4 Presence enabled client with continuous polling
underlying TCP connection is used for bi-directional
client-server communications. If the protocol delivers on While performance related timing information was
this claim it will offer a single socket connection capable available in the traces the goal of the test was not to
of bi-directional messaging between client and server. This measure performance but to determine if asynchronous
sounds like the definition that one might expect for an presence state changes could be successfully delivered to
installed desktop client running natively over TCP. the browser. Trace output in Fig. 5 shows a basic presence
update delivered to the user Colin Flanagan as a result of
Once a websocket is established server sent events can user Keith Griffin changing presence to a “Do Not
be used to push real-time asynchronous events to a Disturb” (DND) state.
browser-based application. This is achieved by using the
event source interface which allows for the creation of an "<presenceList type="basic">[\n]"
event source object in the client and registering an event httpclient.wire.content - <<
listener. "<contact contactURI="[email protected]"
presenceStatus="DND"/>[\n]"
"</presenceList>[\n]"
deliver asynchronous presence events to browser-based [4] Berners-Lee, T., R. Fielding, and H. Frystyk. HTTP – Hypertext
applications. While basic continuous polling techniques Transport Protocol (HTTP 1.0). RFC1945 1996 [cited; Available
can be used, they cannot scale. Web sockets offers the from: ftp://ftp.isi.edu/in-notes/rfc1945.txt.
promise of a “more native” solution in time, however [5] Fielding, R., et al. HTTP – Hypertext Transport Protocol (HTTP
currently it is suggested that long polling techniques
1.1). RFC2616 1999, ftp://ftp.isi.edu/in-notes/rfc2616.txt
(the Bayeux protocol or BOSH) are a suitable mechanism
[6] Brown, K., et al. Web Services Polling (WS-Polling). 2005,
for the delivery of asynchronous real-time events to
browser-based applications. http://www.w3.org/Submission/ws-polling/.
[7] Franklin, M. and S. Zdonik, Data in your face: push technology in
VII. FUTURE WORK perspective, in Proceedings of the 1998 ACM SIGMOD
While a suitable mechanism has been found for international conference on Management of data. 1998, ACM:
asynchronous real-time event delivery to browser-based Seattle, Washington, United States.
applications this mechanism has been identified as a work [8] W3C. XMLHttpRequest. W3C Working Draft 20 August 2009,
around technique rather than a solution. An obvious area http://www.w3.org/TR/XMLHttpRequest/.
for future research would be the pursuit of a true solution [9] Russell, A., et al. Bayeux Protocol Draft 1.0. 2007,
for real bi-directional messaging between web client and [10] Dojo. Cometd Bayeux Ajax Push. 2009,
server which could be used by real-time communication http://alex.dojotoolkit.org/2006/03/comet-low-latency-data-for-
systems.
the-browser/.
ACKNOWLEDGMENTS [11] Paterson, I., et al. XEP-0124: Bidirectional-streams Over
Synchronous HTTP (BOSH). 2005,
With thanks to David Mathieson and Seamus Kerrigan for
http://xmpp.org/extensions/xep-0124.html.
their tutorials on the implementation of JavaScript with
[12] Saint-Andre, P. XMPP Extension Protocols. 2001. XEP-0001.
Comet and XCP with BOSH respectively.
Available from: http://xmpp.org/extensions/xep-0001.html
REFERENCES [13] Paterson, I. and P. Saint-Andre. XEP-0206: XMPP Over BOSH.
[1] Jabber. The Power of Presence. 2008, 2007, http://xmpp.org/extensions/xep-0206.html.
http://www.jabber.com/CE/VideoVoiceCommunications. [14] Hickson, I. and D. Hyatt. A vocabulary and associated APIs for
[2] Rosenberg, J. A Presence Event Package for the Session Initiation HTML and XHTML. 2009, http://www.w3.org/TR/html5/.
Protocol (SIP). The Internet Society 2004, [15] Hickson, I. The Web Sockets API. 2009,
http://www.ietf.org/rfc/rfc3856.txt. http://dev.w3.org/html5/websockets/.
[3] Beltran, V. and J. Paradells, Middleware-based solution to offer [16] Hickson, I. Server-Sent Events. 2009,
mobile presence services, in Proceedings of the 1st international http://dev.w3.org/html5/eventsource/.
conference on MOBILe Wireless MiddleWARE, Operating [17] Hickson, I. The Web Socket protocol, 2009,
Systems, and Applications. 2007, ICST (Institute for Computer http://tools.ietf.org/html/draft-hixie-thewebsocketprotocol-44.
Sciences, Social-Informatics and Telecommunications [18] Cisco, Developer Guide for Cisco Unified Presence v7.0. 2008.
Engineering): Innsbruck, Austria. [19] Jabber. Jabber Extensible Communications Platform. 2008,
http://www.jabber.com/CE/JabberXCP.
An Ontology-based Intrusion Detection for RFID
Systems
M. Esposito1,2, G. Della Vecchia1
1
ICAR-CNR, Via Castellino 111, 80131 Naples, Italy
{massimo.esposito, gennaro.dellavecchia}@na.icar.cnr.it
2
University of Naples “Parthenope”, Centro Direzionale, Isola C4, 80143, Naples, Italy
[email protected]
467
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_81, © Springer Science+Business Media B.V. 2010
468 ESPOSITO AND VECCHIA
their cloned tags indistinguishable from the originals. Once Our approach closely resembles such an idea of applying
legitimate tag data has been obtained, attackers can reproduce the intrusion detection to identify tag cloning. But, differently,
their cloned tags on a wide scale. our research efforts have been primarily focused on
In high cost RFID tags, where resources are not very developing an intelligent misuse detection system based on
restricted, several countermeasures have been devised to ontologies and rules, that, applied in conjunction with an
combat tag cloning, such as deactivation of tags, encryption, inference engine, provide the advanced capacity of
authentication and hash codes [7]. In [8], Juels has automatically reasoning to detect tag cloning. To the best of
demonstrated some techniques for strengthening the resistance our knowledge, the RFID security literature has not yet
of EPC tags against cloning attacks, using PIN-based access to addressed applications of ontologies and rules to implement
achieve challenge response authentication. In [9], Weis et al. intrusion detection in RFID systems, neither system-oriented
have proposed a cryptographic approach to lock the tag researches appear to have been developed in that direction.
without storing the access key, but only a hash of the key on
III. THE ONTOLOGY-BASED INTRUSION DETECTION APPROACH
the tag. The key is stored in a back-end server and can be
found using the tag’s meta-ID. A. Misuse Detection
Avoine et al. [10] have proposed another hash-based RFID Traditionally, there are two basic approaches to intrusion
protocol that provides modified identifiers for improved detection: anomaly detection and misuse detection. In
privacy and that can be applied for authentication. In addition, anomaly detection, the goal is to define and characterize
hash-based RFID protocols for mutual authentication have legitimate behaviours of the users, and, then, detect anomalous
been proposed in [11-12]. All these protocols rely on behaviours by quantifying deviations from the former by
synchronized secrets residing on the tag and back-end server means of established metrics. Differently, misuse detection
and require a one-way hash function from the tag. involves the comparison of a user’s activities with known
In contrast, in low cost tags, due to their small size and behaviours of attackers, formalized in different models.
constrained resources, complex cryptographic solutions like The approach proposed in such a work relies on the
hash functions can not be executed. In [13] the authors application of a misuse detection based on ontologies and
mention scarcity of tag resources in low-cost RFID systems as rules to identify tag cloning in RFID systems. Such a choice is
a primary challenge in providing security and privacy due to a several of issues.
mechanisms, and, also, in combating cloning. To address this Firstly, a misuse detection based on ontologies and rules
issue, few lightweight authentication protocols that do not enables to have a clear understanding of what constitutes a
require cryptographic hash/keys in the tag have been proposed scenario of tag cloning, and, as a result, it is characterized by a
[14-15]. Another approach to combat cloning is the use of a low false positive rate. Secondly, it is more intuitive,
Physical Unclonable Function (PUF) [16]. PUFs significantly particularly in the case when users need to formalize and
increase physical security by generating volatile secrets that understand their knowledge about specific scenarios. This
only exist in a digital form when a chip is powered on and involves that a misuse detection system based on ontologies
running. Its main property is that it is easy to generate but hard and rules knows how tag cloning should manifest itself. This
to characterize. leads to a simple and efficient processing of the audit data.
All the cited solutions aim at preventing tag cloning. The obvious disadvantages of such a model rely on the time-
Additionally, tag cloning can be also detected by correlating consuming task of specifying ontologies and rules and the
information in the middleware layer. In such a sense, a very scarce ability to identify novel situations of tag cloning.
interesting approach has been proposed by Mirowski et al. [6]. Nevertheless, we believe that, in the context of our scenario,
More precisely, they have proposed an intrusion detection such issues do not constitute drawbacks and the proposed
system for RFID systems, named Deckard, based on a misuse detection can be proficiently applied, due to the very
statistical classifier and focused on the detection of change of limited amount of available data that, thus, significantly
tag ownership. This is one of the preliminary research to facilitates the task of specifying and keeping ontologies and
address the need for intrusion detection systems in RFID. rules up to date.
Another remarkable approach, similar to Deckard in its
intrusion detection architecture, has been proposed in [17]. B. The “track & trace” technique
Such an approach focuses beyond change in tag ownership The misuse detection has been actually realized by
and provides a more generic security framework to detect applying the “track & trace” technique. This technique relies
various RFID attacks. on the reasoning that “when you know where the genuine
The approaches outlined in [6, 17] are very remarkable for a tagged object is, the fake/cloned ones can be detected”.
several of ways. Firstly, they go beyond preventing tag In detail, “track & trace” refers to generate and store
cloning by actually determining whether the tag cloning has inherently dynamic profiles of RFID tagged object’s
occurred. Secondly, they avoid the difficulties onboard the movements through the supply chain. Such profiles are built
tags, by working with the middleware. As the reader and exploiting the set of location events related to one tagged
middleware components are typically accepted as the object that can be retrieved from a tracing system, such as the
expensive components of RFID, such proposals are very EPC network [18]. In addition to such dynamic profiles,
practical [6]. tagged objects are also characterized by static profiles, i.e.
fixed paths composed of point of interests (POI) through the
AN ONTOLOGY-BASED INTRUSION DETECTION FOR RFID SYSTEMS 469
Country
is As s oc iatedTo is As s oc iatedTo
RFID Tagged RFID Tagged Dynamic s ubClass Of is PartOf
Static Path Object Path
Object
has Dy namic POI Semantic
has Static POI City
Location
Static Point Dynamic Point maps s ubClass Of is PartOf
Of Interest Of Interest s ubClas s Of
Audit
Record
has Operation has Tag
hasReader
d) Audit Record
C. Ontologies and rules for the “track & trace” model The second part of the ontology models a dynamic profile.
The “track & trace” model specifies information regarding The main concepts and properties are outlined in Fig. 3 (b),
static and dynamic profiles, locations and audit records in a whereas a more detailed list of properties is reported in Tab. 2.
unique and sharable manner. Nevertheless, this information
Role Domain Range Inverse Trans.
must be represented in an unambiguous and well-defined hasDynamicPOI DynamicPath DynamicPoint isDynamic no
formalism and expressed in a machine-readable format in hasDynamicPathID DynamicPath
OfInterest
Datatype: String
POIOf
- -
order to facilitate the application of reasoning techniques to isAssociatedTo DynamicPath RFIDTaggedObject has no
DynamicPath
compare static and dynamic profiles of tagged objects. hasUsageModality DynamicPoint UsageModality isUsage no
As a result, we have defined an ontology to formally OfInterest ModalityOf
hasPhysicalLocation DynamicPoint PhysicalLocation isPhysical no
describe the “track & trace” model, in terms of relevant OfInterest LocationOf
concepts, properties and interrelations. More precisely, Fig.3 hasOperation
Counter
UsageModality Datatype: Integer - -
presents a high level view of the defined ontology, named hasTimestamp UsageModality Datatype: Time - -
hasTagID RFIDTaggedObject Datatype: String - -
“Track and Trace” Ontology, and formalized in standard OWL hasTagData RFIDTaggedObject Datatype: String - -
(Web Ontology Language) [19].This ontology is subdivided in hasValidID RFIDTaggedObject Datatype: Boolean - -
isCloned RFIDTaggedObject Datatype: Boolean - -
four logical parts for sake of clarity. isUnderAttack RFIDTaggedObject Datatype: Boolean - -
The first part of the ontology is intended to model a static hasPhysical
LocationID
PhysicalLocation Datatype: String - -
Moreover, according to such an ontology, we have formalized, usage condition of the corresponding static
in a set of SWRL (Semantic Web Rule Language) [20] rules, POI, in terms of number of operations
the declarative and procedural knowledge modelling the two allowable in a pre-defined time window.
levels of detection previously described. For sake of brevity
and clarity, we have grouped the implemented rules in logical An inference engine is utilized in conjunction with such
sets and reported their description only in natural language: formalized ontologies and rules to automatically reason about
• Analysis and processing of data contained in audit static and dynamic profiles of tagged objects and, thus, to
records. Depending on data contained in an audit implement the two reported levels of detection against tag
record, diverse sets of rules have been formalized for: cloning.
o identifying the dynamic profile to be updated
IV. THE ONTOLOGY-BASED IDS ARCHITECTURE
with new information;
o determining the physical location of the The proposed IDS has been devised to be transparently and
sensed tag for proximity to the reader that seamlessly integrated into a given RFID system by means of
has performed the operation; being positioned at the middleware level in order to effectively
o adding a new dynamic POI in the case when shield back-end systems from cloning attacks. Thus, even if a
the detected physical location belong to no cloned tag enters the reader field, its data will not propagate
POI. Such a POI will be also characterized beyond the IDS, the backend will never be aware of its
by the usage modality and the relative presence and normal operations will continue as if nothing
timestamp, both extracted by the audit happened. The architecture of the proposed IDS consists of the
record; set of components described below:
o updating an existing dynamic POI in the • Target System. It is a typical RFID system capable of
case when the detected physical location producing data that summarizes activity of tags in its
belong to an existing POI, but a new coverage area. It is responsible for recording the
operation has been performed. The existing outcome of a RFID read/write operation into an audit
POI will be characterized by the new usage record.
modality and the relative timestamp, both • Knowledge Base. It contains the formalization of the
extracted by the audit record; domain knowledge regarding the possible scenarios
o updating the usage modality of an existing of tag cloning. It consists of three entities: i) a
dynamic POI in the case when the detected Terminological Box, populated with the previously
physical location belong to an existing POI, described OWL ontologies, ii) the Rule Box,
and the operation has been already populated with the previously described SWRL rules,
performed in the past. The existing POI will and iii) the Assertional Box, populated with the audit
be updated with the timestamp and the records, appropriately formalized in RDF (Resource
occurrences of the last performed operation. Description Framework) [21].
• Execution of the first level of detection. A set of rules • Auditing Module. It is the component responsible for
have been formalized for detecting all the tagged collecting and logging the audit records received
objects that have an invalid tag serial number. from the target system into the Assertional Box of the
• Execution of the second level of detection. Depending Knowledge Base.
on data stored in both the static and dynamic profiles, • Decision Module. It is the smart component that
sets of rules have been formalized for: implements the “track & trace” technique by
o determining the correspondence existing applying ontologies and rules, as described in the
between semantic and physical locations; previous section. It relies on the logic inference
o inferring that a tagged object is cloned in the engine presented in [22], able to integrate and reason
case when it is detected simultaneously in at about ontologies and rules to perform inferential
least two different and not bordering reasoning on the domain knowledge stored into the
dynamic POIs in a too short time window; Knowledge Base.
o inferring that a tagged object is cloned in the
case when a dynamic POI is not expected in With such components in mind, the proposed IDS operates in
the static profile and it is not bordering with the following manner. Firstly, the Target System records the
static POIs; details of a new operation into an audit record. This record is,
o inferring that a tagged object is under attack then, stored by the Audit Module inside the Assertional Box of
in the case when a dynamic POI is not the Knowledge Base. Secondly, according to the data
expected in the static profile, but it is contained in the new audit record, the Decision Module
bordering with static POIs; updates the dynamic profile of the appropriate tag and
o inferring that a tagged object is under attack activates the inference engine in order to compare the
in the case when the usage modality of a appropriate static and dynamic profiles stored into the
dynamic POI is not in accordance with the
472 ESPOSITO AND VECCHIA
Knowledge Base, and, thus, infer if the sensed tag is authentic, [4] M. Lehtonen, D. Ostojic, A. Ilic, and F. Michahelles, “Securing RFID
Systems by Detecting Tag Cloning”, in Proceedings of the 7th
cloned or currently under attack. International Conference on Pervasive Computing, Lecture Notes In
An experimental prototype for testing purpose has been Computer Science, Springer-Verlag, 2009, vol. 5538,pp. 291-308. DOI
implemented at the Institute for High Performance Computing 10.1007/978-3-642-01516-8_20.
and Networking (ICAR) of Italian National Research Council [5] D.C. Ranasinghe, D.W. Engels, and P.H. Cole, “Low-Cost RFID
(CNR). All the described components have been implemented Systems: Confronting Security and Privacy”, in Auto-ID Labs Research
as fully portable Java entities, with the exception of the Audit Workshop, Zurich, Switzerland, 2005.
Module, which also integrate, as native code, Microsoft Visual [6] L. Mirowski, and J. Hartnett, “Deckard: A system to detect change of
C++ 6.0 control drivers provided by the RFID technology RFID tag ownership”, International Journal of Computer Science and
Network Security, 2007, vol. 7, pp. 89-98, ISSN 1738-7906 (2007).
manufacturers. Specifically, the Target System consists of
RFID readers belonging to FEIG’s Passive RFID Long Range [7] T. Karygiannis, B. Eydt, G. Barber, L.Bunn, and T. Phillips, “Guidelines
for securing radio frequency identification (RFID) systems,” April 2007,
Reader series, model ISC.LRU2000-A. These readers are able NIST Special Publication 800-98.
to interact with EPC Class 1 Gen 2 transponders for operating [8] A. Juels, “Strengthening EPC tags against cloning”, in Proceedings of
frequencies in the UHF range 865-870 MHz and reading range WiSe '05 (Cologne, Germany, September 2005), ACM, New York, NY,
of up to 10 meters (approx. 33 feet). FEIG’s UHF antennas 2005, pp. 67-76, DOI 10.1145/1080793.1080805.
model ID ISC.ANT.U250/250-EU have been installed, too. [9] S. Weis, S. Sarma, R. Rivest, and D. Engels, “Security and Privacy
These antennas are circular polarized for operating frequencies Aspects of Low-cost Radio Frequency Identification Systems”. in
in the UHF range 865-870 MHz. Proceedings of SPC 2003, Lecture Notes in Computer Science,
Springer-Verlag, 2004, vol. 2802, pp. 50-59, DOI 10.1007/b95124.
Preliminary tests performed on the experimental prototype
aimed at detecting synthesized cloning attacks gave a proof of [10] G. Avoine, and P. Oechslin, "A scalable and provably secure hash based
RFID protocol”, in Proceedings of PerSec 2005, IEEE Computer Society
the feasibility of our approach, suggesting that an actual Press, 2005, pp. 110–114, DOI 10.1109/PERCOMW.2005.12.
deployment of the proposed IDS can effectively support the
[11] E.Y. Choi, S.M. Lee, and D.H. Lee, “Efficient RFID authentication
detection of cloned tags in RFID systems. protocol for ubiquitous computing environment”, in Proceedings of
SecUbiq 2005, Lecture Notes in Computer Science, Springer-Verlag,
V. CONCLUSION
2005, vol. 3823, pp. 945-954, DOI 10.1007/11596042.
Detection may be seen as the first step in defending against [12] S. Lee, T. Asano, and K. Kim, “RFID Mutual Authentication Scheme
tag cloning and preventing RFID-enabled crime from based on Synchronized Secret Information”, in Proceedings of In
occurring. In this perspective, the present paper has described Symposium on Cryptography and Information Security, Hiroshima,
Japan, January 2006.
a misuse detection approach, based on the “track & trace”
technique, for RFID systems. Ontologies and rules have been [13] S. Sarma, S. Weis, and D. Engels, “Radio-frequency identification:
Security risks and challenges”, RSA Laboratories Cryptobytes, 2003,
used in order to implement such a technique, by enabling the vol. 6, no 1, pp. 2-9.
formalization of the declarative and procedural knowledge
[14] S. Karthikeyan, and M. Nesterenko, “RFID security without extensive
required to model two levels of defence against tag cloning. cryptography”, in Proceedings of SASN'05, ACM, New York, NY,
Moreover, an inference engine has been used to provide the 2005, pp. 63-67, DOI 10.1145/1102219.1102229.
proposed IDS with the advanced capacity of automatically [15] H.-Y. Chien, “SASI: A new ultralightweight RFID authentication
reasoning about ontologies and rules in order to actually apply protocol providing strong authentication and strong integrity”, IEEE
the formalized model of detection, and infer when a tagged Transactions on Dependable and Secure Computing, 2007, vol. 4, no 4,
pp. 337-340, DOI 10.1109/TDSC.2007.70226.
object is cloned or victim of a cloning attack.
As concluding remarks, it is relevant to observe that the [16] S. Devadas, E. Suh, S. Paral, R. Sowell, T. Ziola, and V. Khandelwal,
“Design and implementation of PUFbased “unclonable” RFID ICs for
approach described in this paper represents only a preliminary anti-counterfeiting and security applications”, in Proceedings of the
step of an ongoing research, because, at the moment, an 2008 IEEE international conference on RFID, IEEE Computer Society
experimental prototype has been tested only over proof Press, 2008, pp. 58–64, DOI 10.1109/RFID.2008.4519377.
cloning attacks. Next step of our research will be to minutely [17] G. Thamilarasu, R. Sridhar, "Intrusion detection in RFID systems", in
investigate the feasibility of the approach in real cases, in Proceedings of MILCOM 2008, IEEE Computer Society Press, 2008,
terms of performance evaluation and experimental assessment. pp. 16-19, DOI 10.1109/MILCOM.2008.4753218.
[18] EPCglobal Inc. “EPCglobal Architecture Framework Version 1.3”,
REFERENCES
http://www.epcglobalinc.org/standards/architecture/architecture_1_3-
[1] M. Ciampi, A. Coronato, G. De Pietro, M. Esposito, “A Location framework-20090319.pdf.
Service for Pervasive Grids”, in Proceedings of CISSE 2005, Book
[19] Web Ontology Language, available at www.w3.org/TR/owl-features/ .
“Advances in Systems, Computing Sciences and Software Engineering”,
Springer, 2006, pp. 119-123, DOI 10.1007/1-4020-5263-4_19. [20] Semantic Web Rule Language, available at
www.w3.org/Submission/SWRL/ .
[2] A. Coronato, M. Esposito, G. De Pietro, "A Multimodal Semantic
Location Service for Intelligent Environments: An Application for Smart [21] Resource Description Framework, available at www.w3.org/RDF/ .
Hospitals", Journal of Personal and Ubiquitous Computing (JPUC), [22] M.Esposito, “An Ontological and Non-monotonic Rule-based Approach to
Springer-Verlag, 2009, vol. 13, no 7, pp. 527-538, DOI 10.1007/s00779- Label Medical Images”, in Proceedings of SITIS 2007, IEEE Computer
009-0223-x. Society Press, 2007, pp. 603-611, DOI 10.1109/SITIS.2007.77.
[3] G. Della Vecchia, M. Esposito, "A Pervasive System for Nuclear
Medicine Departments", Journal of Wireless Personal Communications
(WPC), Springer, 2009, DOI 10.1007/s11277-009-9789-x.
A New Algorithm and its Implementations for RC Filters
GUOLIANG ZENG
Phone: (480) 727-1905; Fax: (480) 727-1773; E-mail: [email protected]
Motorola DSP Laboratory
Department Engineering
Arizona State University Polytech
7231 Sonoran Arroyo mall, Santan Hall, Mesa, AZ 85212
Here, 0 ¡ α ¡ 1 is called roll-off factor. The Ts is the symbol period. The spectrum of the
RC pulse is the Fourier transform of equation (1):
Ts , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , ω ≤ π (1 − α ) / Ts
°
°Ts {1 + cos[ 1 (T ω − π (1 − α ))]} =
°2 2α s
(2)
RC (ω ) = ®
°Ts cos 2 [ 41α (Ts ω − π (1 − α ))], , , , , , , , , , , , , π (1 − α ) < ω ≤ π (1 + α )
° Ts Ts
°
¯0, , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , ω > π (1 + α ) / Ts
Therefore the frequency bandwidth of an RC pulse is finite. It can pass some practical
channels without distortion.
473
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_82, © Springer Science+Business Media B.V. 2010
474 ZENG
x(t ) = ¦ k = −∞ d [k ]δ (t − kTs ) ,
∞ CURRENT ALGORITHM: Consider
the input sequence x[n] obtained by (6).
−∞<t <∞ (3)
From equation (3), every N elements of
x[n] , there is only one nonzero element,
Here, k = 0,±1,±2, is the index for
which is the information data d[k]. Other
symbols. Ts is the symbol period. And, N – 1 elements are zeros. In fact, the
d[k] is the information data sequence. It sequence of d[k] is the only physical
consists of 1s and 0s for binary unipolar input. In practical implementations, for
signaling, or 1s and –1s for binary example, in the function “rcosflt” of
bipolar signaling. The output is the MATLAB [2], x[n] is formed by
convolution of input and impulse “upsampling” which pads N – 1 zeros
response: for each d[k]. Since “rcosflt” uses
built-in functions, it is rewritten for
y (t ) = x(t ) ∗ h(t ) = ¦k = −∞ d [k ]h(t − kTs ) (4)
∞
comparison as follows:
0
h1,0, h1,1, h1,2,
-0.5 h1,3, h1,4, h1,5
-1
move #Xbase,r0
-1.5 ;point to states d[k].
0 5 10 15
Time
20 25 30 move #M-1,m0 ;mod(M).
A NEW ALGORITHM AND ITS IMPLEMENTATIONS FOR RC FILTERS 477
479
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_83, © Springer Science+Business Media B.V. 2010
480 OUWAIS ET AL.
including maximizing the number of users with IV. OPTIMIZATION TECHNIQUES FOR RADIO
equal rate requirements [4], maximizing the RESOURCES MANAGEMENT
minimum transmission rate for users with limited
power [5]. The objective is to find a subcarrier allocation
which allows each user to satisfy its rate
The main result of this paper is a class of requirements. Most famous problems can in fact be
algorithms which have been proposed for subcarrier described by: [4], [6].
assignment based on BABS, modified BABS and
our Hybrid BABS algorithm. These algorithms - RA (Rate Adaptive optimization
achieve comparable performance to the BABS problem),
algorithm but do not require intensive computation. - MA (Margin Adaptive optimization
A single cell with one base station and many problem),
mobile stations is considered. The algorithms - Equity optimization problem,
assume perfect information about the channel state - Outage optimization problem.
due to multipath fading as well as path loss and
shadowing effects, and the presence of a medium For more details, [2], [7], [9] could be named as
access protocol to convey information about several researches having been working about the
channel state and subcarrier allocation between the Rate Adaptive optimization problem (RA). We can
base station and the mobile stations. also refer to [3], [10], [11] for the Margin Adaptive
optimization problem (MA). Doufexi and Armour
III. SYSTEM MODEL AND PROBLEM [14] investigate the problem of dynamic multiuser
DESCRIPTION subcarrier allocation in a coded OFDMA system
based on the Equity problem optimization. Finally,
This section introduces the system model that is for Outage problem optimization we name to [4],
used, and formally introduces the problem being [6].
addressed in this paper. The criteria used for resource
allocation are introduced, and the subcarrier allocation A. Rate Adaptive problem formulation (RA)
method is formulated. We consider the downlink
transmissions of a multi-cell OFDMA system. Our objective is to optimize bandwidth allocation in
the system such that a maximum number of users
The system under consideration is an OFDM can achieve their data rate requirements. We aim to
system with frequency division multiple access find a bandwidth allocation that allows each user to
(FDMA). Perfect channel state information is satisfy its rate requirements while maximizing the
assumed at both the receiver and the transmitter, i.e. sum of user data rates subject to total power
the channel gain on each subcarrier due to path loss, constraint. Individual guaranteed rates, is
shadowing, and multipath fading is assumed to be considered [8], [10].
known. Channel parameters are assumed to be
perfectly estimated by a channel estimation The optimization problem can be expressed by:
algorithm for downlink OFDMA [3]. The design
and performance of this algorithm is outside the max ¦ ru with constraint
scope of these studies. The system does not employ ° u=1..U
spreading in either time or frequency, and each
® ¦ p ≤P for b=1...B
°n=1..N b ,n Maxb
subcarrier can only be used by one user at any ¯ b
given time. Subcarrier allocation is performed at the and
base station and the users are notified of the carriers
chosen for them. After the allocation, each user ru ≥ ru0
performs power allocation and bit loading across Where U is the total number of users, N is the
the subcarriers allocated to it to find the number of OFDMA subcarriers, Ptr,max the total
transmission power. power constraint, r and r0 are respectively the data
rate and data rate constraint (minimum data rate)
We consider a system with K users, and N for user u.
subcarriers. Each user k must transmit at least Rk_min
bits per unit time. Let Hk(n) be the channel gain, V. SENSIBLE GREEDY APPROACH
pk(n) the transmission power and rk(n) the
transmission rate for user k on subcarrier n. The problem posed by Wong.et.al [5] is
computationally intractable, and as described
A COMPUTATIONALLY DYNAMIC RESOURCE MANAGEMENT 481
w h ile (
#A * = m * ) do ° max ¦ ¦ ru, b
k k ® b =1 u =1
2 ° P( b ) < P
H k (n) ← 0 ¯ tr tr ,max (1 < u < U ),(1 < b < B)
2 and
k ← arg m ax H k ( n )
1≤ k ≤ K
en d w h ile
r ≥ r0
A
k
*
← A
k
*
∪ {n} VI. SIMULATION SETUP
en d fo r
In simulations Bandwidth Assignment Based on
SNR (BABS), Modified BABS (M-BABS), and
Where mk, is the number of subcarriers allocated to Hybrid BABS (H-BABS) are described. The
each user, and Ak ← { } initially for both channel gain g(u,n), of user u on subcarrier n in the
algorithms. serving cell, is summarized as:
-α
g (u ; n ) = K * d (u ) * ash (u ) * a f (u , n ) where
af has a Rayleigh distribution and represents the
small scale fading. We consider additive white
Gaussian noise (AWGN) which is characterized on
each subcarrier by a Gaussian random variable
with σ 2 = N 0 .W / N . The system under
consideration has parameters given in the following
Table. The channel model and traffic source models
used in simulations are described in more detail
below.
A COMPUTATIONALLY DYNAMIC RESOURCE MANAGEMENT 483
REFERENCES
[1] Didem. Kivanc and Hui. Liu, Subcarrier
Allocation and Power Control for OFDMG, IEEE
Transactions on Wireless Communications,
November 2003, vol. 2, no. 6, pp. 1150-1158.
Symposium on Personal, Indoor and Mobile Radio Allocation in Multiuser OFDM, In Proc. IEEE
Communications (PIMRC'06), 2006. Global Telecommunications Conf., volume 6,
pp. 3648 -3652,2001.
[8] Jiho. Jang, Kwang Bok. Lee, Transmit Power
Adaptation for Multiuser OFDMA Systems, IEEE [12] Lei. Yan, An introduction to Bandwidth
Journal on selected areas in communications, 21. Allocation in OFDM, March 2008.
NO 2, pp. 171- 178, February 2003.
[13] Carle. Lengoumbi, Philippe. Godlewski,
[9] Tommy K. Chee, Cheng-Chew. Lim, Jinho. Philippe. Martins, Comparison of different
Choi, Suboptimal Power Allocation for Downlink subchanalization modes for OFDMA, The 18th
OFDMA Systems, IEEE 60th VTC 2004, 26-29 Annual IEEE International Symposium on Personal,
September 2004. Indoor and Mobile Radio Communications
(PIMRC'07), 2007.
[10] D. Kivanc, H. Liu, Subcarrier Allocation and
Power Control for OFDMA, Conference Record of [14] A. Doufexi, S. Armour., Design considerations
the Thirty-Fourth Asilomar Conference on Signals, and Physical Layer Performance Results for a 4G
Systems, and Computers, pp. 147 -151, 2006. OFDMA System Employing Dynamic Subcarrier
Allocation, The 16th IEEE International
[11] Inhyoung. Kim, Hae Leem. Lee, Beomsup. Symposium on Personal, Indoor and Mobile Radio
Kim, and Yong H. Lee, On the Use of Linear Communications, PIMRC 2005.
Programming for Dynamic Sub channel and Bit
A General Method for Synthesis of Families
of Orthogonal Complementary Codes
B. Y. Bedzhev, Member, IEEE
University of Shumen “Bishop Konstantin Preslavsky”
115 Universitetska Str.
Shumen, Bulgaria 9712, EU
M. P. Iliev
University of Ruse “Angel Kanchev”
8 Studentska Str.
Ruse, Bulgaria 7017, EU
Abstract-Families of orthogonal complementary codes (FOCCs) With regard to above described situation, our paper aims to
are called families of sets of radio signals possessing both zero develop a general method for synthesis of families of orthogonal
autocorrelation functions (ACFs) (except for the zero time shift) and complementary codes (FOCCs), which will allow the most part
zero cross-correlation functions (CCFs) among all pairs of the of techniques for synthesis of FOCCs, known today, to be viewed
members of a family. They are viewed as promising tool for a
radical enhancement of the data transmission rate of the future
in a common theoretical frame. The importance of the studied
wireless communication systems, because they allow the users to problem results from the following facts. First, today the FOCCs
communicate practically without multiple access interferences in a are the only known radio signals that have perfect (both periodic
multipath environment. Due to this reason in the paper a general and aperiodic) correlation properties [1], [3], [4], [5], [8], [9],
method for synthesis of FOCCs is proposed. It allows the [10], [11], [12], [13], [14], [15]. This means that every member
parameters of the FOCCs to be independently managed and the of a FOCC has an ideal ACF, resembling the Dirac delta –
most part of techniques for synthesis of FOCCs, known today, to be function, and the CCFs among all pairs of members of a FOCC
viewed in a common theoretical frame. are zero for every time shift. Second, despite of taken efforts
Index Terms – Synthesis of signals, Families of orthogonal many problems in the field of the synthesis of the FOCCs are
complementary codes. open still.
Paper is organized as follows. First, the basics of the FOCCs
I. INTRODUCTION
are recalled. Second, our method for synthesis of FOCCs is
The wireless communication systems are in very rapid suggested. At the end some important conclusions are given.
progress today. Despite of this the users and industry expect an
II. BASICS OF THE SYNTHESIS OF THE FAMILIES OF
even more significant enhancement of the quality of the services
and the transmission rate, offered by these systems. In order to ORTHOGONAL COMPLEMENTARY CODES
meet these expectations, the experience, obtained during the past Let K ≥ 2, M ≥ 2, N ≥ 1, be integers and let us consider a
twenty years, has been analyzed by the theorist and developers. family V(K, M, N), comprising K sets as every set consists of M
The main conclusion is that the critical role for the whole sequences with (code) length N:
performance of the wireless communication systems has the
system of employed radio signals. Due to this reason great efforts
have been directed to finding families of radio signals possessing
{ak (m, n)}nN=−01 = (1)
both autocorrelation functions (ACFs) with a small level of the = {ak (m, 0), ak (m,1), ..., ak (m, N − 1)}, m = 0,1, ..., M − 1
side-lobes and small cross-correlation functions (CCFs) among
all pairs of members of a family. The focusing of the research Then V(K, M, N) is called family of orthogonal complementary
activities on this problem can be explained by the following facts codes (FOCCs) if the following conditions are satisfied [4], [7],
[1], [3], [4], [5], [8], [9], [10], [11], [12], [13], [14], [15]. First, [9]. First, the aggregated ACFs of the sequences of every set are
the pointed out families of radio signals allow the so-named self- zero except for the zero time shift. Second, the aggregated CCFs
interference (SI), caused by multipath spreading of of the sequences of all pairs of sets are zero for every time shift.
electromagnetic waves, to be reduced by a separate processing of These conditions can be mathematically described as follow:
the direct and reflected signals. Second, it is possible the negative
effect of simultaneous transmission of numerous users, named M .N , k = l ∩ r = 0;
multi user interference (MAI), to be minimized. R Ak , Al (r ) = ® (2)
¯ 0, k ≠ l ∪ r ≠ 0,
487
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_84, © Springer Science+Business Media B.V. 2010
488 BEDZHEV AND ILIEV
N −1− r M −1
where: R Ak , Al (r ) = ¦ ¦ ak (m, n).al* (m, n + r ) =
- r is the time shift, r = −( N − 1), ( N − 2), ..., − 1, 0,1, ..., N − 1 ; n=0 m=0
(6)
N −1− r
- Ak , k = 0,1, ..., K − 1 is the matrix, denoting the k-th set of = ¦ CAk (n) ⊗ CAl* (n + r ).
the family V(K, M, N), i.e.: n=0
Here CAk (n) is the n-th vector – column of the matrix Ak and
ª ak (0, 0) ak (0,1) ... ak (0, N − 1) º
« a (1, 0) the symbol “ ⊗ ” denotes the inner product of the n-th vector –
ak (1,1) ... ak (1, N − 1) »»
Ak = « k
(3) column of the matrix Ak and the complex conjugated (n+r)-th
« ... ... ... ... »
« » vector – column of the matrix Al . Now it should be recalled that
a
¬ k ( M − 1, 0 ) a k ( M − 1, 1) ... ak ( M − 1, N − 1)¼ the evaluation of the CCFs can be simplified by means of
multiplication of polynomials [2]. More specifically, let us
- R Ak , Al (r ) is the aggregated CCF (ACF if k = l) of the rows introduce the following polynomials
of the k-th and l-th sets (matrices), k , l = 0,1, ..., K − 1 .
N −1
The aggregated R Ak , Al (r ) CCF (ACF if k = l) is evaluated by Fk ( x) = ¦ CAk (n).x n , k = 0,1, ..., K − 1,
the well known formula [1], [2], [3], [4], [13], [14], [15]: n =0
, (7)
N −1
R Ak , Al (r ) =
Fl* ( x −1 ) = ¦ CAl* ( j ).x − j , l = 0,1, ..., K − 1
j =0
M −1 N −1− r
° ¦ ¦ ak* (m, n).al (m, n + r ), − ( N − 1) ≤ r ≤ 0, (4) then the coefficients of their polynomial product Fk ( x).Fl* ( x −1 )
°
= ® m = 0M −n1=N0 −1− r will be the aggregated cross-correlations R Ak , Al (r ) for all time
°
°¯ m¦ =0 n =0
¦ ak (m, n).al* (m, n + r ), 0 ≤ r ≤ N − 1. shifts r = −( N − 1), − ( N − 2), ..., − 1, 0,1, ..., N − 1 according to (6).
With regard to (7), the conditions (2) can be rewritten as
Here the symbol “*” means “complex conjugation”. follows
It should be mentioned that FOCCs are mathematical models
of the radio signals for the multicarrier code-division multiple M .N , k = l ;
access (MC-CDMA) wireless communication systems. In such a Fk ( x).Fl* ( x −1 ) = ® (8)
¯ 0, k ≠ l.
system K user can simultaneously communicate practically
without SI and MAI. The k-th set (i.e. the matrix
It should be stressed that multiplication of the coefficients in
Ak , k = 0,1, ..., K − 1 ) of the exploited FOCC is assigned to the k- (8) is performed as multiplication of vectors, i.e.
th user as its signature code. The entry ak (m, n) ,
m = 0,1, ..., M − 1 , n = 0,1, ..., N − 1 of the matrix M −1
CAk (n) ⊗ CAl* (n + r ) = ¦ ak (m, n).al* (m, n + r ) (9)
Ak , k = 0,1, ..., K − 1 is the complex envelope of the n-th m=0
elementary pulse of the m-th sequence. In fact the m-th sequence
describes mathematically the law of the amplitude and phase and due to this reason it is possible CAk (n) ⊗ CAl* (n + r ) = 0 .
modulation of the m-th subcarrier frequency during the every
clock interval in the process of the spreading of the spectrum of III. A GENERAL METHOD FOR SYNTHESIS OF FAMILIES
the radio signals. Most often only phase manipulation (Q-phase OF ORTHOGONAL COMPLEMENTARY CODES
shift keying) is used, i.e. The complementary codes were invented by M. Golay in 1949.
In his origin works [6], [7] K = 1, M = 2 and
∀ak (m, n) ∈ {exp[(2πil ) / Q; l = 0, 1, ..., Q − 1}, i = − 1 , (5) ∀a (m, n) ∈ {− 1, + 1}, m = 0,1, n = 0,1, ..., N − 1 , which gave the
reason these codes to be named complementary pairs. Later, in
but this restriction will not be used in the sequel. 1972, Tseng and Liu [13] extended the Golay’s conception and
For brevity but without loss of generality we shall suppose that introduced the so - named mutually orthogonal complementary
0 ≤ r ≤ N − 1 in (4). Then, after changing the order of sets, consisting of M ≥ 2 complementary sequences.
summations, (4) can be transformed as follows: Due to their unique correlation properties the complementary
codes have been intensively studied. For instance the study in [4]
A GENERAL METHOD FOR SYNTHESIS OF FAMILIES 489
is based on over 300 journal papers and conference reports, are a FOCC with parameters: family size – K, number of
concerning this theme. The complementary codes have also sequences in every set – M’= K.M and length of the sequences –
exploited in some types radars and specialized communication N.
systems and obtained experience has proved the practical Proof: According to (7) let F (x) and F * ( x −1 ) be the
effectiveness of these signals. As a result at the beginning of the polynomials of (N-1)-th degree which coefficients are the vector
new century it was suggested the FOCCs, described in the – columns of the matrix A. As A is a set of M complementary
previous section, to be the base tool for multiple access to the sequences with length N:
communication channel in the next generations wireless
communication systems [1], [3], [4], [5], [8], [9], [12], [14], [15]. F ( x).F * ( x −1 ) = M .N . (13)
At the moment despite of the taken efforts in the study of
complementary codes, it is not known if or not the parameters K,
M, N of a FOCC can be independently managed during the Let CE (n) be the n-th column – submatrix of the matrix A:
developing of a new wireless communication system. For
example in [3], [4], [9], [13], [14] methods for synthesis of CE (n) = [d (0, n). A d (1, n). A ... d ( K − 1, n). A]T , (14)
FOCCs are described, according to which an augmentation of the
number of users (i.e. K) is possible only by enlarging of the where the symbol “T” denotes “transposition”.
length N of the sequences. Unfortunately, this approach leads to After taking into account (11) and (13), the CCF of the n-th
decreasing of the data transmission rate, which is very and l-th column - submatrices CE (n) and CE ( j ) of the tensor
undesirable. Due to this reason in this part of the paper a general
method for synthesis of FOCCs is suggested. In allows the product (12) can be evaluated as follows
parameters K, M, N of a FOCC to be chosen independently
during the process of system optimization. Besides our method ª d (0, n).F ( x) º ª d * (0, l ).F * ( x −1 ) º
provides the opportunity a significant part of known today « d (1, n).F ( x) » « −1 »
techniques for synthesis of FOCCs to be viewed from a common « » ⊗ « d * (1, l ).F * ( x ) » =
theoretical base. « ... » « ... »
« » « −1
»
The suggested general method for synthesis of GOCCs is ¬ d ( K − 1, n ).F ( x ) ¼ ¬ « d * ( K − 1, l ).F * ( x ) »¼
based on the following propositions.
¦ [d (k , n).F ( x)].[d * (k , l ).F * ( x −1 )] =
K −1
Proposition 3.1: Let A be a set of M complementary sequences = (15)
with length N and let D be a matrix k =0
K −1 (K .M ).N , n = l
ª d (0, 0) d (0,1) ... d (0, K − 1) º = M .N ¦ d (k , n).d * (k , l ) = ®
« d (1, 0) k =0 ¯ 0, n ≠ l.
d (1,1) ... d (1, K − 1) »»
D=« (10)
« ... ... ... ... » Equation (15) proves the Proposition 3.1.
« » It should be emphasized on the following facts.
¬ d ( K − 1, 0) d ( K − 1,1) ... d ( K − 1, K − 1)¼
First, the rows of D are mutually orthogonal also, because D is
a square matrix [1].
which columns are mutually orthogonal, i.e. Second, the Proposition 3.1 remains true if we reduce some of
columns of D and use only K’ (K’<K) columns for creating the
K −1 K , n = l; tensor product (12). Anyway in this case the size of the
¦ d (k , n).d * (k , l ) = ® 0, n ≠ l , (11)
synthesized FOCC will be smaller – namely, it will be K’.
k =0 ¯
Third, on the base of a known initial set of M complementary
for 0 ≤ n, l ≤ K − 1 . Then the column - submatrices of the tensor sequences with length N the Proposition 3.1 proves a technique
product for synthesis of a FOCC with parameters: family size – K,
number of sequences in every set – M’= K.M and length of the
sequences – N.
ª d (0, 0). A d (0,1). A ... d (0, K − 1). A º Proposition 3.2: Let V(K, M, N) be a FOCC consisting of K
« d (1, 0). A d (1,1). A ... d (1, K − 1). A »» sets A0 , A1 , ..., AK −1 as every set comprises M complementary
E=« (12)
« ... ... ... ... » sequences with length N. Besides, let D be a matrix, which
« » columns are orthogonal, according to (10) and (11). Then the
¬d ( K − 1, 0). A d ( K − 1,1). A ... d ( K − 1, K − 1). A¼
column - submatrices of the all tensor products
490 BEDZHEV AND ILIEV
are a FOCC with parameters: family size – K’ = K2, number of (K .M ).N , k = l ∩ n = j
=®
sequences in every set – M’= K.M and length of the sequences – ¯ 0, k ≠ l ∪ n ≠ j.
N.
Here the all rows of the matrix U
Equation (20) proves the Proposition 3.2.
It should be pointed out that the Proposition 3.2 suggests a
ª u (0, 0) u (0,1) ... u (0, K − 1) º
« u (1, 0) technique for expanding K times the size of an initial FOCC
« u (1,1) ... u (1, K − 1) »» without changing the length of the sequences. This is a very
U= (17)
« ... ... ... ... » useful opportunity, because the known today methods [3], [4],
« » [9], [13], [14] expand the size of an initial FOCC only by a
¬u ( K − 1, 0) u ( K − 1,1) ... u ( K − 1, K − 1)¼
simultaneous enlargement of the length of the sequences.
Anyway, in order to show that parameters of a FOCC can be
are permutations of the set {0,1, ..., K − 1} , i.e. independently managed it is necessary to prove the following
{u (k , 0), u (k ,1), ..., u (k , K − 1)} are the numbers {0,1, ..., K − 1} in propositions.
Proposition 3.3: Let V(K, M, N) be a FOCC, consisting of K
a possibly changed order.
sets A0 , A1 , ..., AK −1 as every set comprises M complementary
Proof: According to (7) let Fk , m ( x) and Fl*,m ( x −1 ) be the
sequences with length N. Besides, let D be a matrix which
polynomials of (N-1)-th degree which coefficients are the vector columns are orthogonal, according to (10) and (11). Then the row
– columns of the matrices Au ( k , m) and Au*(l , m) respectively. As - submatrices of the tensor product
all matrices Au ( k , m) are sets of M complementary sequences with
H=
length N:
ª d (0, 0). A0 d (0,1). A1 ... d (0, K − 1). AK −1 º
« d (1, 0). A d (1,1). A1 ... d (1, K − 1). AK −1 »» (21)
M .N , k = l ,
Fk , m ( x).Fl*,m ( x −1 ) = ® (18) =« 0
« ... ... ... ... »
¯ 0, k ≠ l , « »
¬d ( K − 1, 0). A0 d ( K − 1,1). A1 ... d ( K − 1, K − 1). AK −1 ¼
for 0 ≤ k , l ≤ K − 1 .
Let CEk (n) be the n-th column – submatrix of the matrix Ek, are a FOCC with parameters: family size – K, number of
i.e.: sequences in a set –M and length of the sequences – N’ = K.N.
Proof: According to (7) let Fk (x) and Fk* ( x −1 ) be the
CEk (n) = polynomials of (N-1)-th degree which coefficients are the vector
,(19) – columns of the matrix Ak. As A0 , A1 , ..., AK −1 are a FOCC with
[
= d (0, n). Au ( k ,0 ) ... ... d ( K − 1, n). Au ( k , K −1) ] T
parameters K, M, N:
CH k (m) = [d (m, 0). A0 d (m,1). A1 ... d (m, K − 1). AK −1 ] .(23) with parameters K’ = 2, M’ = K.M = 2 and N’ = 1.
Now, using the Proposition 3.3, FOCC from (27) and D,
After taking into account (22) and (23), the CCF of the m-th defined by (26), we create the following FOCC
and l-th row - submatrices CH k (m) and CH l (m) of the tensor
ª+ 1 + 1º ª+ 1 − 1º
product (21) can be evaluated as follows A' '0 = « », A' '1 = «+ 1 + 1» (28)
¬ + 1 − 1¼ ¬ ¼
ª K −1 k .N º
ª K −1 º with parameters K’’ = 2, M’’ = 2 and N’’ = K.N’ = 2.
« ¦ d (m, k ).Fk ( x ).x » « ¦ d * (l , j ).F j ( x ) x
* −1 − j . N
»= At the end, employing the simplest matrix U, which rows are
¬k =0 ¼ ¬« j = 0 ¼»
the numbers {0,1, ..., K − 1} in their natural order (here K = 2)
K −1
= ¦ d (m, k ).d * (l , k ).Fk ( x).Fk* ( x −1 ) +
k =0 ª0 1º
K −1K −1 (24) U =« » (29)
+ ¦ ¦ d (m, k ).d * (l , j ).Fk ( x).F j* ( x −1 ).x ( k − j ).N = ¬0 1¼
k = 0 j = 0,
j≠k and applying the Proposition 3.2, FOCC from (28), again D,
K −1 M (K .N ), m = l defined by (26), we create the following FOCC
= M .N ¦ d (m, k ).d * (l , k ) + 0 = ®
k =0 ¯ 0, n ≠ l.
ª+ 1 + 1º ª+ 1 + 1º
Equation (24) proves the Proposition 3.3. «+ 1 − 1»» «+ 1 − 1»»
Proposition 3.4: After interleaving of the columns of the row – A' ' '0 = « , A' ' '1 = « ,
«+ 1 + 1» «− 1 − 1»
submatrices CH k (m) , defined by (23), a FOCC with parameters: «+ 1
¬ − 1»¼ «− 1
¬ + 1»¼
family size – K, number of sequences in a set –M and length of (30)
the sequences – N’ = K.N will be obtained. ª+ 1 − 1º ª+ 1 − 1º
The Proposition 3.4 can be proven in a truly similar way as «+ 1 + 1»» «+ 1 + 1»»
above and due to this reason the proof is left to the reader. A' ' '2 = « , A' ' '3 = «
«+ 1 − 1» «− 1 + 1»
IV. RESULTS «+ 1 + 1»¼ «− 1 − 1»¼
¬ ¬
It should be pointed out that the Proposition 3.3 and 3.4
suggest techniques for expanding K times the length of the The parameters of the FOCC, presented by its sets of
sequences of an initial FOCC without changing the number of the complementary sequences (30), are: K’’’ = K.K’’ = 4,
sequences. Consequently, Propositions 3.2, 3.3 and 3.4 together M’’’ = K.M’’ = 4 and N’’’ = N’’ = 2.
prove that the parameters of the FOCCs can be independently It should be stressed that the analogous FOCC, synthesized by
managed. the methods, known from [3], [4], [13], [14] is
The Propositions 3.1, 3.2 and 3.3 will be explained by the
following example.
ª+ 1 + 1 + 1 + 1º ª− 1 − 1 + 1 + 1º
Example 4.1: Let us make use of the simplest set of «+ 1
complementary sequences + 1 − 1 − 1»» «− 1 − 1 − 1 − 1»»
B0 = « , B1 = « ,
«− 1 + 1 − 1 + 1» «+ 1 − 1 − 1 + 1»
A = [+ 1] (25) «− 1
¬ + 1 + 1 − 1»¼ «+ 1
¬ − 1 + 1 + 1¼»
(31)
ª− 1 + 1 − 1 + 1º ª+ 1 − 1 − 1 + 1º
with parameters K = M = N = 1 and the simplest orthogonal «− 1
matrix + 1 + 1 − 1»» «+ 1 − 1 + 1 − 1»»
B2 = « , B3 = «
«+ 1 + 1 + 1 + 1» «− 1 − 1 + 1 + 1»
«+ 1 + 1 − 1 − 1»¼ «− 1 − 1 − 1 − 1»¼
ª+ 1 + 1º ¬ ¬
D=« ». (26)
¬+ 1 − 1¼
As seen from (30) and (31), the classical methods for synthesis
After applying the Proposition 3.1 we obtain the FOCC of FOCCs extend the size of a FOCC only by the cost of a
directly proportional enlargement of the length of the sequences,
ª+ 1º ª+ 1º used for spreading the spectrum of the radio signals. This
A'0 = « », A'1 = « » (27) peculiarity often has negative effect over the performance of the
¬+ 1¼ ¬− 1¼ developed wireless communication system as it leads to an
492 BEDZHEV AND ILIEV
493
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_85, © Springer Science+Business Media B.V. 2010
494 OYEYINKA ET AL.
requires explicit network assistance to determine a practicable The remaining of this paper is organized as follows:
starting sending rate called Quick Start [4]. section II presents the existing TCP slow start mechanism.
According to [4], the experimental Quick start TCP Section III introduces our proposed approach called e-speed
extension is currently the only specified TCP extension that start, the design of e-speed start and its mechanism. Section IV
realizes a fast startup. A large amount of work has already discusses initial results obtained. Section V concludes the
been done to address the issue of choosing the initial paper.
congestion window for TCP. RFC 3390 [1] allows an initial
window larger than one packet. Quick start is based on the fact II THE EXISTING MODEL
that explicit feedback from all routers along the path is Window based congestion control has four stages; this is
required to be able to use an initial window larger than those to say that the life time of a connection passes through
specified by RFC 3390. possibly four stages before connection termination. These are
In quick start proposals, a sender (TCP host) would the slow start, congestion avoidance, fast retransmit and fast
indicate its intention to send at a particular rate in bytes per recovery. Most research efforts in the past have been
second. Each router along the path could approve, reduce, concentrated on the congestion avoidance stage and several
disapprove or ignore the request. Approval of the request by advances have been made until recently when research efforts
the router indicates that it is being underutilized currently and began to focus on the slow start stage due to advances in
it can accommodate the sender’s requested sending rate. The network speed technologies which made it possible to have
quick start mechanism can detect if there are routers in the gigabit networks operating at a speed of 1 gigabit/second (gbs)
path that disapproved or do not understand the quick start and above, compared to the maximum speed of 100 megabits
request. The receiver communicates the response to the sender per second (mbps) that existing networks transmit. The
in an answering TCP packet. If the quick start request is implication of this higher speed of transmission is that the
approved by all routers along the path, then the sender begins current congestion control techniques being used on the
transmission at the approved rate. Subsequently, transmissions internet may not be suitable for the best effort network in the
are governed by the default TCP congestion control future. The slow start stage is modeled as described below.
mechanism. If the request is not approved, then the sender The algorithm below applies:
transmits at the normal TCP startup speed [15].
According to [15], TCP is effective for both flow control The Slow Start Algorithm
and congestion control. TCP flow control is a receiver-driven 1. Start with a window size of 1 i.e cwnd=1
mechanism that informs the sender about the available 2. Increase the window size by 1 i.e. cwnd=cwnd+1; for
receive-buffer space and limits the maximum amount of every ACK1 received. Repeat step 2 until the ssthresh is
outstanding data. Flow control and congestion control are reached OR Packet loss is detected
independent while size of receive buffer space depends on the 4. If ssthresh is reached, go to congestion avoidance phase.
capacity of the receiver network. However, if the TCP is used 5. Else If a packet loss is detected;
with a link with large bandwidth-delay product, both set to ssthresh2 = 0.5*cwnd3
congestion window and flow control window need to be large set to cwnd =1
in other for TCP to perform well [4]. This makes both flow go to step 2
control and congestion control to overlap. The quick start TCP 6. Endif
extension assumes independence of flow and congestion Note: Set the initial value of ssthresh to a fraction of the
control which is not true and makes it not appropriate for the maximum window size. This is determined at the beginning of
global internet. transmission. Table I shows the simulation of the slow start
Another issue with the quick start approach is that the algorithm stated above. From table I the window size is
starting speed of a connection using quick start may send expressed as
thousands of packets in one RTT. The implication of this is W(nT + m/c) = 2n – 1 + m + 1, 0 ¡ m ¡ 2n –1 (1)
that, thousands of packets may be dropped in one RTT, And the maximum window size in a cycle is derived as
thereby necessitating a retransmission of these packets which W(m) = 2n (2)
may in turn be dropped and retransmitted over and over again, There will be two slow start phase. The first phase, buffer
thereby saturating the network which can lead to congestion. overflows and there is a packet loss which reduces the window
In addition, most routers and other network components on the size to 1 and slow start is re-entered. Equations 3 and 4
global internet are not built with capabilities to be quick start describe the ssthresh in both phases respectively.
aware. The implication is that they need either to be replaced ssthresh = (cT + B)/2 (3)
or modified before deployment. Hence, we propose a milder ssthresh = Min (2B-1, (cT + B)/4) (4)
approach to fast startup which called e-speed start. This is a where B is buffer size of link, c is the link capacity, T is sum
form of fast startup that does not require routers’ response or of propagation and queuing delay, W(m) is maximum window
modification. Rather, it employs end-to-end principles to size in a cycle, n is the number of cycles and m is number of
determine a suitable startup speed. This proposal is adaptable
to the current internet as well as the future gigabit internet. 1
Acknowledgment
2
Slow start threshold
3
Congestion window
E-SPEED START- A WINDOW BASED ADDITIVE INCREASE, MULTIPLICATIVE DECREASE NETWORK 495
C. Speed Determination Stage The system transits into bandwidth estimation mode if
At this stage, two main environmental variables are the value of Time-of-day variable is off-peak or if user-input
introduced and their values determined in the following order: is yes in response to the holiday prompt. The bandwidth
• Time-of-day: a binary value of peak or off-peak estimation module estimates the available unused bandwidth.
• User-input: a binary value of workday or holiday. If bandwidth usage is low and the estimated available
In standard TCP, after handshake, the connection transits bandwidth is high, the system enters e-speed start otherwise
from idle to slow start state where data transmission proceeds slow start is activated.
immediately. In e-speed start, successful handshake transits From figure 3 below, the time function captures the time
into the speed determination state. At the speed determination of connection, determines if it can be categorized as off-peak
state, the system time is captured into the variable Time-of- or peak period, and passes control to bandwidth estimation
day and the value is compared with a table that contains data function or user input as the case may be. The bandwidth
collected from network usage statistics. If bandwidth usage is estimation function uses the packet pair algorithm to
high at this time (as contained in the network usage statistics determine the available bandwidth and thereby passes control
table), the value of Time-of-day is set to peak and if to either the slow start or e-speed start. If control is passed to
bandwidth usage is low, it is set to off-peak. It is assumed that the slow start, it proceeds as in conventional TCP. If the user-
if network usage is low for a period of time at a particular time input module is activated, the user is prompted with an
of the day, it is likely to be low always at such period of time optional message e.g. “press spacebar if it is an holiday”. If
and vice versa. However, the determination of the period of user responds to this message the bandwidth estimation
the day to be categorized as low or high bandwidth usage time module is activated else the slow start state is entered into and
is not trivial. We intend to use fuzzy logic to determine the data transmission starts. If the bandwidth estimation function
probability of low or high bandwidth usage time. If the value returns a high value for available bandwidth, e-speed start can
of Time-of-day is peak the connection activates user input be used for startup for data transmission.
module and if it is off-peak it proceeds to bandwidth
estimation. The e-speed Start Algorithm
The second environmental variable is user-input, the user 0. Estimate the value of using available bandwidth. If =1
optionally responds to the prompt, “is it an holiday?” for go to 1else if >1 go to 7
example, during connection attempt. If yes, the speed 1. Start with a window size of 1 i.e cwnd=1
determination module default to bandwidth estimation to 2. Increase the window size by 1 i.e. cwnd=cwnd+1; for
determine available bandwidth, otherwise it proceeds to slow every ACK received. Repeat step 2 until the ssthresh is
start. Time-of-day and User-input are two variables that reached OR Packet loss is detected
determine if a connection proceeds to bandwidth estimation 4. If ssthresh is reached, go to congestion avoidance phase.
state or directly to slow start state. 5. Else If a packet loss is detected;
set to ssthresh = 0.5*cwnd
Congestion set to cwnd =1
Avoidance
go to step 2
Fast 6. Endif
Recovery
7. If =n, Increase the window size by 2 i.e.
Idle Slow Start
cwnd=cwnd+2; for every ACK received. Repeat step 7
until ssthresh is reached OR Packet loss is detected
ack of 8. If ssthresh is reached, go to congestion avoidance phase.
Exponential 9. Else if a packet loss is detected;
b
set to ssthresh = 0.5*cwnd
set to cwnd =1
go to step 2
Fig 1: TCP state transition using Slow Start 10. Endif
Using decision table to describe these scenarios, we
, Congestion
Avoidance obtain table II below: From table II, it is observed that E-
Speed
speed is used for startup at 37.5% of the time while slow start
start Fast
Retransmit
still dominate most of the connections at 62.5% of the time.
This implies that the link/network would have been
Idle Environmen
tal factor
underutilized 37.5% of the time. It is clear that e-speed start
optimizes the available bandwidth utilisation at these times for
faster connection and data transfer.
Slow Start
Fast Recovery
IV DISCUSSION OF E-SPEED START MECHANISM.
In e-speed start mechanism, new connection starts by
TCP sending an initial packet of 1. Thereafter, the congestion
Fig 2: TCP state transition using e-speed start
window is grown by (5);
E-SPEED START- A WINDOW BASED ADDITIVE INCREASE, MULTIPLICATIVE DECREASE NETWORK 497
TABLE III
E-SPEED START TRANSITION CYCLES WITH =2 and 3
70
Slow start with B = 1
E-Speed start with B = 2
E-Speed start with B = 3
60
Congestion Congestion Congestion
Window for Window for E- Window for E- 50
Slow-Start speed Start speed Start
NUMBER OF PACKETS
Cycle (=1) (=2) (=3)
40
0 1 1 1
1 2 4 8 30
2 4 16 64
3 8 64 512 20
4 16 256 4,096
10
5 32 1,024 32,768
6 64 4,096 262,144 0
7 128 16,384 2,097,152 0 0.2 0.4 0.6 0.8 1
CYCLE
1.2 1.4 1.6 1.8 2
499
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_86, © Springer Science+Business Media B.V. 2010
500 GRECA
¦x
d =1
dl ≤1 ∀l ∈ E , λ ∈ Λ
3. each demand uses the same wavelength through its path
E1
¦x
l =1
dl ≤1 ∀d ∈ D, λ ∈ Λ
4. There are no loops
E1 D E1 D
¦¦ x
l =1 d =1
dl in + ¦¦ xdl out ≤ 2
l =1 d =1
∀n ∈ V , λ ∈ Λ
5. Conservation law
1 if n is the origin
E1 D E1 D
°
¦¦ x dlin − ¦¦ xdlout = ®− 1 if n is the destination
l =1 d =1 l =1 d =1 ° 0 otherwise
¯
6. All variables can be either 1 or 0
x dl = 1 or 0 ∀d ∈ D, λ ∈ Λ
REFERENCES
[1] P. Demeester, M. Gryseels, K. van Doorselaere, A. Autenrieth, C.
Brianza, G. Signorelli, et al. “Network resilience strategies in
SDH/WDM multilayer networks”, 24th European Conference on Optical
communication, Vol. 1, pp. 579-580, September 1998.
[2] C. Ou, K.Y. Zhu, B. Mukherjee, “Survivable traffic grooming in WDM
mesh networks”, IEEE OFC, Vol. 2, pp. 624-625, March 2003.
[3] J. Zhang, et al IEEE OFC, Vol. 2, pp. 622- 623, March 2003.
[4] A. Greca, et al IEEE Workshop on HPRS, pp. 146-149, Dallas 2001.
[5] S. Ramamurthy, B. Mukherjee, “Survivable WDM mesh networks, part I
– protection”, Proc. IEEE Infocom 99.
[6] S. Ramamurthy, B. Mukherjee, “Survivable WDM mesh networks, part I
– restoration”, Proc. ICC 99.
[7] O. Gerstel, R. Ramaswami, “Optical layer survivability – an
implementation perspective”, IEEE Journal of Selected Areas in
Communications, Vol. 18, No. 10, pp. 1885 – 1889, October 2000.
[8] J. Manchester, P. Bonenfant, C. Newton, “The evolution of transport
network survivability”, IEEE Communication Magazine, August 1999.
[9] O. Gerstel, R. Ramaswami, “Optical layer survivability: a service
Fig. 5. Wavelength requirement vs. traffic connectivity in regular network
perspective”, IEEE Communication Magazine, pp. 104 –113, March
topologies. 2000.
[10] Ch. Li, R. Ramaswami, “Automatic fault detection, isolation, and
number of wavelengths. This increase is almost linear. These results recovery in transparent all-optical networks”, Journal of Lightwave
Technology, Vol. 15, No. 10, pp.1784 –1793, October 1997.
are in line with other observations given in literature. [11] T.M. Chen, T.H. Oh, “Reliable services on MPLS“, IEEE
Communication Magazine, pp. 58 – 62, December 1999.
[12] K. Owens, S. Makam, V. Sharma, B. Mack-Krane, “A path
V. CONCLUSIONS protection/restoration mechanism for MPLS networks”, IETF draft, draft-
chang-mpls-path-protection-02.txt, July 2000.
In this paper we studied the design of “common-wavelengths” [13] V. Sharma, B.M. Crane, S. Makam, et al., “Framework for MPLS-based
scheme which can be used to build cost-effective survivable recovery”, IETF draft, draft-ietf-mpls-recovery-frmwrk-01.txt, November
2000.
WDM networks. We used ILP to solve the problem. Simulation [14] L. Mo, “General considerations for bandwidth reservation in protection”,
results confirmed that the proposed scheme reduces the number IETF draft, draft-mo-mpls-protection-00.txt, work in progress, July 2000.
of required wavelengths when a failure happens in the network. [15] D. Haskin, R. Krishman, “A method for setting an alternative label
switched path to handle fast reroute”, IETF draft, draft-haskin-mpls-fast-
reroute-05.txt, November 2000.
[16] P. Demeester, M. Gryseels, A. Autenrieth, et.al., “Resilience in
Multilayer Networks“, IEEE Communication Magazine, August 1999.
ICMP Covert Channel Resiliency
Kristian Stokes, Bo Yuan, Daryl Johnson, and Peter Lutz
Department of Networking, Security, and Systems Administration
B. Thomas Golisano College of Computing and Information Sciences
Rochester Institute of Technology
Rochester, New York 14623
Abstract—The ICMP protocol has been widely used and II. C URRENT C OUNTERMEASURES
accepted as a covert channel. While the ICMP protocol is very
simple to use, modern security approaches such as firewalls, deep-
Countermeasures exist within the ICMP covert channel
packet inspection and intrusion detection systems threaten the use realm, however they come at a cost. As with all aspects of
of ICMP for a reliable means for a covert channel. This study security, tradeoffs exist.
explores the modern usefulness of ICMP with typical security Blocking all ICMP traffic from entering the network pre-
measures in place. Existing ICMP covert channel solutions are vents all aspects of ICMP communication, including covert
examined for compliance with standard RFCs and resiliency with
channels. This methodology may not be acceptable due to
modern security approaches.
the loss of network troubleshooting abilities. Blocking ICMP
communications at a central firewall through the use of ACL
I. I NTRODUCTION mechanisms may not solve the problem. The capability to send
ICMP messages in an intranet environment may still exist.
Blocking specific ICMP messages from entering the net-
The Internet Control Message Protocol (ICMP) [1] is de- work. This methodology also incorporates the use of segment-
signed to provide feedback about problems in the communi- ing incoming and outgoing connections. If this countermeasure
cation environment. ICMP relies on the basic support of IP is enabled, both parties involved in the covert channel can
as part of a higher level protocol. Due to this dependency, develop fuzzing techniques to modify ICMP messages used
both ICMPv4 and ICMPv6 exist for both versions of IP. for communication or party initiation.
Many message codes exist within ICMP to properly diagnose Restricting the size of ICMP packets. By blocking large
network problems and traffic flow. ICMP messages are sent ICMP packets, an ICMP message with an extensive data field
under many different circumstances such as an unreachable will be dropped and perceived as a crafted packet with a
destination or general congestion control on the network. malicious or unknown payload. Large ICMP packets can be
Simple network troubleshooting utilities such as ping and used to test a network for proper handling of large packets.
traceroute utilize explicit ICMP messages to gather informa- An adversary can also overcome this limitation by fragmenting
tion about a network. their ICMP covert channel to smaller packets.
Covert channels often refer to a hidden information stream Traffic normalization. Traffic normalization techniques such
and more specifically, hidden streams embedded in IEEE as state preservation will reply and generate new packets on the
802 networks. Lampson [2] originally defined covert channels senders behalf. The normalizer, typically a firewall, will serve
under a number of categories such as storage, timing, termi- as a proxy, rebuild messages and construct new payloads. This
nation, resource exhaustion and power. Most covert channels activity essentially disrupts ICMP covert channel communica-
that involve the use of ICMP are largely storage channels tion by stripping out the message payload. State preservation
where unused fields are utilized for covert communication. requires significant computational power and may not scale to
ICMP as a covert channel provides many benefits due to particular environments with high traffic loads.
the overall simplicity. Only several fields exist within most
ICMP messages which enables quick implementation and III. R ELATED W ORK
simple channel setup/teardown. The idea of using ICMP as Research in this area focuses on existing solutions currently
a covert channel is to use an atypical communication protocol available for ICMP covert channel communication. Loki, an
rather than ordinary TCP or UDP. This will have a smaller ICMP tunneling back door application, tunnels remote shell
footprint across the network and may go unnoticed by network commands in ICMP echo reply / requests and DNS query /
administrators and traffic analyzers. What really makes the reply traffic. This proof of concept was originally published
ICMP protocol a viable covert channel is the use of data fields in Phrack World News [3] to demonstrate the vulnerabilities
or payloads within certain messages. By generating packets within these protocols. This implementation is very easy to
based on specific message codes and embedding the actual deploy and thusly carries a security risk of ICMP tunneling.
covert channel message in the data field enables ICMP to serve Another implementation named Ping Tunnel [4], is focused
as an alternate use for covert channels. These simple factors on reliably tunneling TCP connections using ICMP echo
enable ICMP to be considered as stealth traffic. request and reply packets. This tool can be with an outside
503
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_87, © Springer Science+Business Media B.V. 2010
504 STOKES ET AL.
proxy to tunnel traffic using ICMP messages back to the remote host. This solution can be utilized as a covert channel
requesting client. The exterior proxy serves all TCP requests when the operating network is heavily restricted.
and forwards the data back to the client via ICMP echo reply. Each solution will be tested and further explored in the
Although this solution can be viewed as subverting typical following sections in terms of resiliency with modern security
network communication, it is also important to realize other devices such as firewalls and intrusion detection systems.
potential uses for ICMP covert communication.
ICMP-Chat [6] implements a basic console-based chat V. E XPERIMENT
mechanism that utilizes ICMP packets for communication. A minimalistic test environment was created in a lab en-
A unique aspect of this solution incorporates the use of an vironment using VMware Workstation. Two CentOS work-
encrypted data field using AES-256. By implementing an stations are used as the endpoints of the covert channel. A
encrypted payload this further secures the covert channel, but Checkpoint NGX R65 firewall separates the network between
adds increased suspicions on an abnormal ICMP payload. internal and external clients. The internal network simulates a
With the increasing awareness of IPv6, covert channel tools small business with a firewall and IDS in place. The external
further expanded into the realm of ICMPv6. A tool v00d00N3t network simulates an Internet node communicating to the
[7] was developed to operate specifically over IPv6. This business. ICMP communication is permitted to the interior
tool focuses on the infancy of IPv6 protection technology. node, all remaining traffic is dropped at the firewall.
Dual IPv4/IPv6 routers are utilized to route IPv6 traffic to The Checkpoint firewall serves as a network layer inspection
send messages and files using ICMPv6. Useful information is engine to detect fields and patterns in the network layer.
embedded into fields other than the data field. For instance, the This firewall provides sufficient inspection for a network layer
sequence number tells the covert channel receiver if it should covert channel.
read the packet and ID field is used to identify how many A Snort IDS [8] is deployed inline on the network listening
bytes out of the payload to read. in promiscuous mode. All traffic passing through the internal
Additional defensive research has been performed to further network is captured and analyzed by the IDS. The standard
limit the capabilities of ICMP covert channels. A Linux kernel Snort ruleset was updated to the latest version from Sourcefire.
module was developed to scan ICMP messages for specific The basis for this experiment focuses on the resiliency of
signatures such as passwd, root, etc, ls and dir [5]. If these ICMP as a viable covert channel. ICMP-Chat and Ping Tunnel
signatures were detected, the ICMP message was scrubbed will be used in this environment to test resiliency against a
by zeroing out the data field while being processed by the modern firewall and IDS. Resiliency will be examined for
network stack. This technique is similar to normalizing traffic, each tool on connection establishment, connection integrity
however the action of data scrubbing is performed on the and overall covert design.
end nodes. This solution can also be deemed as a very
VI. R ESULTS
computational intensive process involving bi-directional deep-
packet inspection. The added network processing overhead Given the provided experiment, both ICMP-Chat and Ping
may not be acceptable for performance reasons. Tunnel are examined for ICMP resiliency. Each tool was
With these current solutions outlined, little information is installed and tested for functionality on both internal and
available for actual survivability of the tools and general external workstations.
ICMP covert channel message resiliency with common se-
A. ICMP-Chat
curity appliances. The use of intrusion detection, intrusion
prevention and firewalls are commonplace within a production By default, ICMP-Chat uses echo reply packets for primary
environment. To fully understand the capabilities of ICMP communication. Figure 1 illustrates the network topology of
as a covert channel, it must contain a moderate amount of the experiment. When comparing standard operating system
resiliency in modern networks. echo reply packets to echo reply packets generated by the
ICMP-Chat application many parameters follow standard RFC
IV. E XISTING S OLUTIONS compliance.
The standard CentOS echo reply, shown in Figure 2 consists
ICMP-Chat [6] is based on a simple console-based chat in- of several key focus areas such as packet size and data
terface which uses ICMP packets for communication. Features content. A total of 98 bytes are captured on the wire with
of this solution include the use of AES-256 encryption for a 56 byte data field. The data field also standardizes on the
encrypting chat data across the channel. ICMP-Chat provides following sequence of characters: “!#$%&’()*+,-./01234567”.
several protection mechanisms such as password protecting the For a perfect covert channel, the ICMP communication should
channel using SHA-256 and supporting the ability to change match this similar format to reduce detection.
usage of ICMP codes within the application. This allows for An ICMP-Chat echo reply session was initiated, sent with
mobility between ICMP codes to further obfuscate the channel. the message “test” and captured in Figure 3.
Ping Tunnel [4] allows a user to tunnel TCP connections Major differences with the reply packet in Figure 3 centers
to a remote host using ICMP echo request and reply packets. on the overall packet size of 318 bytes with a data size of
This is achieved through the use of a proxy serving as the 276 bytes. The dramatic increase in data size contributes to
ICMP COVERT CHANNEL RESILIENCY 505
507
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_88, © Springer Science+Business Media B.V. 2010
508 SULTANA ET AL.
layer through cover glass, an electric field is applied to fundamental mechanisms of optical propagation is
the layer to change the refractive index that depends on through the process of diffraction. In phase only
molecular orientation of liquid crystal layer block. The hologram or grating diffraction is done without
index change induces a phase-only modulation of the blocking photons. By using a hologram or grating to
input light. It is then reflected from the aluminum steer the light by diffraction, we can correct for any
mirror and passes back through the liquid crystal layer motion or misalignments and create an adaptive optical
[7]. Multilevel phase modulation can be achieved using interconnect. The SLM must maintain DC balance
a nematic material with the correct birefringence and during the display of each corrective pattern. [8]. A
molecular orientation [8]. The SLM chip is mounted in controller algorithm is used to design a suitable
a ceramic Pin Grid Array (PGA) package. This PGA adaptive binary holograms to calculate the amount of
package connects to a signal controlling board for phase modulation required for alignment of the light
receiving and sending data, power and control signals beam. In order to achieve active alignment we need to
[9]. The signal is collimated by the lens onto the SLM. generate holograms in real time thus we can use a non-
By controlling the deflection angle by SLM, the beam iterative algorithm [8]. When the light beam incidents
may be made to return to a selected point by on the SLM then the phase modulation causes it to split
collimating again by a lens. into a number of reflected beams and a phase
Following figures shows a cross-sectional view of a modulation pattern is created. Each beam corresponds
liquid-crystal-on-silicon spatial light modulator to a different diffraction order of the phase modulation
illustrates how the liquid-crystal layer causes a voltage- of which only one is created intentionally to carry the
dependent phase shift upon entering light. Each pixel is signal into the selected output point. The other
fully programmable, offering a high-resolution pure diffraction orders are unwanted and they may lead to
phase-modulating array [6]. crosstalk and back reflection. The first-order diffraction
efficiency of a binary-phase hologram with LC layer
thickness, and molecular tilt angle θ is given by
equation (1), where Δn is the birefringence and λ is
the operating wavelength [3].
§ 2π Δ nd ·
η = sin 2 2θ sin 2 ¨ ¸ (1)
© λ ¹
IV. PROPOSED MODEL Figure 6 shows the second interconnect scenario, where
multicasting interconnects can be established by using
The descriptions and functions of SLM and hologram multicasting phase hologram. Both SLMs are loaded
design were stated above. The complete mechanism of with proper steering holograms to steer the beams of
our proposed model is give here: VCSEL #2 to PDs #1 and #2.
Each chip has a VCSEL & a Photo Detector (PD) with
it. To communicate with another chip, each chip send
light beam by using VCSEL & in the designated chip
this beam of light is detected by using a PD. For each
VCSEL an individual pixel of SLM is assigned. Each
PD also has an individual pixel assigned to it. Light
from a VCSEL is collimated and focused by using a
collimating lens and it incidents on the first SLM. The
addressing & other information are sent to the signal
controlling board. Then the hologram pattern is
generated using Fourier series analysis and the required
diffraction angle is calculated by using an appropriate
controller algorithm. Light beam is then phase shifted Fig. 7. N-to-N parallel interconnect scenario.
and directed toward the pixel of the second SLM which
Figure 7 shows the third interconnect scenario, where
is assigned to the destined PD. The second SLM again
parallel optical interconnects can be established
do the same things and direct the light beam to the
simultaneously by two SLMs using the VCSELs and
specific PD. A collimating lens is used to collimate and
the PDs. Both SLMs are loaded with proper steering
focus the light beam to the PD for a chip. The different
holograms to steer the beams of VCSEL #1, #2, and #3
scenarios of this model that can be implemented are as
to PD elements #3, #1, and #2.
follows:
V. CONCLUSION
relatively insensitive to pixel errors there are excellent [2] Katsuhiko Hirabayashi, Tsuyoshi Yamamoto, Shigeki
prospects for high yield fabrication of the devices. It is Hino, Yoshitaka Kohama, and Kouta Tateno, “Optical
Beam Direction Compensating System for Board-to-
also possible to correct for optical aberrations in the Board Free Space Optical Interconnection in High-
system. Finally, the prospect of liquid crystal on silicon Capacity ATM Switch”. Journal of Lightwave
(LCOS) devices means that, in principle, the necessary Technology, Vol. 15, No. 5, May 1997.
control electronics could be integrated on to the device [3] W. A. Crossland, I. G. Manolis, M. M. Redmond, K. L.
to make a very compact system [4]. Tan, T. D. Wilkinson, M. J. Holmes, T. . Parker, H. H.
There are two key challenges of implementing this Chu, J. Croucher, V. A. Handerek, S. T. Warr, B.
Robertson, I. G. Bonas, R. Franklin, C. Stace, H. J.
model. First, as the SLM is pixellated, so the possible White, R. A. Woolley, and G. Henshall, “Holographic
hologram periods must be an integer multiple of the Optical Switching: The “ROSES” Demonstrator”;
pixel pitches. Second, the unwanted diffraction orders Journal of Lightwave Technology, Vol. 18, No. 12,
that can lead to crosstalk and back reflection must be pp 2,3, December 2000.
routed and controlled appropriately, otherwise these [4] C.J. Henderson, B. Robertson, D. Gil Leyvaa, T.D.
will greatly reduce the performance of the system. Wilkinson, D.C. O’Brien, G. Faulkner Cambridge
Univeristy Engineering Department, Trumpington St,
Finally, SLM consists of a layer of nematic liquid- Cambridge, CB2 1PZ, UK Department of Engineering
crystal, but the nematic devices have a significantly Science, University of Oxford, Parks Road, Oxford, OX1
slower switching speed than ferroelectric devices. 3PJ, UK. “Control of a free space adaptive optical
In this preliminary study we have considered only chip- interconnect using a VCSEL array transmitter and FLC
to-chip interconnection. This proposed model of optical SLM for beam steering”; [email protected], pp 1.
interconnects can also be applied in board-to-board [5] Tod Laurvick, ”Chip-To-Chip Optical Interconnection
Using MEMS Mirrors”, March 2009.
communication.
[6] http://www.laserfocusworld.com/articles/255505.
[7] http://jp.hamamatsu.com/en/rd/publication/scientific_am
REFERENCES erican/common/pdf/scientific_0806.pdf.
[8] T. D. Wilkinson, C. D. Henderson, D. Gil Leyva and W.
[1] Muhsen Aljada and Kamal E. Alameh , Yong-Tak Lee A. Crossland, “Phase modulation with the next
and Il-Sug Chung, “High-speed (2.5 Gbps) generation of liquid crystal over silicon technology”,
reconfigurable inter-chip optical interconnects using Journal of Materials Chemistry, pp 1-4, June 2006.
opto-VLSI processors”, Centre for
MicroPhotonicSystems, Electron Science Research [9] Boulder Nonlinear Systems, “32-channel Spatial Light
Institute Joondalup, WA, 6027, Australia; pp 3-8, 2006 Modulator Driver Subsystem”, Boulder Nonlinear
Optical Society of America. Systems, pp 2, June 1, 2001.
GA Based Neural Network Retraining
Using Adaptive Genetic Operations
C.I.Bauer; H.Yu; B.Boffey
Faculty of Computing, Engineering and Technology, Staffordshire University, Stafford, England
[email protected]; [email protected]; [email protected]
511
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_89, © Springer Science+Business Media B.V. 2010
512 BAUER ET AL.
system to predict future movement, through the information To create the initial generation to start the GA the
in the distribution probability matrix – maintained for each weight information of the trained and out-of-date NN was
cell within the network grid – and the previous movement used as a basis, as it can be argued that the ideal set of
step, with a certain level of accuracy. However, as the weights for the new traffic patterns are very close to the new
movement patterns of mobile devices change over time this optimal solution. Hence this holds the potential to further
will affect the prediction success rate of the system. Without reduce training times. To generate the individuals of the
retraining with new route information that reflects the initial population, each gene of the original set of weights
changed movement pattern of mobile devices the prediction was modified by a small, randomly selected value.
success rate will drop over time. As full retraining using The remaining parameters used for the GA were a
traditional training mechanisms such as the backpropagation mutation rate of 1/N (with N being the length of the
algorithm can be very time-consuming and inconvenient at chromosome), single point crossover with a crossover rate
times, GAs can offer an alternative solution to the problem, of 50% using the roulette wheel selection algorithm with
or at least provide a temporary measure to allow the system elitism as well as the population size of 180 individuals. The
to continue with acceptable performance levels until full number of individuals selected for the next generations was
retraining is possible. This algorithm could be employed set to 50% of the population size. These parameters were
once the prediction success rate has dropped below a established through initial simulations to optimise the GA
defined threshold to improve the prediction success rate. retraining approach. The simulations were run with 10, 20,
50 and 100 generations and the results for the parameter set
When applying GAs to a problem there are two main optimised for the retraining approach compared against the
requirements as defined in [11]: firstly, a suitable way to standard settings as defined by DeJong [12]. The changing
represent the problem/solution needs to be found. Secondly, traffic patterns were simulated through data sets with
in order to be able to evaluate the solutions of each varying levels of known regular route information ranging
generation, a fitness function needs to be defined that allows between 100%–50% in steps of 10% the NN had
for selecting the best solutions from each generation with encountered during training. The remaining route
which to continue the evolutionary process. In addition to information within the data sets consisted of unknown
this parameters such as population size, number of regular route information that had not been experienced
generations, how to implement the genetic operations and previously. The results of the simulations can be seen in the
apply them to the defined chromosomes as well as any diagram below which is showing the achieved fitness for the
additional parameters that arise due to the scenario-specific different number of generations averaged across the
implementation have to be identified. In the case of the investigated data sets for ease of representation:
given scenario using NNs in combination with GAs, two
parameters are indirectly given through the design of the 82
NN and the way its performance is evaluated. As the aim of
the system is to successfully predict the next visited cell of a 81
mobile device, the prediction success rate is used as the
Fitness in %
As can be seen from the diagram the results received for the
customised parameter set using 20, 50 and 100 generations
for varying levels of new unknown regular route
information within the data set clearly outperformed the
standard DeJong settings for Neural Network retraining.
However, although the simulation results showed that the
GA retrained network can improve the fitness value
compared to the original network without retraining, the
Fig. 1. Conversion of NN weight information into chromosome results obtained were lower than expected. Therefore, the
characteristics of the provided route information were
This representation was chosen because the vital investigated further to identify a reason for this. The weight
information of a trained network is stored in the weights for information of networks trained with route information of
the connections between the different nodes and these are 100%, 95% and 90% of regular known routes was
the parameters that determine the prediction success rate for investigated. It was observed that with each progressing
a given traffic pattern. When the traffic patterns change this layer of hidden nodes, the weight difference of the network
is the information that needs updating to maintain prediction trained with 100% regular routes compared to 95% and 90%
performance levels. as well as the network trained with 95% regular routes
GA BASED NEURAL NETWORK RETRAINING USING ADAPTIVE GENETIC OPERATIONS 513
compared to 90%, increased, as depicted in the diagram As in the case of the non-adaptive GA investigation, the
below: parameters for the adaptive implementation of GA for the
purpose of neural network retraining, were tested overall to
18 allow comparison and performance analysis. With the
Layer 3
16 exception of the parameters for the adaptive elements
Layer 2
14 identified through the preliminary investigation, all
Layer 1
12 remaining parameters remained the same as in the case of
Difference
100
Degradation over time
90 Adaptive Retraining
Fitness in %
80
70
60
50
100 90 80 70 60 50
Fig. 4. Chromosome with weights encoded per network node Percent of Known Regular Route Information
Time in Seconds
run without retraining. As expected the improvement is more 600
significant for data sets with a higher percentage of new
400
unknown regular route information as it is for data sets with
a lower percentage of new unknown regular route 200
information. A maximum of over 15% for the data set with 0
50% known route information was established. This is 50 60 70 80 90 100
confirmation that using an adaptive GA retraining scheme is Percent of Known Regular Route Information
a viable solution to provide temporary relief to NN-based Fig. 9. Comparison Training Times Neural Network versus GA
prediction systems in the presence of changing traffic
patterns until full retraining is possible. Looking at the time measurements for the two
retraining techniques in figure 9 the Backpropagation
approach shows steadily increasing time requirements with
As the improvement in the prediction success rate is
a rising percentage of unknown route information within the
only one of the deciding factors on the success of the GA
data sets. For the GA approach there also seemed to be
based NN retraining approach for the given problem further
rising time requirements for data sets with higher levels of
studies need to be conducted. To investigate whether the use
known route information which, after the first few data sets,
of GAs for retraining would provide a time benefit for the
however, seemed to level off. Overall seen, the retraining
prediction scheme, time measurements were taken and
time requirements of the GA approach were well below
compared to the prediction success rates. As a comparison
those of the Backpropagation approach clearly confirming
algorithm the original Backpropagation approach was used.
the time benefit of the GA scheme. This is down to the fact
Subject to the investigation were data sets with varying
that the GA algorithm in previous experiments seemed to
levels of known regular route information ranging from 50 –
converge after a relatively small number of generations as
100% in steps of 10%. To obtain representative time
can be see in figure 7 which shows the best and the average
measurements three simulation runs were conducted and the
population score for an example run of the GA.
results averaged across the three runs per data sets. Obtained
simulation results are shown below:
Considering the results of both graphs it can be concluded
that looking at the prediction success rate alone, the
90 Backpropagation algorithm appears to be the better choice
85 of retraining the network. However, considering the original
80 aim of the GA retraining, this approach has a clear
advantage on the training times. As the GA retraining is
Fitness in %
75
70 only meant to be used as a temporary measure to extend the
65 life of the Neural Network until full retraining is possible
60 again this proves that this is a viable way to achieve this. In
55 light of the obtained simulation results, taking into
50 consideration the time saving and the improved prediction
1 11 21 31 41 51 61 71 81 91 101 111 121 131 141 151 161 171 181 191
Number of Generations
success rate, it can further be concluded that when using the
GA approach for retraining to set a threshold value to a
Fig. 7. GA Training Cycle
suitable value to benefit from the better prediction success
rate achieved by the GA algorithm.
100
90
IV. CONCLUSION
Fitness in %
80
In this paper a GA-based retraining algorithm using
70 adaptive mutation and crossover to assist the NN based
60 NN success rate prediction systems in maintaining its performance levels in
50
GA Success Rate the presence of changing traffic patterns is presented. As
50 60 70 80 90 100 simulation work has shown, the adaptive layer-based
Percent of Known Regular Route Information implementation of the genetic operations provided
Fig. 8. Comparison NN versus GA prediction success rate improvements to the system’s capabilities to increase the
prediction success rate (fitness) compared to the standard
mutation genetic operations. This is a strong indication that
when applying genetic operations, the structure of the
weight information contained within a chromosome should
be taken into consideration.
GA BASED NEURAL NETWORK RETRAINING USING ADAPTIVE GENETIC OPERATIONS 515
Further to the investigation of enhancing the prediction [4] Doulamis, A.D.; Doulamis, N.D.; Kollias, S.D., “An
Adaptable Neural-Network Model for Recursive Nonlinear
success rate through adaptive genetic operations when Prediction and Modeling of MPEG Video Sources”; IEEE
retraining a NN, time measurements were taken to compare Transaction on Neural Networks; Volume 14, Issue 1; January
the traditional Backpropagation algorithm and the GA based 2003 Page(s): 150 – 166.
retraining algorithm with respect to retraining time and [5] Ichibuchi, H.; Nakashima, T.; Nii, M., “Learning of Neural
achieved prediction success rate. Simulation showed that Networks with GA-Based Instance Selection”; Joint 9th IFSA
World Congress and 20th NAFIPS International Conference,
looking at the prediction success rate the Backpropagation 2001; Volume 4, 25-28 July 2001; Page(s): 2102 - 2107 vol.4.
algorithm clearly outperformed the GA retraining approach. [6] Ling, S.H.; Lam, H.K.; Leung, F.H.F.; Tam, P.K.S., “A
However, with respect to retraining times, the GA retraining Genetic Algorithm Based Neural-Tuned Neural Network”;
approach showed a clear time benefit over the traditional IECON ’03; The 29th Annual Conference of the IEEE –
retraining approach, although not achieving the same Industrial Electronics Society, 2003; Volume 3, 2-6 Nov.
2003 Page(s): 2423 – 2428 Vol.3.
accuracy levels as the Backpropagation approach. As the
[7] Zuo, Guoyu; Liu, Wenju; Ruan, Xiaogang, “Genetic
adaptive GA based NN retraining approach is intended as a Algorithm Based RBF Neural Network for Voice
temporary measure to bridge the time until full retraining is Conversion”; WCICA 2004; Fifth World Congress on
possible, again this proves that this is a viable way to Intelligent Control and Automation, 2004; Volume 5, 15-19
June 2004 Page(s): 4215 – 4218 Vol.5.
achieve this.
[8] Zhangang, Yang; Yanbo, Che; Cheng, K.W. Eric, “Genetic
Algorithm-Based RBF Neural Network Load Forecasting
Model”; IEEE Power Engineering Society General Meeting,
REFERENCES 2007; 24-28 June 2007 Page(s): 1 – 6.
[1] Nastac, I.; Matei, R. “Fast retraining of artificial neural [9] Ji, Tao; Pang, Qingle; Liu, Xinyun, “Study of Traffic Flow
networks”; Rough Sets, Fuzzy Sets, Data Mining and Forecasting Based on Genetic Neural Network”; Sixth
Granular Computing, Springer-Verlag in the series of Lecture International Conference on Intelligent Systems Design and
Notes in Artificial Intelligence; 2003; pp. 458.462; ISBN: Applications – ISDA ’06; Volume 1, Oct. 2006; Page(s): 960
978-3-540-14040-5. – 965.
[2] Nastac, I.; Costea, A. “A Retraining Neural Network [10] Bauer, C.I.; Yu, H.; Boffey, B. “Adaptive Genetic Algorithm
Technique for Glass Manufacturing Data Forecasting; for Neural Network Retraining”; International Joint
International Joint Conference on Neural Networks”, 2004; Conferences on Computer, Information, and System Sciences
Proceedings 2004; Volume 4; 25-29 July 2004 Page(s): 2753 – CISSE ’08; 5-13 December 2008
– 2758. [11] Núñez, Edwin; Banks, Edwin Roger; Agarwal, Paul, “High
[3] Nastac, I.; Cristea, P., “ANN Flexible Forecasting for the Performance Evolutionary Computing”; HPCMP Users Group
Adaptive Monitoring of Multi-Tube Reactor”; 6th EURASIP Conference 2006, June 2006; Page(s): 354 - 359.
Conference focused on Speech and Image Processing, [12] DeJong, K.A. and Spears, W.M. "An Analysis of the
Multimedia Communications and Services; 14th International Interacting Roles of Population Size and Crossover in Genetic
Workshop on Systems, Signals and Image Processing 2007; Algorithms," Proc. First Workshop Parallel Problem Solving
27-30 June 2007; Page(s): 193 – 196. from Nature, Springer-Verlag, Berlin, 1990. pp. 38 - 47.
Decision Support for IP Uplink and Routing Choice
A Whole View of the Internet
Peng Su Zhengping Wu
Department of Computer Science and Engineering Department of Computer Science and Engineering
University of Bridgeport University of Bridgeport
221 University Avenue, CT 06604, USA 221 University Avenue, CT 06604, USA
[email protected] [email protected]
517
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_90, © Springer Science+Business Media B.V. 2010
518 SU AND WU
˄2˅To get rid of the repeated AS number generated by The second algorithm is designed to regulate and classify
AS prepending (append AS_NUM to AS path), the BGP prefixes. It is named “Class Algorithm,” as
search the database by AS number, and update the illustrated in figure 6. The tasks of this set of programs can
database. be described as:
520 SU AND WU
Output
Database BGP, table "ROUTES_BY_AS"
configurations [8]. When C1 sets its main link to AS2 and issue. As a result, the customer stub ASes suffer from
undergoes some remote congestion or oscillations [4], this them.
˄2˅Even though one stub AS uses more than two links to
AS will not automatically switch its link to AS3. However, if
the uplink AS, these links functions only load balance
putting this situation to the support decision system, the fruits
will be like this: [12] and failover, which will not be helpful to the
previous issues.
Key = AS1 ˄3˅In many cases, the traffic between two stub ASes is
Neighbor number = n huge, which hold the most bandwidth of these areas.
Neighbor information of AS1: <AS1, ASn>, <AS1, AS3>, However, these packets have to go through the hub
<AS1, ASj> … <AS1, ASk> area and affected by traffic policies and unexpected
routing questions.
Key=C1 The first and second questions may occur when some
Neighbor number = 2 uplink ASes (maybe some ISPs or domains) conceal the
Neighbor information of C1: <C1, AS3>, <C1, AS2> topology by summary action or stub areas only use default
route. Then these stub ASes are blind to the topology outside
Key=AS2 local area. Nevertheless, in our decision support system, the
Neighbor number=2 adjacency relationships of stub ASes and other IXP are
Neighbor information of AS2: <AS2, C1>, <AS2, ASn> obvious. Thus, as described in the previous section, stub
ASes can get the nearest IXP information, at least in the
Key=AS3 aspect of AS-PATH, and then make a proper decision for the
Neighbor number=2 network design.
Neighbor information of AS3: <AS3, C1>, <AS3, AS1> Nowadays, it is pretty common for the third question over
the Internet. Consequently, traffic has to endure unnecessary
Key=ASn hops between stub areas. Figure 8 and 10 illustrate this issue.
Neighbor number=2 Figure 9 and 11 provide a solution.
Neighbor information of AS3: <ASn, AS1>, <ASn, AS2>
; ;
Let A and B { אNeighbor information of C1, AS2,
AS3, ASn}, suppose that x A ¤ B if and only if x A and
x B, and let M = A ¤ B, x=AS1, then (xęM) Æ ( ęǡ
ę) Æ AS1 ę Neighbor information of AS3 OR AS1 ę
Neighbor information of ASn. < = < =
From the calculation of ASes’ adjacency relationships, it is
obvious that the congestion may occurred either in <AS1, Figure 8. Unnecessary hops between Figure 9. Refined network
ASn> or in <AS2, ASn>. Furthermore, AS3 is closer to AS1, two stub ASes and one hub AS architecture between stub ASes
from the AS-PATH aspect. Thus, changing main link of C1
will solve the problem. For another question, as illustrated in Figure 8 presents a simple “hub and spoke” topology,
figure 4, although C1 may not be sensitive to the impacts of where X is the hub AS, and stub area Y as well as Z work as
AS2’s routing policies, after checking the topology between spoke. When X forwards the packets from Y to Z, the hops
source and destination, C1 still can avoid the redundant hops between Y and Z is 2. However, when the decision support
by switch its main link to its backup link. Therefore, by system shows that AS Y and AS Z connect to the same hub
introducing the decision support system using the whole view AS - AS X, a more appropriate solution emerges, that is to
of the Internet, these two previous issues can be resolved. add a link between Y and Z, as shown in figure 9.
Besides, in the view of IXP, there will be similar issues to This situation could be more complicate when introducing
AS1, as described in figure 3 and figure 4. While the more hub ASes into this scenario, as figure 10 shows.
difference between C1 and AS1, in this scenario, is that the
situations of IXP would be more complex and more difficult ; 1 ˞
to deal with. However, following the normal forms explained
in the previous example, the adjacency relationships, EBGP
peer numbers, route entry scales make the relationships
among neighbors and the entire Internet topology clearer.
Moreover, when take the ranks of ASes, the stub areas, and < = ˟ ˠ
stability of ASes into account, this decision system is able to
provide more useful ranking information to make decisions. Figure 10. Unnecessary hops between two stub ASes and various hub
Let H(i, j) defines the hops between two different ASes.
B. The future of stub AS Suppose the traffic between AS Y and AS holds a great
According to the discovery of whole adjacency amount of each AS. When Y needs to access , then the least
relationships of ASes over the Internet, the percentage of stub hops between from Y to are:
AS over the Internet reaches 57.27%, so that the star
topology is still dominant and prevalent in the Internet. But
as a reality, this is an unstable structure, for below reasons:
˄1˅Many stub ASes use only a default route to other Therefore, each hop along the road from Y to could
ASes, whereas their unique uplink ASes could arise influence the quality of transmission. As mentioned above,
congestion, redundant hops, as well as packet loss after knowing the adjacency relationships between Y and ,
some solutions, as figure 11 shows, could be either
522 SU AND WU
Abstract. Blow-CAST-Fish is a sixteen-round Feistel cipher in the key, we will analyze the properties of F function under
which the F function is a part of the private key. In this paper, we weak key assumptions. We will try to identify differential
show that the disclosure of F allows perform a differential characteristics with high probability for round function F in
cryptanalysis against a number of rounds. We firstly identify the which a non-zero inputxor resulted in a zero outputxor under
properties of F function with three active S-boxes to construct some weak keys, then the differential cryptanalysis for
the 6-round differential characteristic with the probability 2 -61
8-round and 16-round of Blow-CAST-Fish will be given. As
and the differential weak keys occur with the probability 2-12. the attack assumption in [4], we also assume the opponent
Then we identify the properties of F function with only one knows the part of the key which describes the F function, that
active S-box to construct the 14-round differential characteristic is the four S-boxes. And the main idea of the 8-round attack is
with the probability 2-49 and the differential weak keys occur from reference [9].
with the probability 2-52. With the above differential The paper is organized as follows. Section 2 introduces the
characteristics, we can attack 8-round and 16-round description of Blow-CAST-Fish. We present how to attack
Blow-CAST-Fish respectively. 8-round and 16-round Blow-Cast-Fish in Section 3 and
Section 4 respectively. We conclude this paper in Section 5.
Introduction
Blow-CAST-Fish is a block cipher designed by Description of Blow-Cast-Fish
Krishnamurthy G.N et.al in 2008[1]. It uses the features of As a Feistel block cipher, Blow-Cast-Fish uses a block size
Blowfish [2] and CAST-128 [3]. of 64 bits, and the key size can vary from 32 bits to 448 bits.
S.Vaudenay presented the differential cryptanalysis for The cipher consists of two parts: a key expansion part and a
Blowfish based on the differential characteristics for round data encryption part.
function with non-zero inputxor and a zero outputxor [4]. J. Key expansion converts a key of at most 448 bits into
Nakahara Jr and M. Rasmussen presented the first concrete several subkey arrays totaling 4168 bytes following a
linear cryptanalysis on reduced round CAST-128 and scheduling scheme which works as a pseudo random
CAST-256 in [6], for CAST-128, they can recover the subkey generator. As this scheme is very complicated, one has to
for the 4th round with 237 known plaintexts and 272.5 times of store definitely the expanded key. Our cryptanalysis is not
4-round CAST-128 encryption. M. Q. Wang and X. Y. Wang related to the key schedule, so we don’t present it in detail.
et.al presented the linear cryptanalytic results for 6-round The expanded key consists of:
CAST-128 with 253.96 known plaintexts and 288.51 times of z 18 32-bit round subkeys P1,ĂĂP18˗
6-round encryption in [8]. Haruki Seki and Toshinobu z Four arrays of 256 32-bit values which describe
Kaneko found the 2-round differential characteristics for F2 four S-boxes S1, S2, S3 and S4 with 8-bit inputs and
round function of CAST-128 in [5], in which a non-zero 32-bit outputs.
inputxor resulted in a zero outputxor, then they could attack The four S-boxes define 32-bit to 32-bit round dependent
5-round CAST-128. function F as follows:
As we know, there is no attack on Blow-CAST-Fish until (Divide the input of F into four 8-bit quarters: a, b, c and d)
now. Due to the S-boxes of Blow-CAST-Fish are related with Rounds˖1ˈ4ˈ7ˈ10ˈ13ˈ16
523
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_91, © Springer Science+Business Media B.V. 2010
524 SUN AND WANG
+ S2[b])-S3[c])+S4[d]
F1= ((S1[a]ƻ function in round 2 is as follows:
Rounds˖2ˈ5ˈ8ˈ11ˈ14 I=IaýIbýIcýId=(L1ƻ + P2)<<<kr2
+ S4[d]
F2= ((S1[a]-S2[b])+S3[c])ƻ + S4[Id]
F2= ((S1[Ia]-S2[Ib])+S3[Ic])ƻ
Rounds˖3ˈ6ˈ9ˈ12ˈ15 The mapping (Ia,Ib ė61>,a@62>,b@ is a 16 to 32 bits
+ S3[c])-S4[d]
F3= ((S1[a]+S2[b])ƻ function, so it may have a collision61>,a@62>,b@ 61>,a’@
Where “ ” is the bit-wise xor, “+” is the addition modulo 62>,b’@with high probability.
232 and “ˉ” is the subtraction modulo 232. Let ¥=Ia+Ia’, =Ib+Ib’, assuming there is only one collision
The encryption process is defined as follows: for S1-S2ˈso the probability of characteristic ¥00ė0000
z Split the plaintext into left and right 32-bit halves + S4[Id] is 2-15. So 2-round
from I to ((S1[Ia]-S2[Ib])+S3[Ic])ƻ
XL and XR . differential characteristic for Blow-Cast-Fish with the
z For 1¡i¡16 probability 2-15 can be produced,
XL=XLƻ + Pi R1: (0000ý(¥00>>>kr2))((¥00>>>kr2ý0000)
XL=XL<<<Kri R2: ((¥00>>>kr2) 0000)(0000ý(¥00>>>kr2))
(Kri=last 5-bit of Pi) We search for 210 keys randomly, each key produces four
XR=F(XL)ƻ + XR S-boxes according to the key schedule of the cipher, then
Swap XL and XR (Undo the last swap) test the number of keys which make S1-S2 have a collision.
z XR=XRƻ
+ P17 We tested for twenty times to know that about half of every
+ P18
XL=XLƻ 210 random keys will produce S-boxes which have a
Recombine XL and XR as the ciphertext. collision with S1-S2. The value of ¥ý is different for
The encryption process is shown in figure 1. differential keys, and for most time there is only one value
of ¥ý for the same weak key. So the weak keys that cause
Differential attack on 8-round Blow-Cast-Fish differential characteristic ¥00 ė 0000 for the round
For convenience, we denote the input of ith round as function in the second round occurred with probability 2-1.
(Li-1,Ri-1). We consider the second round at first, the round We will try to extend the above 2-round differential
characteristic to 6 rounds. First we only consider kr2 as zero,
Plaintext the other cases will be considered later.
Round 1
As Iƻ+ I’=¥00<<<kr4 and kr4 varies from 0 to 31, we find 2. For every four S-boxes which have a collision with
+ I’ be zero, i.e. one of
out that there is always a byte of Iƻ S1-S2 and the corresponding inputxor ¥ý , we exhausted
the S-boxes always be non-active. So the probability of the 232 values of L3 and the 25 values of kr4 in the 4th round.
differential characteristic ¥00<<<kr4 ė 0000 from I to If , the key used is the weak key that results the
+ S2[Ib])-S3[Ic])+S4[Id] is 2-23 and the probability of
((S1[Ia]ƻ differential characteristic ¥00<<<kr4 ė0000 for the 4th
4-round differential characteristic is 2-38 . round.
We will test the proportion of weak keys together with 3. For the keys which satisfy both condition 1 and 2, we
the 6th round later. exhausted the 232 values of L5 and the 25 values of kr6 in the
6th round. If , the key used is weak key that results the
Then we extend the above 4-round differential differential characteristic ¥00<<<kr6 ė0000 for the 6th
characteristic to 6 rounds. The inputxor of the 5th round is round.
zero, so the outputxor of the 5th round must be zero and the The key which satisfies the three conditions is the weak
inputxor of round function F3 in the 6 th
round must be key for the 6-round differential characteristic.
¥00(kr2=0), and the outputxor has not been decided. The 4. Test for twenty times to get the steady proportion of
differential characteristic of F3 in the 6 th
round can be weak keys.
denoted as follows,
I=IaýIbýIcýId=(L5ƻ + P6)<<<kr6 We tested that the weak keys occurred with the
’ ’ ’ ’ ’
+ ¥00)ƻ
I =Ia ýIb ýIc ýId =((L5ƻ + P6)<<<kr6 probability 2-2. There is at least one solution for kr4 which
+ (((S1[Ia’]+S2[Ib’]) ƻ
+ S3[Ic])-S4[Id])ƻ
2=(((S1[Ia]+S2[Ib])ƻ + satisfies the differential characteristic in the 4th round for
’ ’
S3[Ic ])-S4[Id ]) the same keys, so the satisfied kr4 occurred with probability
We consider the case that 2 equals to zero. 2-5. Similarly, the satisfied kr6 occurred with probability 2-5.
+ I’=¥00<<<kr6 and kr6 varies from 0 to 31, we find
As Iƻ The probability for the above 6-round differential
’
+ I be zero, i.e. one of
out that there is always a byte of Iƻ characteristic is 2-61 (see figure 2), and the 6-round
the S-boxes always be non-active. So the probability of differential weak keys occur with the probability 2-52-52-5
differential characteristic ¥00<<<kr6 ė 0000 from I to 2-2 2-17 .
+ S3[Ic])-S4[Id] is 2-23 and the probability of
((S1[Ia]+S2[Ib])ƻ In order to further increase the proportion of the
the 6-round differential characteristic is 2-61 (see figure 2).
differential weak keys for 6-round differential characteristic,
we try to test the differential characteristics with the
Due to S-boxes are key dependent, in order to obtain the
inputxor ¥00>>>kr2 and zero outputxor for round
proportion of weak keys for the 6-round differential
functions in the 4th and 6th round as kr2 is not zero.
characteristic, we need to identify the weak keys which
As kr Į 0, the differential characteristics for round
cause the following three conditions :
functions in the 4th and 6th round are (¥00>>>kr2)<<<kr4ė
1. There is a collision for S1-S2 with differential
0000 and (¥00>>>kr2)<<<kr6 ė 0000 respectively.
characteristic ¥ýė0ý0. 5
Obviously, for each value of kr ę 2 , there must be
2. 1=0 for the 4th round differential characteristic.
solutions of kr4 and kr6 that satisfy the differential
3. 2=0 for the 6th round differential characteristic.
characteristics. It means that for each non-zero kr, the
We test the proportion of weak keys for the 6-round
proportion of weak keys in the 4th and 6th round is same.
differential characteristic as follows:
So the differential weak keys which produce the 6-round
1. Select 27 keys of Blow-Cast-Fish randomly, and differential characteristic occur with the probability 25
identify the keys produce four S-boxes which have a 2-17 2-12.
collision ¥ýė0ý0 with S1-S2 .
526 SUN AND WANG
0000 δμ 00
3.3 Complexity Analysis
P1
In this attack, k = 37 is the total subkey bits of the 8th round,
Kr1
<<<
0000
F1 p=2-61, =2-32 because the ciphertext pairs can be filtered
according to their left halves XOR value ¥00>>>kr2. For a
p=2-38
random ciphertext pair, we denote C=(CLýCR), since we
have
0000 δμ 00 F2((CLƻ + F2((CL’ƻ
+ P10)<<<kr8)ƻ + P10)<<<kr8) =xyzt
P5 Kr5
We can exhaustively try all the 237 possible values for (P10,
0000 p=1
<<< F2 kr8) until the above equation holds. Given the ciphertext pair,
on the average there are 25 solutions for the equation, so the
δμ 00 0000
average count of keys per analyzed pair is equal to 25.
P6 Kr6
Consequently,
βγσθ
<<< F3 p=2-23 2−61 × 237
S N= = 23
25 × 2−32
0000 δμ 00 The data complexity is 263 plaintexts. The time complexity
p = 2−61
¥000>>>¢ 0000
δμ 00 >>> kr2 0000 P2 Kr2=¢
<<< ¥000 F2
-7
p=2
P6 Kr6
βγσθ
<<< F3
<<<
0000
F1 p =1 <<< ¥000 F2 p=2
0000 ¥000>>>¢
References
P15
[1] Krishnamurthy G.N, Dr. V. Ramaswamy, Leela G.H
Kr15 p=1
<<<
0000
F3 and Ashalatha M.E, “Blow- CAST-Fish: A New 64-bit
Block Cipher”, Bapuji Institute of Engineering and
Techology, Davangere-577004, Karnataka, India.
¥000>>>¢ 0000
[2] B.Schneier, “Description of a New Variable-Length Key,
P16 Kr16
βγσθ p=1 64-Bit Block Cipher (Blowfish)”,Fast Software
<<< F1
xyzt Encryption, Cambirdge Security Workshop proceedings
P18 P17 (December 1993), Springer-Verlag, 1994, pp.191-204.
¥000>>>¢ xyzt [3] Adams, C. “The CAST-128 Encryption Algorithm”.
RFC 2144, May 1997.
Figure 5: 16-round Blow-CAST-Fish [4] S. Vaudenay, “On the weak keys of Blowfish,” Fast
Software Encryption (FSE’96), LNCS 1039, D. Gollma-
2−49 × 237 nn, Ed., Springer-Verlag, 1996, pp. 27–32.
S N = 5 −32 = 215
2 ×2 [5] Seki. H and Kanekko. T, “Differential Cryptanalysis of
where k=37,p=2-49,=25,=2-32 ,.The data complexity is CAST-256 Reduced to Nine Quad-rounds”, IEICE
52
2 chosen plaintexts. The time complexity is about Transactions on Fundamentals of Electronics Communi-
25×237×220=262 times of one-round encryption which is cations and Computer Sciences, E84A(4), pp.913-918,
equivalent to 258 times of 16-round encryption. The attack 2001.
is efficient for Blow-CAST-Fish with key sizes lager [6] J. Nakahara Jr and M. Rasmussen, “Linear Analysis of
than 58 bits. And the proportion of weak keys is 2-52. Reduced-round CAST-128 and CAST-256”,
SBSEG2007, pp.45-55, 2007.
5 Summary [7] E. Biham and A. Shamir, “Differential Cryptanalysis of
Until now, there is no analysis result for the algorithm. In DES-like Cryptosystems”, Journal of Cryptology, vol.4,
this paper, we first studied the properties of round function no.1, pp.3-72, Springer, 1991.
F1 and F3 with non-zero inputxor and zero outputxor in [8] M. Q.Wang, X. Y.Wang, C. H. Hu, New Linear
the 4th and 6th round of Blow-CAST-Fish separately, then Cryptanalytic Results of Reduced-Round of CAST-128
6-round differential characteristic with the probability 2-61 and CAST-256, appeared on SAC’2008.
-12
can be identified which are resulted by 2 of the total key [9] M. Q.Wang, X. Y.Wang, K.P. Chow, Lucas C.K. Hui,
space. Based on the 6-round differential characteristics, we Differential Crypanalysis of Reduced-Round of
can attack 8-round Blow-CAST-Fish. CAST-128, Preprint.
Then we studied the properties of S1 with non-zero
inputxor and zero outputxor in round function of Blow-
A Keyless Polyalphabetic Cipher
William J. Spezeski
Department of Computer Science
Massachusetts College of Liberal Arts
375 Church Street, North Adams MA 01247 USA
Phone: 413.662.5591
Email: [email protected]
Abstract-A polyalphabetic cipher is one in which the letters of the standard alphabet as shown in Figure1
multiple alphabets (monoalphabets) are used to encode below. To encode the word ‘CIPHER’, we embark on a
a plaintext message. Each letter of plaintext is encoded ‘search and count’ process.
by selecting one of the alphabets with which to
determine a substitution. A key word or phrase is often For illustration, we begin at position 0 (the letter ‘A’); any
used to select different alphabets to encrypt the pre-designated letter position will do.
plaintext. This paper considers three concepts that,
The first letter of plaintext is obtained by counting the
when combined, lead to a keyless polyalphabetic cipher.
number of letters that we pass over as we go (left to right)
INTRODUCTION – POLYPHABETIC CIPHERS from the starting position ‘A’ until we find the first
plaintext letter ‘C’. It is coded by the resulting count, in
There are many variations of polyalphabetic cipher types this case, 2.
with varying degrees of complexity. Most rely on a key or
keyword to form the substance of the ciphertext. Consider The second letter of plaintext is obtained by counting the
this very simple example: A polyalphabetic cipher with the number of letters that we pass over as we go from the first
key BLUEBIRD would encode its first letter of plaintext plaintext letter ‘C’ to the second plaintext letter ‘I’. The
using its B-alphabet, its second letter using its L-alphabet, second letter is coded by the second count 6. And so on. If
its third letter using its U-alphabet and so on. The ninth the search process takes us beyond letter25 (Z), we return to
letter thru the sixteenth letters of plaintext would reuse the letter0 (A) and continue the count from right to left. Thus
B,L,U,E,B,I,R,D-alphabets for encipherment, respectively. letter26 = letter0 in our circular count.
Every eight letters of plaintext would use the key for
encipherment until the plaintext message is exhausted. A B C D E F G H I J K L MN O P Q R S T U VWX Y Z
Both the sender and receiver would need seven distinct 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25
alphabets available designated by B, L, U, E, I, R, D, C 2
respectively. Some classic multiple-alphabet ciphers I 6
include the Vigenere Cipher [6], Gronsfeld Cipher [3], P 7
Porta Cipher [3] and the Enigma Cipher of WWII [4]. H 17
E 23
A KEYLESS POLYALPHABETIC CIPHER
R 13
The polyalphabetic cipher being considered here is the Fig. 1 The Search and Count Algorithm
byproduct of combining three unrelated concepts.
Consequently, we will look at the methodology of this Using this scheme, the plaintext word ‘CIPHER’ is encoded
cipher in three stages. as the substitution ciphertext 2 6 7 17 23 13. This is
the general idea that we shall use for substituting characters
A. A Monoalphabet Substitution Cipher
of plaintext with numbers to form the ciphertext. Note that
To illustrate the concept behind a keyless alphabetic cipher,
it can be made more obtuse by setting up patterns of
we first consider a simple alphabetic cipher which relies on
alternating clockwise and counter-clockwise counting.
529
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_92, © Springer Science+Business Media B.V. 2010
530 SPEZESKI
Some examples of various monoalphabets are shown on. Better yet, an arbitrary (random) number can determine
below. what occurrence of each letter of plaintext we will search
A1 = defghijklmnopqrstuvwxyzabc for in the polyalphabet. A computer can randomly generate
A2 = jlpawiqbctrzydskegfxhuonvm these integers or they can be pulled off-the-top of the
A3 = juliscaertvwxyzbdfghkmnopq sender’s head. The selection of these numbers results in
A4 = xzavoidbygerspcfhjklmnqtuw different versions of the encoded plaintext, but does not
A5 = fzbvkixaymeplsdhjorgnqcutw affect the decoding of the ciphertext. The original plaintext
and its corresponding decoded ciphertext are always the
B. A Substitution Cipher with a Polyalphabet same. For example, each ciphertext below
To develop our polyalphabet, we string several alphabets,
say {A1, A2, A3, A4, A5}, together to form the Plaintext: happy birthday
polyalphabet. The same basic assignment method of
converting plaintext to ciphertext shown above is used with Ciphertexts:
the polyalphabet. A starting position must be identified;
1. 36 75 16 55 10 57 57 35 64 30 16 78 52 52
then the ‘search and count’ process begins. If the
2. 9 48 16 28 37 30 3 62 10 3 43 51 79 25
polyalphabet is a string of length N, then after reaching
3. 63 75 43 55 37 57 3 8 64 57 16 78 25 79
letterN-1 in a count, the count continues with letter0 , letter1 ,
letter2 and so on; thus letterN = letter0 . is obtained by using the “Search and Count” algorithm with
a standard 26- letter alphabet plus a ‘space’. If we use five
27-character unsorted alphabets to construct a polyalphabet,
A1 A2 A3 A4 A5
some resulting ciphertexts for ‘happy birthday’ appear as
Fig. 2 – Polyalphabet Structure
follows:
The alphabets making up the polyalphabet need not be
1. 48 14 60 29 10 33 17 45 60 96 51 76 74 8
ordered or of the same length. The number of alphabets
2. 48 37 12 54 56 14 9 26 60 27 11 63 24 67
that constitute the polyalphabet may vary as well. Below is
3. 5 27 90 29 56 41 4 58 61 60 32 41 72 17
an example of a polyalphabet formed by concatenating
4. 74 86 56 17 81 69 4 58 31 43 79 41 79 60
alphabets A1, A2 and A3 above.
All four ciphertexts above decrypt back to the original
defghijklmnopqrstuvwxyzabcjlpawiqbctrzydskegfxhuonvm
plaintext. Notice that the last series of plaintext ciphers
juliscaertvwxyzbdfghkmnopq
consists of (with leading zeros) all 2-digit numbers with the
largest number being 96, even though there are 135
In some sense, this polyalphabet can be thought of as a characters in the polyalphabet that is being used. This is
large monoalphabet. Because its construction (as I use it) being controlled by the fact that the random number choices
makes use of several existing monoalphabets , I use the are restricted to 1, 2, or 3. This helps to disguise the
term polyalphabet to describe it. number of ‘alphabets’ that form the polyalphabet as well as
the length of the polyalphabet.
C. A Keyless Polyalphabetic Cipher
Most polyalphabetic ciphers require a key - a word or DECODING THE CIPHERTEXT
phrase that distinguishes which particular monoalphabet a
Deciphering the coded message is very easy. One starts at
particular plaintext letter is being enciphered with. The
the pre-designated letter position of the same polyalphabet
classic Vignere Cipher [5] is a case in point. However,
and simply counts, associating the final number of each
using the ‘search and count’ process mentioned above,
count with the associated plaintext letter. Figure 1 above
there is no need for a key.
illustrates both how the word CIPHER can be encoded and
The ‘search and count’ process essentially looks for the how the ciphered word can be decoded. In the encipherment
first occurrence of the plaintext letter as it traverses the process, searching for a plaintext letter results in a count ,
polyalphabet. A match results in a ciphertext number. which in turn becomes the ciphered substitute; in the
However, there is no need to look for only first occurrences decipherment process, the count becomes the means of
of a plaintext letter. We could search for the second locating the corresponding plaintext letter. As mentioned
occurrence of a plaintext letter, a third occurrence, and so
A KEYLESS POLYALPHABETIC CIPHER 531
before, an enciphered text can have many possible variants. deciphering algorithms in a computer language like Java is
However, each variant will decipher to the same plaintext. a straight-forward process using String arrays. Here are
some things to keep in mind:
SOME OBSERVATIONS
The number of alphabets used to form the polyalphabet
Let us suppose that a five letter word is to be encoded with
should be a prime number. This is to ensure that the
a 100-character polyalphabet using the random search
searches do not get confined to any cyclic subset of
‘modifiers’ 1,2 and 3. Then there will be 100 possible
alphabets comprising the polyalphabet.
starting positions and
100 x 3 x 32 x 33 x 34 x 35 or 100 x 315 The number of distinct random numbers that get generated
should be strictly less than the number of alphabets making
possible variants of the enciphered five letter word. More up the polyalphabets. For example, assume that a particular
generally, if a plaintext message consisting of N characters polyalphabet consisted of 5 alphabets each having 25
(including spaces) is encoded with a P-character characters. Assume further, that the distinct random
polyalphabet using K random search modifiers, then the numbers that get generated are 1, 2, 3. The largest possible
number of possible enciphered variants is numeric substitution for a plaintext letter would be 100,
even though there are actually 125 characters in the
P KN(N + 1)/ 2 . polyalphabet. This helps to conceal the actual string length
of the polyalphabet. In fact, the 5 alphabets mentioned can
If a 135-character polyalphabet to code ‘happy birthday’
be rearranged so that only1- or 2-digit numbers will be
using 3 random search modifiers, the number of variants
generated for substitution. If the ciphertext is converted to
would be 135 x 3105 . Four of the variants were shown
ASCII characters , 3-digit numbers (< 256) can be
above. For any fixed starting point, the number of variant
accommodated without the ciphertext getting unwieldy.
ciphertexts for ‘happy birthday’ is 3105 .
IMPLEMENTATION Use the Java string method length() (or equivalent) to make
any reference to the length of the active polyalphabet. It
The implementation of the Search and Count Cipher (SCC) helps to further conceal its actual string length.
requires that the both the sender and receiver use the same
polyalphabet as well as an agreed upon starting position The alphabets that comprise the polyalphabet can vary in
within the polyalphabet. There are 26! (over 1026 ) regular number, length, letter arrangement and order of appearance
26-letter monoalphabets available for building a within the polyalphabet. The details, of course, must be
polyalphabet. That number grows rapidly with an increase agreed upon in advance by both sender and receiver.
in alphabet size. Below is the ciphertext that is generated from a
polyalphabet formed by the 5 alphabets above (A1, A2, A3,
Polyalphabets can be easily constructed by having a A4, A5 each with an extra ‘space’) using modifiers 1, 2, 3
collection of monoalphabets available for concatenation. and starting at the first position of alphabet A1.
The computer language Java, for example, provides string
concatenation (+) that allows a wide variety of 74 52 17 78 38 44 73 25 34 41 29 5 14 57 48 77 67 26 18
polyalphabets to be built very simply from existing 61 79 64 13 25 39 1 20 75 23 30 41 11 3 29 14 72 78 30 4
monoalphabets. Again, this is a primary reason for 78 20 8 63 56 32 42 78 40 79 64 68 36 52 23 47 23 19 55
referring to the cipher as a polyalphabetic cipher. In turn, 76 34 15 36 63 79 23 16 27 26 41 3 59 39 72 15 9 64 39 8
multiple alphabets can be constructed from a single one by 21 50 13 2 68 28 56 38 30 57 4 80 80 21 80 68 23 9 25 23
using a shuffle algorithm to rearrange an alphabet’s letters. 12 4 2 63 51 63 37 12 8 21 40 51 46 41 15
If both sender and receiver have the same set of numerically
identifiable alphabets at hand, the sequence of alphabets It is the ciphertext for the verse
used to construct the operable polyalphabet can be encoded
speak roughly to your little boy
within the numerical ciphertext. The starting position can
and beat him when he sneezes
be part of the ciphertext as well. And like the polyalphabet
he only does it to annoy
construction, the plaintext message string can be built by
because he knows it teases
using concatenation. Coding the enciphering and
532 SPEZESKI
The SCC produces random-like sequences of numbers for Will the SCC survive the test of cryptanalysis? That
ciphertext that help to diffuse the process of cryptanalysis. question remains.
The randomness of the ciphering algorithm reduces the REFERENCES
recognizable statistical patterns that a cryptanalyst might
use to a minimum. The lack of a key further frustrates [1] Baldoni, Ciliberto & Cattaneo, Elementary Number Theory,
Cryptography and Codes, Springer-Verlag, (2009).
cryptanalysis efforts.
[2] Friedman, William F., Military Cryptanalysis - Part II: Simpler
Varieties of Polyalphabetic Substitution Systems, Filiquarian Publishing,
In the previously encoded verse above, there are 108 LCC./Qontro (1938).
integers in the ciphertext whose values fall in the interval [3] Gaines, Helen F., Cryptanalysis, Dover Publications, (1956)
[4] Kahn, David, The Code-breakers, Macmillan Publishing, (1979)
[1, 80]. Figure 3 illustrates the frequency counts of numeric
[5] Shannon, Claude, “Communication Theory of Secrecy Systems”, Bell
ciphertext in the interval [1, 80] and compares it with 108 System Technical Journal, (1949).
randomly generated numbers in the same interval. [6] Singh, Simon, The Code Book, Fourth Estate Ltd., (1999).
[7] Sinkov, Abraham, Elementary Cryptanalysis,2nd edition, The
Mathematical Association of America, (2009).
Vertical Handover between WiFi Network and
WiMAX Network According to IEEE 802.21 Standard
Mariem Thaalbi, Nabil Tabbane
Research Unit MEDIATRON
SUP’COM
Tunis, Tunisia
[email protected] [email protected]
Abstract—IEEE 802.21 is one of the emerged standards designed BSSs connected to the same wired LAN so called Distribution
to build handover between heterogeneous technologies while System (DS).
optimizing session continuity. In this paper we provide a
performance evaluation of the vertical handover between WiFi In ad hoc mode, devices can communicate directly with
and WiMAX using IEEE802.21 for different traffic flows: VoIP, each other without Access Point support.
video streaming and FTP. The performance metrics used for this In our paper, we will focus on infrastructure mode.
evaluation include throughput, end-to-end delay and packet loss
rate. The performance evaluation is done using the Network
Simulator NS2.
I. INTRODUCTION
The rapid expansion of heterogeneous wireless and mobile
networks raises interesting issues such as vertical handover. To
insure vertical handoff between WiFi and WiMAX networks,
we will propose, in this paper, the IEEE 802.21 as an alternative
to maintain uninterrupted user sessions during handovers
between these two different networks. The IEEE 802.21
supports media Independent handover (MIH) in order to
provide seamless mobility between heterogeneous networks.
The purpose of this work is to evaluate the performances of
IEEE 802.21 in vertical handover between WiFi and WiMAX
technologies. The remainder of this paper is organized as
follows: The background of WiFi and WiMAX are briefly
viewed on Section II and III. An overview of IEEE 802.21 is
given in Section IV. Simulation Scenario and experimental
Fig.1. WiFi infrastructure mode.
results are presented in Section V. The paper is concluded in
Section VI. III. WIMAX NETWORKS
Worldwide Interoperability for Microwave Access
II. WIFI NETWORKS (WiMAX) [2] is a Wireless Metropolitan Access Network
Wireless Fidelity (WiFi) [1] is a wireless local area network (WMAN). It is used generically to describe wireless systems
over limited range. It is also known under IEEE 802.11, and based on IEEE 802.16.
aims to connect devices such as Personal Computer, PDA,
WiMAX offers high data rates and large area coverage. It
laptop, printers, etc.
supports fixed and mobile broadband access. In our work, we
IEEE 802.11 allows connectivity in infrastructure mode or will focus on Mobile WiMAX (IEEE 802.16e) [3].
ad hoc mode. In infrastructure mode, the WiFi stations
The Mobile WiMAX Network, described in Fig.2, is
communicate through the Access Point (AP). In this mode, the
divided into three parts: the Mobile Station (MS) which is the
802.11 network architecture, illustrated in Fig.1, is hierarchical.
user’s device, the Access Service Network (ASN) is the radio
Its basic element is the Basic Service Set (BSS), which is the
access and includes one or more Base Stations (BS) and one or
group of stations covered on the area covered by one AP. A
more ASN Gateways (ASN-GW). The Connectivity Service
BSS may also be part of larger network element called
Network (CSN) is the core of WiMAX Network. It provides IP
Extended Service Set (ESS). The ESS consists of one or more
connectivity to the MS.
533
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_93, © Springer Science+Business Media B.V. 2010
534 THAALBI AND TABBANE
In order to provide mobility, 802.16e supports handover Commands originate from the upper layers down to the MIHF.
mechanisms. The IEEE 802.16e defines three different types of Link Commands are specific to the lower layers. MIIS provides
handover: Hard Handover (HHO), Fast Base Station Switching a framework by which MIHF can discover homogenous and
(FBSS) and Macro Diversity Handover (MDHO). The HHO is heterogeneous network information existing within a
often referred to as a break-before-make handover: first the MS geographical area to facilitate seamless handover when roaming
disconnects from the serving BS and then connects to the target across these networks. The MIIS provides a bidirectional way
BS. In FBSS and MDHO, the MS maintains a diversity set for the two layers to share information such as current QoS,
which includes numerous active BSs in its range. The BS, to performance information and availability of service. The figure
which the MS has the connection, is called the anchor BS. In bellow illustrates the MIH architecture:
FBSS, the MS transmits to and receives data from a single
serving BS. The transition from the serving anchor BS to the
target anchor BS in FBSS is done without invocation of the
normal handover procedure, only the anchor BS update
procedure is needed. In MDHO, the MS maintains a
connection to one or more BSs in the diversity set
simultaneously.
End-to-end delay
The end to end delay refers to the time taken by data packet
to be transmitted across a network from source to destination.
We calculate in this section the end-to-end delay with the same
cases already described. The average delay for different
simulated traffic is shown in Fig.6 (a) - (b) - (c).
When the MS speed is equal to 10 m/s, the delays degrade (b) End-to-end delay (ms) - Video streaming traffic
in WIFI network because it supports only the low mobility
speed. The curves obtained with VoIP and FTP traffic during
the handover process exceed slightly the threshold which is
fixed to 100 ms for the VoIP and FTP traffics [8]. The curves
obtained with streaming video traffic exceed very largely the
threshold which is fixed to 100 ms [8].
We note that, with the MS speed (10 m/s) the delays
offered by WIMAX network are better than those offered by
WIFI network because the mobile WIMAX network can
support the low and the high mobility speed.
Packet loss rate In this figure, we see that with MS speed equal to 1 m/s, the
two handover cases curves reflect good results, unlike the
The packet loss rate parameter allows to determine the lost medium mobility speed equal to 10 m/s the handover curves
packet rate compared to those sent. The percentage of packet exceed the threshold (fixed to 1%) for the VoIP and video
loss for different simulated traffics is illustrated in Fig.7 (a) - streaming traffics [8].
(b) - (c). With low mobility, the packet loss rate is less in WIFI
network and the better results are obtained with the case when
the MS makes a handover from WIMAX to WIFI. With MS
speed equal to 10 m/s, the results are better in WIMAX
network and the handover case from WIFI to WIMAX gives
better packet loss rate values than the opposite case.
VI. CONCLUSION
The interoperability and the vertical handover between
different networks present currently a real challenge to
overcome.
In this paper, we have evaluated the performance of
vertical handover between WiMAX and WiFi according to
802.21 for FTP and real-time traffic such as VoIP and video
streaming. Seeing the results obtained, we can notice that the
speed of the MS affects the handover performance. Also, the
(a) Packet loss rate (%) - VoIP traffic WiFi supports only the lox mobility speed. Finally, we can
conclude that with the low speed the handover from WiMAX
to WiFi generate best results than the opposite case of
handover. But with medium speed, it is the otherwise because
the 802.16e supports better the mobility management.
REFERENCES
[1] IEEE Std 802.11™-2007, IEEE Standard for Wireless LAN Medium
Access Control (MAC) and Physical Layer (PHY) Specifications, June
2007.
[2] IEEE Std 802.16™-2004, IEEE Standard for Local and Metropolitan
Area Networks—Part 16: Air Interface for Fixed Broadband Wireless
Access Systems, October 2004.
[3] IEEE Std 802.16e™-2005, IEEE Standard for Local and Metropolitan
(b) Packet loss rate (%) - Video streaming traffic Area Networks—Part 16: Air Interface for Fixed and Mobile Broadband
Wireless Access Systems, February 2006.
[4] IEEE Std 802.21™-2008, IEEE Standard for Local and Metropolitan
Area Networks- Part 21: Media Independent Handover, January 2009.
[5] Network Simulator- NS2- http://www.isi.edu/nsnam/ns, February 2009.
[6] National Institute of Standards and Technology (NIST), “The Network
Simulator NS-2 NIST mobility package add-on,”, March 2009.
[7] National Institute of Standards and Technology (NIST), “The Network
Simulator NS-2 NIST IEEE 802.16 Implementation,”, March 2009.
[8] Jeffery G. Andrews, Arunabha Gosh, Rias Muhamed, “Fundamentals of
WiMAX, 1.7.3 Quality of Service”, June 2007.
Abstract – In this paper we analyze the total energy energy and present the setting of the energy model. We
consumption of a wireless sensor network (WSN) per single consider a sensor network that is hierarchically clustered.
communication round as a function of the node density and base The LEACH (Low Energy Adaptive Clustering Hierarchy)
station coordinates. We provide performance results of several protocol maintains such clustering hierarchy [2, 3]. In LEACH,
simulations with WSN using both single hop and multi hop
mechanisms for data transfer, and than discuss the relationships
the clusters are re-established in each round. New cluster heads
between the node density, the base station location and the total are elected in each round and as a result the load is well
energy dissipation per round. distributed and balanced among the nodes of the network.
Keywords – WSN, Energy dissipation, Cluster, Cluster Head Moreover each node transmits to the closest cluster head so as
to split the communication cost to the sink [4]. Only the cluster
I. INTRODUCTION head has to report to the base station and may expend a large
amount of energy, but this happens periodically for each node.
Sensor networks play a major role in many aspects of In LEACH there are an optimal predetermined percentage
society including home automation, consumer electronics, p opt of nodes that has to become cluster heads in each round
military application, agriculture, environmental monitoring, assuming uniform distribution of nodes in space. Since the
health monitoring and geophysical measurement. Usually nodes are homogeneous, which means that all the nodes in the
sensor devices are small and inexpensive, so they can be field have the same initial energy, the LEACH protocol
produced and deployed in large numbers. Their resources of guarantees that everyone of them will become a cluster head
energy, memory, computational speed and bandwidth are exactly once every 1 / popt rounds. Throughout this paper we
severely constrained. Therefore, it is important to design refer to this number of rounds as epoch of the clustered sensor
sensor networks aiming to maximize their life expectancy. network. Initially each node can become a cluster head with a
Different aspects of sensor networks such as data aggregation probability popt [5]. On average, n × p opt nodes must become
or fusion, packet size optimization, target localization, design
challenges, network protocols are discussed in the literature cluster heads per round per epoch. Nodes that are elected to be
with respect to crucial energy limitations and network lifetime cluster heads in the current round can no longer become cluster
maximization [1]. heads in the same epoch. The non-elected nodes belong to the
This paper presents a study focused on the total energy set G and in order to maintain a steady number of cluster heads
dissipation of a sensor network per a communication round. per round, the probability of nodes ∈ G to become a cluster
For this purpose we have simulated a WSN with a various head increases after each round in the same epoch. The
number of sensors and with a base station, which coordinates decision is made at the beginning of each round by each node
are first at the center of the sensor field, than at the edge of the s ∈ G independently choosing a random number in [0,1]. If the
field and finally outside of the field. random number is less than a threshold T (s ) then the node
The rest of the paper is organized as follows. In Section 2 becomes a cluster head in the current round. The threshold is
we have provided the communication model, which is being set as:
used, together with the energy model. Section 3 presents the popt
performance measures used in this paper. Section 4 provides ° if s ∈G
° § ·
the simulation scenarios and the various settings we use for the T (s ) = ®1 − popt ¨ r mod 1 ¸ (1)
simulations. In Section 5 we provide the results of the ° ¨ popt ¸¹
©
conducted simulations as well as a discussion on the obtained ° 0 if s∉G
¯
results. Section 6 contains the conclusions on the conducted
studies and directions for future work. where r is the current round number. The election probability
of nodes ∈ G to become cluster heads increases in each round
II. MODEL OVERVIEW in the same epoch and becomes equal to one in the last round
of the epoch. Note that by round we define a time interval
In this section we describe our model of a wireless sensor where all clusters members have to transmit to the cluster head
network with nodes, homogenous in their initial amount of once. For the purpose of our study we use the energy model
presented in Figure 1.
539
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_94, © Springer Science+Business Media B.V. 2010
540 ZAHARIEV ET AL.
n−k k
ETotal = ¦ EiSensor + ¦ ECH
j (5)
1
BS
i =1 j =1
0.9
CHn 0.7
Multi hop
CH2 0.6
0.5
0.4
BS
0.3
MH
CH1 N 0.2
2
0.1 Single hop
BS 0
0 100 200 300 400 500 600 700 800 900 1000
1
Number of nodes, n
Fig.3 Multi hop wireless sensor network Fig.4 Total energy dissipation of the network, when the base station is in the
center of the network
1.8
V. SIMULATION RESULTS
Total energy dissipation per round, E Total, J
1.6
1.4
On Figure 4 we present the results of the simulation, whit 1.2 Multi hop
the base station in the center of the sensor network 1
( x = 50 , y = 50 ). It can be clearly seen that when the node 0.8
density is relatively small ( n < 100 ), the total energy 0.6 Single hop
consumption of the network for both multi hop and single hop 0.4
scenarios is almost the same. This is due to the fact that the 0.2
distance of every node in the field is smaller than d 0 and due
0
to the fact that the number of clusters is relatively small, which 0 100 200 300 400 500 600
Number of nodes, n
700 800 900 1000
Base station x=50, y=200 transmit to the base station compared to the single hop
2.4 scenario, where all do. Compared to the results from the
2.2
previous simulation here even the cancellation effect of the
Total energy dissipation per round, E Total, J many hops that the data travels from the edge cluster heads to
2
1.8
1.6
ones closer to the base station shows little effect, due to the
1.4 Multi hop larger distance to the base station.
1.2
1
Single hop VI. CONCLUSIONS AND FURTHER WORK
0.8
0.4
0.2
node density and base station location in wireless sensor
0 networks. From the obtained simulation results it is clear to
notice that with the relocation of the base station away from
0 100 200 300 400 500 600 700 800 900 1000
Number of nodes, n
Abstract-Families of generalized orthogonal complementary A promising candidate for employment in the future wireless
pairs (FGOCPs) are called families of pairs of radio signals, which communication systems are the so – named Families of
autocorrelation functions (ACFs) and cross-correlation functions generalized orthogonal complementary pairs (FGOCPs). They
(CCFs) among every two pairs of a family have so-called zero are families of pairs of radio signals, which ACFs and CCFs
correlation windows (ZCWs) in the area around the zero time shift.
They allow the users to communicate practically without multiple
among every two pairs of a family have so-called zero correlation
access interferences in a multipath environment if the wireless windows (ZCWs) in the area around the zero time shift [4], [5],
communication systems works in a quasisynchronous mode. As a [6]. They allow the users to communicate practically without
result the employment of the FGOCPs in the future wireless MAI in a multipath environment if the wireless communication
communication systems will significantly increase the data systems work in a quasisynchronous mode. It should be
transmission rate. Due to this reason in the paper an algorithm for mentioned that the so – named families of orthogonal
synthesis of FGOCPs is proposed. It distinguishes by the fact that it complementary codes (FOCCs) provides possibility the users to
provides small level of the lobes of the CCFs without of the ZCW, communicate practically without any SI and MAI in a truly
which is a valuable feature. asynchronous mode [1], [4]. Anyway, in a comparison with
Index Terms – Synthesis of signals, Families of quasiorthogonal FOCCs valuable advantages of the FGOCPs are the simplicity of
complementary pairs. their technical realization and the higher resistance to the
frequency selective fading.
I. INTRODUCTION
With regard to above described situation, our paper aims to
Today users and industry expect a significant enhancement of develop an algorithm for synthesis of FGOCPs, which
the quality of the services and the data transmission rate, offered distinguishes by the fact that it provides small level of the lobes
by the wireless communication systems. This is a very hard of the CCFs without of the ZCW. The importance of the studied
technical challenge for the theorists and developers. From the problem results from the following facts. First, today it is
information in the open literature [4], [5], [9] it can be concluded possible to provide a quasisynchronous mode of operation of the
that a critical role for the whole performance of the wireless wireless communication systems by a real time measurement of
communication systems has the system of employed radio the location of the users relatively to the base stations of the cells.
signals. Due to this reason great efforts are directed to finding of Second, despite of taken efforts many problems in the field of the
families of radio signals possessing both autocorrelation synthesis of the FGOCPs are open still [4], [5], [6].
functions (ACFs) with a small level of the side-lobes and small Paper is organized as follows. First, the basics of the FGOCPs
cross-correlation functions (CCFs) among all pairs of members are recalled. Second, our algorithm for synthesis of FGOCPs is
of a family. The focusing of the research activities on this suggested. At the end some important conclusions are given.
problem can be explained by the following facts [1], [2], [4], [5],
[6], [9], [10], [11], [13], [14], [15], [16], [17]. First, the pointed II. BASICS OF THE SYNTHESIS OF THE FAMILIES OF
out families of radio signals allow the so-named self-interference GENERALIZED ORTHOGONAL COMPLEMENTARY PAIRS
(SI), caused by multipath spreading of electromagnetic waves, to Let K ≥ 2, N ≥ 1, be integers and let us consider a family
be reduced by a separate processing of the direct and reflected W(K, 2, N, Z0), comprising K pairs as every pair consists of two
signals. Second, it is possible the negative effect of simultaneous sequences with (code) length N:
transmission of numerous users, named multi user interference
(MAI), to be minimized.
543
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_95, © Springer Science+Business Media B.V. 2010
544 BEDZHEV AND ILIEV
ª K −1 ª K −1 º where s is a divisor of Q.
k .N º
« ¦ d (m, k ).F ( x).x » « ¦ d * (l , j ).F * ( x ) x
−1 − j . N
»= It should be mentioned that in the open literature detailed
¬k =0 ¼ ¬« j = 0 ¼» tables with complementary pairs and orthogonal matrices are
given [4], [8].
ª K −1 º ª K −1 º
= F ( x).F * ( x −1 ) « ¦ d (m, k ).x k . N » « ¦ d * (l , j ).x − j. N » (19) 3) Form the tensor product E, according to (13).
¬k =0 ¼ ¬« j = 0 ¼» The rows of E are a FGOCP with parameters: family size – K,
length of the sequences – N’ = K.N, width of the ZCW – Z0 = N.
= 2 N .H m ( x N ).H l* ( x − N ) = 2.N .PH m , H l ( x N ).
IV. RESULTS
Equation (19) proves the Proposition 3.1, because the It should be stressed that the above Algorithm 3.2 gives the
coefficients of the polynomial in the right side of (19) satisfy the valuable opportunity to minimize correlation lobes without of the
following restrictions ZCW of the synthesized FGOCPs by an appropriate selection of
the matrix D. With regard to the known upper bounds of the level
of the lobes of the CCFs of the binary sequences [12], an
° 0, s ≠ r.N ;
2.N .RH m , H l ( s) = ® exhaustive examination of the correlation properties of the square
°̄2.N .RH m , H l (r ), s = r.N , (20) orthogonal matrices have been conducted. As a result we have
s = −( K − 1).N , − ( K − 1).N + 1, ..., − 1, 0,1, ..., ( K − 1).N found that the most appropriate matrices D are these, which are
generated by the so – named Frank’s method [2]. This conclusion
and issues from the following propositions.
Proposition 4.1: Let g be a K-th primitive root of the unity:
K , r = 0 ∩ l = m;
RH m , H l ( r ) = ® (21) g = exp[(2πil ) / K ], i = − 1 (25)
¯ 0, r = 0 ∩ l ≠ m,
where l and K are relatively prime numbers.
as the rows of the matrix D are mutually orthogonal.
AN ALGORITHM FOR SYNTHESIS OF FAMILIES 547
Consequently V. CONCLUSION
In the paper an algorithm for synthesis of FGOCPs is
K suggested on the base of Proposition 3.1. It allows the users to
r≤ . (30) communicate practically without MAI in a multipath
2
environment if the wireless communication systems work in a
From (27) and (28) follows: quasisynchronous mode. It should be stressed that a
quasisynchronous mode can be provided by a measurement of
the locations of the users respectively to the base station and
− RCG n ,CGl (r ) = RCGn ,CGl ( K − r ) . (31)
introducing a time schedule for sending messages to the base
station. This is not a hard technical task and as a result it is
The magnitude of the right side of (31) can be estimated from possible to exploit quasisynchronous mode in the next
the chain of inequalities: generations wireless communication systems.
The FOCCs provides possibility the users to communicate
practically without any SI and MAI in a truly asynchronous mode
548 BEDZHEV AND ILIEV
[1], [4], [5], [6]. Anyway, in a comparison with FOCCs valuable REFERENCES
advantages of the FGOCPs are the simplicity of their technical [1] B. Y. Bedzhev and M. P. Iliev, “A general method for synthesis of families of
realization and the higher resistance to the frequency selective orthogonal complementary codes,” unpublished
fading. [2] J. J. Benedetto and J. J. Donatelli, “Ambiguity function and frame-theoretic
properties of periodic zero-autocorrelation waveforms,” IEEE J.of Selected
The algorithm for synthesis of FGOCPs, suggested in the Topics in Signal Processing, vol.1, No.1, pp. 6-20, June 2007.
paper, has the following positive features. [3] R .E. Blahut, Theory and Practice of Error Control Coding, Addison-Wesley,
First, it gives a general approach to the synthesis of the Reading, Mass., USA, 1987
FGOCPs. Namely it is a direct construction for generating [4] H.-H. Chen, Next Generation CDMA Technologies, John Wiley & Sons,
Chichester, West Sussex, England, 2007
FGOCPs, providing both the desirable number of users K and the [5] P. Fan, “Spreading sequence design and theoretical limits for
width Z0 of the ZCW. quasisynchronous CDMA systems,” EURASIP J. on Wireless
Second, in contrast with large number of constructions [5], [6] Communications and Networking, pp. 19 – 31, no. 1, 2004
it is developed on the base of the aperiodic correlations which is [6] L. Feng, P. Fan, X. Tang, and K.-K. Loo, “Generalized pairwise
Z-complementary codes,” IEEE Signal Processing Letters, vol. 15, pp. 377-
the typical case for the present wireless communication systems. 380, 2008
Third, it allows minimizing of the correlation lobes without of [7] M. J. E. Golay, “Multislit spectroscopy,” J. Opt. Soc. Amer., vol. 39, pp. 437 –444,
the ZCW of the synthesized FGOCPs by an appropriate selection 1949
of the matrix D. For example, if we use matrix D = G in [8] M. J. E. Golay, “Complementary series,” IRE Trans. Inform. Theory, vol. 7,
pp. 82–87, £2, 1961.
Algorithm 3.2, as G is generated by the so – named Frank’s [9] L. Hanzo, M. Munster, B. J. Choi, and T. Keller, OFDM and MC-CDMA for
method [2] (i.e. by (26)), then the maximum side – lobes do not Broadband Multi-User Communications, WLANs and Broadcasting, John
Wiley & Sons and IEEE Press, New York, NY, USA, 2003
exceed K .N . Here K is the size of the matrix D, determining
[10] S. P. Maity, M. Mukherjee, “A high capacity CI/MC-CDMA system for
the number of the users of the developed wireless communication reduction in PAPR,” in Proc. Seventh Annual Communication Networks and
system, N is the length of the used pair of complementary Services Research Conference, pp.297-304, Moncton, New Brunswick,
sequences and K.N is the amplitude of the central peak of the Canada, 11-13 May, 2009
[11] A. Milev and V. Mutkov, “Research over correlation properties of families
aperiodic ACFs of the FGOCP, synthesized by the optical orthogonal codes,” in Proc. International Conference on
Algorithm 3.2. CompSysTech’09, pp. III.20.1- III.20.6, Ruse, Bulgaria, EU, June 2009
With regard to all above stated we hope the algorithm for [12] D. Y. Peng and P. Z. Fan, “Generalized Sarvate bounds on periodic
synthesis of FGOCPs, suggested in the paper, to be useful in the autocorrelations and crosscorelations of binary sequences,” Electronics
Letters, vol. 38, no. 24, pp. 1521 – 1523, 2002
process of development of the new generation wireless [13] T. R. Qureshi, M. D. Zoltowski and R. Calderbank, “A MIMO-OFDM
communication system, which aim to provide a radical channel estimation scheme utilizing complementary sequences,” in Proc.
enhancement of the data transmission rate. The 2009 IEEE International Conference on Acoustics, Speech and Signal
Processing, pp.2677-2680, Taipei, Taiwan, 19-24 April, 2009
ACKNOWLEDGMENT [14] R. G. Tsakov, V. A. Mutkov, N. R. Nikolov and L. . Staneva, “A survey of
The authors wish to thank Prof. Dr Kunchev, Prof. E. the correlation properties of the generalized Barker codes”, unpublished
[15] R. G. Tsakov, V. A. Mutkov, N. R. Nikolov and L. . Staneva, “An
Ferdinandov, Prof. Dr H. Hristov and Prof. Dr B. Tsankov. improved method for synthesis of families of Costas arrays”, unpublished
Finally, the authors appreciate the thoughtful critique by the [16] C.-C. Tseng and C. L. Liu, “Complementary set of sequences,” IEEE Trans.
referees. Inf. Theory, vol. 18, pp. 206–209, Sept. 1972.
[17] M. D. Zoltowski, T. R. Qureshi and R. Calderbank, “Complementary codes
based channel estimation for MIMO-OFDM systems,” in Proc. Forty-Sixth
Annual Allerton Conference, pp. 133-138, Allerton House, UIUC, Illinois,
USA, 23-26 Sept., 2008
Low Message Overhead Clustering Algorithm
for Wireless Sensor Networks
Batista, Othon Marcelo Nunes; Giozza, William Ferreira
Abstract—A Wireless Sensor Network (WSN) is defined by a deaths occur from the farthest to the nearest ones. Direct
set of sensor nodes and a Base Station (BS). Typically, sensor diffusion forms a plain topology.
nodes are distributed over a geographical area, sense some Another way to send information to the BS is via multi hop
information, process it and send it to the BS via a wireless link.
routing. Each sensor must configure its transmission power to
Since sensor nodes are small, may be distributed in large scale or
in a dangerous area, their battery are small and may not be reach one or more neighbor sensor node as close as possible.
recharged or replaced. So the network lifetime is prolonged when The sensed information is sent to the BS via intermediate
the battery energy is used wisely. The function of the BS is to sensor nodes, in a multi hop way. The energy spent to send an
receive information sensed by the nodes, therefore it's a sink. information to the BS is the summation of the energies spent
One way to send information is to cluster the WSN, dividing the to send the information to the sensor nodes in a route toward
sensor nodes in two exclusive sets: Cluster Heads (CH) or Cluster
the BS. If a sensor node is very near to the BS, it spends more
Members (CM). Therefore, the clustered WSN presents two
types of communication: one from CM to CH, inter-cluster energy than the others, since it aggregates informations from
communication, and other from CH to BS, intra-cluster all other nodes. The sensor nodes deaths occur from the
communication. Once a WSN is clustered, energy consumption nearest to the farthest ones.
and coverage are improved. The first step in clustering a WSN is Coverage is the degree of the geographical area in which
a CH election algorithm. We propose a distributed clustering the sensor nodes were distributed that are in fact being sensed.
algorithm for WSN named Low Message Overhead Clustering
As the sensor nodes die, coverage is gradually lost. When the
Algorithm (LMOCA) that has the goal to spend as minimum
energy as possible on the clustering process. Beyond that it is information is sent in one hop or multi hop, the coverage is
very intuitive and easy to implement. lost really fast, only varying the direction, from the farthest
sensor node to the base station or from the nearest sensor node
Index Terms— Wireless Sensor Networks, Clustering to the base station, respectively.
Algorithm, Simulation Results. Another way to send information to the BS is by clustering
the WSN. Clustering a WSN means to divide it in two sets of
nodes: Cluster Heads (CH) and Cluster Members (CM). This
I. INTRODUCTION division happens also in the communication level: one
549
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_96, © Springer Science+Business Media B.V. 2010
550 BATISTA ET AL.
The algorithm is divided in five phases: on to get the node degree messages.
phase 1: node degree calculation; Received node degrees are stored in an array in a position
phase 2: node degree exchange; correspondent to the node identification number in the
phase 3: initial clusters; neighbors array. At the end of this phase, all sensor nodes
phase 4: network rearrangement; know their neighbors and neighbour degrees.
phase 5: cluster member transmission range adaptation.
C. Initial Clusters
For testing purposes, the phases were configured to run for
5 seconds each. In the future, we intend to find the best time The initial clusters are formed by nodes with higher degree.
that suits the protocol. Since all sensor nodes are initialized as CM, they must choose
This algorithm was simulated in Castalia, a discrete event a node as a CH, if one exists. Each sensor node, checks its
WSN [16] simulator based in OMNET++ [17]. For a better node degree array and sends a message to the highest degree
explanation, one example is presented in a step-by-step basis sensor node. This message is called DOMINATE ME.
with the aid of the WSN graph shown in Fig. 1. This graph Once a sensor node receives a DOMINATE ME message, it
represents 100 sensor nodes in an area of 900m2 (30m x 30m). changes its status to CH and stores sensor node identification
Sensor nodes are white circles with a number (sensor node number in an array of CM sensor nodes.
identification number). The red circle in coordinate (0, 0) is At the end of this phase if a node neither generate or
the BS. receive a DOMINATE ME message, it self elects CH. This
Sensor nodes have a initial transmission power of 0 dBm. guarantees that, all nodes are either CM or CH at the end of
This transmission power applied to Friis and Two Ray Ground the clustering phase.
propagation models lead to a range of approximately 8.59m. The algorithm could stop after this phase, however, as Fig.
This range indicates the distance until the transmission power 2 shows, the clusters are not well distributed in the
is reduced to the receiver sensitivity of -98 dBm. Also, it geographical area, what will raise the interferences and
indicates the radius of each cluster. collisions between clusters. So, there's a needing to reorganize
The five phases describing the operation of LMOCA are clusters to minimize interference and maximize energy usage.
presented as follows. Notice that cluster heads are the red triangles and cluster
members are blue circles.
clusters, which must aid in the future inter and intra-cluster Hence, message complexity for the whole network in the
communications, regarding that the interferences decrease worst case is O(n). n is the total number of nodes in the
with the transmission power adjustment (Fig. 3). network.
80
Therefore, message complexity for each sensor node is
O(1). 70
O(n). 44,63
50
nodes send three messages and CH nodes send four messages. 22,01
20,62
0
2,04 3,62 4,83 5,74 6,44 7,23 8,11 8,59 11,46 15,28 48,31
50 1,2
45
43,59
1 1 1 1 1 1 1 1 1 1 1
1
40
35
0,8
25 0,6
23,19
19,98
20
16,69
0,4
15 14,49
13,87
13,51
10 9,38
0,2
4,54
5
2,55
0 0
2,04 3,62 4,83 5,74 6,44 7,23 8,11 8,59 11,46 15,28 48,31 2,04 3,62 4,83 5,74 6,44 7,23 8,11 8,59 11,46 15,28 48,31
Graph 2. Quantity of CH after 100 simulations for 50 sensor nodes. Graph 4. Quantity of CH after 100 simulations for 2 sensor nodes.
The results of the third simulated scenario are presented in Although not graphically shown in this paper, the scenario
Graph 3. In this scenario, 10 sensor nodes were used, each with only one sensor node were simulated. All the times the
simulation in a aleatory different position. The quantity of CH quantity of CH were 1. This also shows that LMOCA works
is inversely proportional to transmission power. for this special situation.
VII. CONCLUSION
10
saves energy. Some lemmas were proved for the LMOCA and
7
6,64
6
LMOCA Average
Cluster Quantity
simulation results were presented. In these results, we
observed that the quantity of CH depends on the transmission
5,46
5
power. Bigger transmission powers lead to fewer CH.
4 LMOCA is part of a clustering protocol and by this way,
future works consist of finding a CH rotation policy based in
3,47
REFERENCES
[1] Brooks, R. R., Ramanathan, P. and Sayeed, A. M. (2003) “Distributed
0
2,04 3,62 4,83 5,74 6,44 7,23 8,11 8,59 11,46 15,28 48,31
[4] Krishnamurthy, L., Adler, R., Buonadonna, P., Chhabra, J., Flanigan, [11] Dali, W. and Chan, A. C. (2005) “A Survey on Cluster Schemes in Ad
M., Kushalnagar, N., Nachman, L. and Yarvis, M. (2005) “Design and Hoc Wireless Networks”, Proceedings of the IEEE Mobility Conference.
Deployment of Industrial Sensor Networks: Experiences from a Guang Zhou, China.
Semiconductor Plant and the North Sea”, Proceedings of the 3rd ACM [12] Heinzelman, W. R., Chandrakasan, A. and Balakrishnan, H. (2000)
Conference on Embedded Networked Sensor Systems (SenSys). San “Energy-Efficient Communication Protocol for Wireless Sensor
Diego, California, USA. Networks”. Proceedings of the 33rd Hawaii International Conference on
[5] Szewczyk, R., Mainwaring, A., Polastre, J., Anderson, J. and Culler, D. Systems Sciences .
(2004) “An Analysis of a Large Scale Habitat Monitoring Application”, [13] Vlajic, N. and Xia, D. (2006) “Wireless Sensor Networks: To Cluster or
Proceedings of the 2rd ACM Conference on Embedded Networked Not to Cluster?”, Proceedings of the 2006 International Symposium on a
Sensor Systems (SenSys). Maryland, USA. World of Wireless, Mobile and Multimedia Networks (WoWMoM'06).
[6] Su, H. and Zhang, X. (2006) “Optimal Transmission Range for Cluster- [14] Younis, O. and Fahmy, S. (2004) “Distributed Clustering in Ad-hoc
Based Wireless Sensor Networks with Mixed Communication Modes”, Sensor Networks: A Hybrid, Energy-Efficient Approach”, Proceedings
Proceedings of the 2006 International Symposium on a World of of the. Twenty-third Annual Joint Conference of the IEEE Computer and
Wireless, Mobile and Multimedia Networks (WoWMoM'06). Communications Societies, vol. 1, p. 640.
[7] Heinzelman, W. B. and Chandrakasan, A. P. (2002) “An Application- [15] Wang, W. and Jantsch. A (2006) “A new protocol for electing cluster
Specific Protocol Architecture for Wireless Microsensor Networks”, head based on maximum residual energy”. In Proceedings of the IEEE
IEEE Transactions on Wireless Communications, volume I, Issue 4, Oct International Cross-Layer Designs and Protocols Symposium.
2002, p.p. 660 - 670. [16] Pham, H. N, Pediaditakis, D. and Boulis, A. (2007) “From Simulation to
[8] Kuhn, F., Moscibroda, T. and Wattenhofer, R. (2004) “Unit Disk Graph Real Deployments and Back”. in Proceedings of t2pWSN.
Approximation”, Proceedings of the 2004 Joint Workshop on [17] Omnet++ (2009) “OMNET++ Discrete Event Simulation System”.
Foundations of Mobile Computing, Philadelphia, PA, USA, p.p. 17 - 23. http://www.omnetpp.org.
[9] Younis, O., Krunz, M. and Ramasubramanian, S. (2006) “Node [18] XBow (2009) “Crossbow Technology”. http://www.xbow.com. San
Clustering in Wireless Sensor Networks: Recent Developments and Jose, CA. USA.
Deployment Challenges”, IEEE Network. May/June. [19] Ye, M., Li, C., Chen, G. and Wu, J. (2005) “EECS: An Energy Efficient
[10] Funke, S., Kesselman, A., Meyer, U. and Segal, M. (2006) “A Simple Clustering Scheme in Wireless Sensor Networks”. the 24th IEEE
Improved Distributed Algorithm for Minimum CDS in Unit Disk International Performance, Computing, and Communications
Graphs”, ACM Transactions on Sensor Networks, Vol. 2, No. 3, August, Conference. p.p. 535-540.
p.p. 444-453.
On the Design of SOHO Networks
Lawan Ahmed Mohammed
King Fahd University of Petroleum and Minerals
HBCC, 31991 Hafr Batin, Saudi Arabia
[email protected]
Abstract — Small networks such as small office/ home office and portable computing nodes into existing fixed
(SOHO) face the same threats as large enterprise networks. infrastructure based networks increases the heterogeneity of
However, they also contend with the challenge of limited the resulting network, making the problem of intrusion
resources and budgets for IT expenditures. In such networks detection even harder especially as these devices can be
environment, the role of administering the system often falls on connected to each other and also to Internet by using wired or
the business owner or on the default setup when the system was wireless technology. For example, wireless technology like
installed. In most cases, the owners do not usually have the time,
Bluetooth, 802.11, IrDA can provide LAN and Internet access
resources or expertise to work on complex security problems.
Similarly, most of the default setup use to have some loopholes
to a handheld computer wirelessly through access points. The
which were not initially identified. This paper describes the mobility aspect of these devices makes it harder to protect
primary security challenges facing SOHO networks today, and them. Further, different components and networking
suggests simple easy to use security solutions to resolve these technologies support different security mechanisms, which
challenges. may not be compatible and which have different security
characteristics. This implies the need for multi vendor’s
Keywords— Network security, Design of SOHO, Flooding attack, device and security management. For instance, usually,
Security measures. Access Point (APs) from manufacturers comes with a set of
default configuration parameters. These default parameters
I. INTRODUCTION leave some loopholes for attacks. For example (depending on
the manufacturer), most APs have a default administrator
The development of cost-effective networking password, SSID, channels, authentication/encryption settings,
technologies has provided large amount of new possibilities SNMP read/write community strings, etc. Since these default
for the users in small office and home (SOHO) environments. values are available in user manuals, vendor’s website and
Some of the benefits of these networks include: Improved installation guide, they are well known to the general public
efficiency and access to mobile employees, Offers mobility and may be used by hackers to compromise WLAN security.
and flexibility, Increase productivity and reduced costs, Furthermore, these environments involves different category
Optimizing business processes, Increase competitive of users with different access right and privacy requirements.
advantage, Improve customer satisfaction and retention etc. For example, home users may include parents, children,
As with all types of networking environments, SOHO guests, and external administrators while offices may have
environments and technologies are also vulnerable. Security employees with different roles. Finally, as the system may
issues such as authentication, authorization, and access control involve different type of users, individual skills and expertise
requirements should be considered when designing services challenges should also be considered.
for such networks. As SOHO networks can be either wired or As SOHO is generally small and require fewer
wireless, peculiar vulnerabilities associated with the chosen infrastructures, attack on this network can easily make the
technology should also be taken into consideration. In view of entire system inoperable. In view of these, the focus of the
these, implementing and supporting scalable SOHO faces paper is to examine the security threats and requirements
many challenges. For instance, these environments consist of especially in the design of wireless SOHO networks. First, the
variable hardware and software components, coming from paper provides an overview of SOHO design. Later, it
different vendors or manufacturers some being more discusses security threats and attacks for these environments.
trustworthy than others. In this case issues to be managed may Finally, it presents some methods of preventing the attacks.
include integration management, rapid technological
expansion and changes etc. II. OVERVIEW OF SOHO DESIGN
Lightweight SOHO computing devices like Personal
Digital Assistants (PDAs), mobile phones, and laptops are Wired SOHO uses either dialup or Ethernet. This consists
also prone to large set of software attacks like the denial of of network adapters, hub, and simple configuration software.
service (DOS) or buffer overflow attacks. These devices when Unlike in wired SOHO, wireless SOHO uses WLAN
connected to infrastructure based networks like office LAN, technology based on the IEEE 802.11 standard. Users can
make them vulnerable to the attacks similar to that on freely move inside or outside the building and still remain
infrastructure based networks. Addition of mobile, handheld, connected to the network and other hardware devices on the
network. This network components can be set up anywhere in
555
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_97, © Springer Science+Business Media B.V. 2010
556 MOHAMMED
the building. Further, wireless SOHO is very flexible and also On the negative side, some of the key challenges can be
allows users to move computers and other devices without the summarized as follows:
need to reconfigure the network. Flexibility comes in as users • Integration management
can get connected through simple steps of authentication • Multi vendors device and security management
without the hassle of running cables. Also compared to the • Rapid technological expansion and changes
wired network, wireless network installation costs are minimal • Skill and expertise challenges
as the number of interface hardware is minimal. As illustrated • Future proofing challenges etc
in table 1 below, WLANs are less expensive to install even for
relatively modest workforces. Those savings grow along with III. SECURITY THREATS
the size of the workforce as the cost associated with WLAN
controllers is further amortized. Further, wireless SOHO does SOHO computing environment share all of the security
not require any computer running Windows server 2003, it issues of traditional networked applications. These include
fully supported in Windows XP (even home edition), it is authentication of devices and users, privacy of data or
relatively easy for even a novice user to deploy, it does not information, defence against malicious code such as viruses,
require the purchase of a server certificate, and also provides worms, Trojan horses etc, and access control mechanisms.
simple methods to control or manage workgroup members. However, the SOHO computing environment combines both
These reasons make wireless SOHO more popular than the wired and wireless network security challenges as the system
wired SOHO. can be based on both types of environment. When wireless
TABLE 1 network is employed, information that is usually confined
COST OF WIRED VS WLAN (Source: www.Sonicwall.com) behind a corporate firewall will be winging its way through
Equipments Ethernet WLAN the air, possibly spending some time on hosted servers or
Ethernet/PoE Port $35 $35 wireless gateways. Further, the techniques of hacking
Installation $200 $250 wireless and mobile devices such as laptops, cell phones,
AP - $1,300 (50 users) PDAs etc is already spreading. In view of these, adding
Controller - $15,000 (50 APs) security to such environment presents challenges at different
Installed costs - 100 users $23,500 $16,585 levels. For instance, having a central authority for a single
Installed costs - 1,000 users $235,000 $46,700 building or even a group of rooms is infeasible because every
possible access right will have to be specified for every user.
In a typical WLAN configuration, a transmitter/receiver Authenticating the identity certificate of a previously
(transceiver) or AP, connects to the wired network from a unknown user doesn’t provide any access control information.
fixed location using standard Ethernet cable. At a minimum, Simple authentication and access control are only effective if
the Access Point receives, buffers, and transmits data between the system knows in advance which users are going to access
the WLAN and the wired network infrastructure. A single a particular subject or stored information and what their
Access Point can support a small group of users and can access rights are. Portable handheld and embedded devices
function within a range of less than one hundred to several have severely limited processing power, memory capacities,
hundred feet. Fig. 1 below shows the basic components of a software support, and bandwidth characteristics. Also,
SOHO wireless network infrastructure. hardware and software environments are becoming
increasingly heterogeneous [1]. In view of these, the security
requirements can be roughly divided into three main
categories; network security requirements, device security
requirements, and finally the user security requirements.
However, in this paper we focus more on the network security
requirements.
SOHO computing aims to provide users with computer-
supported capabilities like traditional network environment. In
order to achieve this, correct identification and verification of
users is paramount. Authentication is equally important when
allowing a user access to a secure file space. Its purpose is to
ensure that each entity is correctly matched to its
Fig 1 Basic Wireless SOHO Infrastructure corresponding privileges. If someone is not accurately
identified then their identification is left unverified. In
Some of the well-known SOHO benefits, including:
wireless SOHO computing environment, the risk of someone
• Improved efficiency and access to mobile employees
fraudulently performing input using authorized user’s login
• Offers mobility and flexibilty details is greater and, for reasons such as this, access to these
• Increase productivity and reduced costs workspaces and the computers within them may require more
• Optimizing business processes physical and information security measures than the normal
• Increase competitive advantage wired network environment. Details of WLAN vulnerabilities
• Improve customer satisfaction and retention. etc can be found in [2], [3] and [4]. Moreover, one of the main
ON THE DESIGN OF SOHO NETWORKS 557
difficulties in designing a secure SOHO environment is in amount of network activity serves as an indicator for the
ensuring that the functionality of the environment is not over occurrence of a large event. Secondly, the attacker can find
constrained by security. Brief descriptions of some specific information about the location of APs in the surrounding area.
attacks targeting the entire network and its infrastructure when This is because unless turned off, APs broadcast their service
the devices are connected to the SOHO network are given set identifiers (SSIDs) for identification. Thirdly, the attacker
below: may learn the type of protocols being used in the transmission.
Web Attacks – The web browser is a complex application This knowledge can be obtained based on the size, type and
that can be extended through entire programming languages number of packets in transmission over a period of time.
such as Javascript, Java applets, Flash applications and Analysis of TCP three-way handshake is described in [5].
ActiveX controls. While these technologies make for a richer Unauthorized Access – this attack involves gaining free
web experience for the end user, each of them is a new attack access to the network and also using the AP to bypass the
vector in a program running inside a firewall. Filtering firewall and access the internal network. Once an attacker has
network traffic cannot protect against all attacks. Further, even access to the network, he/she can then launch additional
Web sites that are legitimate for use in a business context can attacks or just enjoy free network use. Although free network
serve as a source of threats. use may not be a significant threat to many networks, however,
Impersonation attacks - An intruder assumes the access is a key step in ARP based man-in-the-middle attacks.
identity and privileges of another node in order to consume AP Overloaded - This attack is based on the observation
resources or to disturb normal network operation. An attacker that in 802.11 a client must be successfully authenticated and
node achieves impersonation by misrepresenting its identity. associated to an AP before using wireless communication
This can be done for example by changing its own MAC services. AP maintains the client state information in a client-
address to that of some other legitimate node. Strong association table. When the table reaches the permitted level
authentication procedures can be used to stop attacks by of associated clients, the AP start rejecting new association
impersonation. requests. This attack is facilitated by the open system
Eavesdropping – This implies the interception of authentication method in 802.11where an AP authenticates
information/data being transmitted over the network. In case anyone who requests authentication. An attack can be
of wireless SOHO, when the wireless link is not encrypted, an launched if the adversary does a large number of associations
attacker can eavesdrop the communication even from some with an AP, using random MAC addresses. Since an AP can
few miles away. The attacker can gain two types of maintain a limited number of associations, this will prevent
information from this attack. The attacker can read the data other stations from joining the AP.
transmitted in the session and can also gather information Layer-Level Attack - Every year, network attacks
indirectly by examining the packets in the session, specifically become more widespread, more intelligent and more difficult
their source, destination, size, number, and time of to detect. Because many of today's attacks are blended attacks
transmission. Eavesdropping can also be active; in this case which use multiple techniques at different layers to try to
the attacker actively injects messages into the communication infiltrate the network, they can bypass outdated firewalls that
medium in order to assist him/her in determining the contents only provide packet inspection for network traffic. With
of messages. packet inspection, firewalls mainly ensure that connections are
Poor network administration - A surprising number of valid. However, they perform absolutely no inspection beyond
machines are configured with an empty or easy to guess the first few networking layers and thus do not concern
root/administrator password. One of the first things an themselves with the content carried within these data streams,
attacker will do on a network is to scan all machines for empty allowing both desired and malicious data through. Further, as
or commonly used passwords. the amount of traffic being scanned by firewall as well as the
System Intrusion – Also known as “Privilege increasing amount of threats and malware lurking in that
Escalation”, this type of hacking assumes the attacker already traffic is quickly becoming more than many firewalls can
has a low-privilege user account on the system. If the system handle. Inspecting every byte of every packet can overwhelm
doesn't have the latest security patches, there is a good chance some firewalls and bottleneck network performance. This not
the attacker will be able to use a known exploit in order to only keeps users from getting the most out of the available
gain additional administrative privileges. bandwidth, but also degrades and disrupts streaming
TCP/IP protocol flaws - The TCP/IP protocol was applications like Voice over IP (VoIP) phone and video.
designed before we had much experience with the wide scale Modern Malware - The goal of malware is money, not
hacking we see today. As a result, there are a number of recognition, and stealth is highly prized. The methods and
design flaws that lead to possible security problems. Some techniques for making malware pay expand and evolve at a
examples include smurf attacks, ICMP Unreachable rapid pace, with the more general risks of infected systems
disconnects, IP spoofing, and SYN floods. being absorbed into botnets for use in spam campaigns and
Traffic Analysis – Traffic analysis allows the attacker to DDoS attacks, draining bandwidth and resources, and halt
obtain three forms of information. The attack primarily device performance, supplemented by the more personal
identifies that there is activity on the network. Similar to dangers of data theft. A more dangerous malicious program
standard radio communications, a significant increase in the can transmit itself across the network, bypassing some of the
558 MOHAMMED
list of ICMP requests, it will allow these packets to their Against Spoofing - Attackers launching spoofing usually
destination. If there is an internal router, the router will hide the identity of machines they used to carry out an attack
respond with a ICMP “Host Unreachable” for every host that by falsifying the source address of the network
it cannot reach, thus giving the attacker knowledge of all hosts communication. This makes it more difficult to identify the
which are present behind the filtering device. sources of attack traffic. It is therefore important to use
Crafted ICMP Messages: An attacker can send a network switches that have MAC binding features that store
specially crafted ICMP message to an affected system. An the first MAC address that appears on a port and do not allow
attacker who successfully exploited this vulnerability could this mapping to be altered without authentication. To prevent
cause the affected system to reset existing TCP connections, IP spoofing, disable source routing on all internal routers and
reduce the throughput in existing TCP connections, or use ingress filtering. DNS spoofing can be prevented by
consume large amounts of CPU and memory resources. These securing the DNS servers and by implementing anti-IP
attacks, which only affect sessions terminating or originating address spoofing measures. Some vendors have added access
on a device itself, can be of three types: 1. Attacks that use control lists (ACL), implemented through MAC address
ICMP "hard" error messages, 2. Attacks that use ICMP filtering, to increase security. MAC address filtering amounts
"fragmentation needed and Don't Fragment (DF) bit set" to allowing predetermined clients with specific MAC
messages, also known as Path Maximum Transmission Unit addresses to authenticate and associate. While the addition of
Discovery (PMTUD) attacks, 3. Attacks that use ICMP MAC addresses filtering increases security, it is not a perfect
"source quench" messages. Successful attacks may cause solution given that MAC addresses can be spoofed. Also, the
connection resets or reduction of throughput in existing process of manually maintaining a list of all MAC addresses
connections, depending on the attack type. can be time consuming and error prone. Therefore MAC
address filtering is probably best left for only small and fairly
V. DEFENSE MECHANISMS static networks [10].
Against Flooding Attack - A combination of Host-
The following countermeasures can be taken to minimize or based Intrusion Detection System (HIDS) and Network-based
thwart the attacks mentioned above. Intrusion Detection System (NIDS) can greatly help against
this attack. HIDS can be placed on critical servers and NIDS
Against Unwanted Web Contents - An effective way that can be placed on one or more network segments. Signature
can be implemented to block Web-based threats is content detection scheme would be good at detecting any known
filtering designed to block unwanted file types. Blocking file attacks. Alerts arising from any suspicious activity can be
types based on their content can be useful in preventing some intimated to the administrator immediately. NIDS reactions
types of Web threats from entering a network, particularly can also be TCP resets, IP session logging and Blocking.
files that are traditionally known to be associated with HIDS approach looks into log files, file checksums and
malware, such as .scr or .pif. These systems can also block file intercepting requests to the operating system for system
types that are generally not used in a legitimate business resources before they can proceed. Signatures and generic
context, such as .mp3, .jpg or .mov files. rules help in anomaly detection. Open server ports can also be
Against Access to Harmful Websites - Another option monitored for excess or abnormal traffic. Firewalls are an
that should be considered is the deployment of URL filtering excellent form of protection; however, they must leave some
tools that will block access to non-approved Web sites. Many ports open to allow the operation of the web, mail, ftp, and
organizations have deployed these filters, albeit with varying other Internet based services, and which are the paths
levels of success. While URL filters can be useful, they can exploited by most of the vulnerabilities. A more effective
rarely keep up with the new threats that enter the Web on an measure may involve a hybrid IDS system. This is a
hourly basis and for which no signature has been created in distributed system where each and every node in the network
the tool. Further, URL filters can generate significant levels of collectively cooperated to form a team of IDS agents for the
false positives – blocking Web sites that appear to be entire network. This can easily be done by placing an IDS
suspicious but might have a legitimate business purpose. intelligent agent on every node in the network. Each IDS
Against Malware - A new approach is required to thwart agent runs independently and monitors local activities (user,
the escalating threats posed by targeted malware and botnets system, or communication) in the network. It will then
that render traditional security techniques obsolete. Deploying program to detect any intrusion from local traces (especially
an anti-botnet protection system that uses global intelligence communication activities). If any anomaly or suspected
combined with local analysis and control ensures that botnets activity is detected, a report is send to the host server and
do not promulgate within the enterprise network. An adequate other nodes. These individual IDS agents form the system’s
protection system scans all network traffic and executes IDS to defend the entire network.
relevant traffic streams in virtual victim machines – safe from Against Layer-Level Attack - One of the simple was to
real clients – and conclusively determines if botnet malware is secure SOHO network today is the use of Unified Threat
present. This is the only real defense against network-borne Management (UTM) tools that layer on antimalware
attempts at spamming, spear-phishing, coordinated attacks, protection, content filtering, antispam and intrusion prevention,
identity theft, or theft of intellectual property. It eliminates the because deploying a single, multi-function device reduces
underlying cause for all these events—botnets
560 MOHAMMED
costs and simplifies configuration. Thus, UTM is a network capabilities. SSL VPNs are best suited for secure remote
security tool providing a new level of scrutiny to network access during an emergency because they allow workers and
traffic passing into an organization. Simply put, UTM partners to connect safely to corporate network resources
firewalls combine the effectiveness of various point defences using a Web portal, without having to pre-install clients.
to add protection at every networking layer. The power of Modern continuous data protection (CDP) solutions can
UTM comes from its simplicity: a single appliance that automatically backup data and applications to discs, avoiding
provides a powerful defence against a wide range of security the complexity and human error involved with tape backup.
threats, while extending beyond, the capabilities of regular Backup to a secure secondary business location or third-party
packet inspection firewalls. This makes network protection site means business systems can be restored and operational
more complete, affordable and easy to manage. For optimal even in the primary site is compromised. Bare metal recovery
performance while maintaining maximum security, solutions (BMR) technology enables entire operating systems, such as
such as reassembly-free deep packet inspection UTM help to database or file servers, be recovered to new or different
deliver throughput that won't bog down performance. Clever hardware platforms if the original device can’t be restored.
algorithms help to reduce latencies to non-noticeable levels.
Also, advances in microprocessor technology, especially those VI. CONCLUSION
designed for network processing, allow UTM appliances
designed for SOHO environment to gain incredible network This paper described some vulnerabilities associated with
efficiency. SOHO networks. It shows that implementation of wireless
Alternatively, a firewall can be considered when SOHO posses more loopholes than wired SOHO if proper
connecting that network to a second network (such as virtual security measures are not taken into consideration. The paper
private network or VPN) which has a less stringent security discuses some essential areas required for a complete and
policy. Implementing firewall along with VPN over the robust protection policy. These include anti-malware
SOHO LAN/WLAN can strengthen the system’s security. protection on all platforms, along with firewalling, intrusion
IPSec can be used as the security protocol and the secure prevention, network access control, data leak prevention and
tunneling with authentication and encryption. In [11], we data encryption among others.
demonstrated this idea using a simple network topology. Two
nodes A and B had to access database and web pages on a ACKNOWLEDGMENT
Server through a network with firewall and VPN server The author wishes to acknowledge Hafr Al-Batin
implemented. The firewall would not permit database access, Community College and King Fahd University of Petroleum
but would allow HTTP or web service access. A compulsory and Minerals (KFUPM) Saudi Arabia for their support in
VPN tunnel is implemented between the routers (source router providing the various facilities utilized in the presentation of
and destination routers) that connect node A to the Server. So this paper.
even if the firewall blocks a service, a node with VPN tunnel
enabled can go through the firewall as data communication REFERENCES
happens as encapsulated packets (packet within a packet). [1] L. Kagal, F. Tim, and J. Anupam, “Trust-Based Security in Pervasive
Against Wireless Breaches -When wireless network is Computing Environments”, Computer v34, n12, Dec 2001, pp:154-157T.
deployed, it is recommended to check all default settings, [2] T. Takahashi, “WPA passive dictionary attack overview” [Online]:
including administration password and access path, and http://www.3c.com.vn/Uploaded/huyenctt/New%20Folder/WPA_Passiv
change them or disable functions as needed. Turn off all e_Dictionary_Attack_Overview.pdf
access point services that are not being used (ftp, http, etc…). [3] P. Liza, “WPA PSK Crackers: Loose Lips Sink Ships”, [Online].
Also WEP encryption must be turned on and must use 128 bit www.wi-fiplanet.com/tutorials/article.php/3667586
keys. Moreover, SSIDs must be changed from defaults, must [4] E. Tews ‘Gone in 900 Seconds, Some Crypto Issues with WPA’,
PacSec 2008 Conference, Nov. 2008, Tokyo, Japan.
not reflect any information about the organization or location
[5] W. Donald, “Wireless Security Threat Taxonomy”, IEEE Workshop on
of the Access Point, and SSID cloaking must be turned on. Information Assurance. US Military Academic West Point N.Y, June
MAC address filtering must be used to restrict wireless access 2003, pp: 76-83.
to only approved hardware addresses. The AP locations [6] Internet Control Message Protocol (ICMP)., [Online]:
should be in interior positions and away from exterior walls http://www.erg.abdn.ac.uk/users/gorry/course/inet-pages/ icmp.html
and windows, ensure they are placed in secured locations to [7] ICMP Attack, [Online]: http://www.javvin.com/networksecurity/
prevent unauthorized physical access. ICMPAttacks.html
[8] Cisco White Paper: “Crafted ICMP Messages Can Cause Denial of
Disaster Recover Plan - No matter how secure the Service “ [Online]: http://www.cisco.com/warp/public/707/cisco-sa-
network is, the need for a disaster recovery solution is 20050412-icmp.shtml
essential. Major disasters have demonstrated how exposed [9] SANS White Paper, “ICMP Attacks Illustrated” [Online]:
small businesses can be to unexpected events. Building fires, http://www.sans.org/info/36914
[10] L. A. Mohammed and B. Issac, “Detailed DoS Attacks in Wireless
broken water pipes, power outages, equipment failures, Networks and Countermeasures” International Journal of Ad Hoc and
terrorist, or even lost or stolen handheld devices, can mean Ubiquitous Computing, (IJAHUC), Vol 2. No 3, pp. 157-166, 2007
disaster for small businesses networks such as SOHO. Modern [11] B. Issac and L. A. Mohammed , “War Driving and WLAN Security
UTM appliances can feature integrated IPSec or SSL VPN Issues Attacks, Security Design and Remedies”, Information Systems
Management, Vol 24, pp. 289-298, 2007
Study of UMTS Handover Scenarios
B. Issac
Swinburne University of Technology (Sarawak Campus), Malaysia
[email protected]
well as service providers and that will create added interest and
Abstract— Universal Mobile Telephone system (UMTS) is a it can be expected that different forms of service providers will
third generation mobile and wireless communications standard play a more definite role in the future. This could in turn boost
which is also being developed into a fourth generation technology.
We wanted to study the performance of such a wireless cellular new business models and new forms of revenue sharing
network during soft and hard handover scenarios with single and through the advent of new services.
multiple towers. An attempt is thus made to use existing The content services that will be provided by UMTS will
simulation models to study the related performance graphs as the evolve from the services provided by GSM. These services
mobile node moves around. offered by UMTS can be classified as “mobile Internet”, where
display text or other contents are optimized for the limited
graphical capabilities of the mobile devices, along with the
I. INTRODUCTION location awareness functionality. A range of services have
been suggested, for example: Images and video: streaming
T he European countries have chosen Universal Mobile
Telephone system (UMTS) as one of the standards for the
third generation mobile communication system. The third
pictures and video, entertainment, games, lottery, gambling,
and video-conferencing, Short range and Location based
generation in the Nordic countries has been preceded by the services: Information of nearby restaurants, shops and special
first generation mobile telephone system NMT (Nordic Mobile offers, Context based services, or push: Notification of
Telephone), and the second generation GSM (Global system interesting sports event if you are in front of the TV,
for Mobile Communication). UMTS is planned to coexist or advertising if you are close to the advertiser, Simultaneous
work with 2G for several years, as NMT coexisted with GSM transfer of speech, data, text, pictures, audio and video, High-
from 1992 to 2002, when NMT finally was phased out. As speed, mobile access to the Internet, Customized infotainment,
with NMT and GSM, GSM handsets will not be compatible Travel information: congested roads, flight departures, and if
with the UMTS. On the other hand, the different networks will you get lost, find your current location etc.
not be totally separated as UMTS handsets will be compatible When mobile users are roaming (moving from one area or
with GSM providing a dual mode of communication. This is cell to another) the handset must attach itself to the closest
because UMTS, as NMT, will provide less coverage than base-station. When one base-station hands over the user to
GSM, probably only rural areas with high density of users. another base-station, multiple transactions must be made. If the
UMTS is defined by European Commission [1] as: “... a user is roaming among base-stations belonging to the
third-generation mobile and wireless communications system subscribed operator, this is only a technical issue. But when
capable of supporting in particular innovative multimedia roaming implies other operators, national or international, the
services, beyond the capability of second generation systems picture becomes more complicated. The vision of UMTS
such as GSM ...” . This definition is not that specific on what presupposes that it will be possible for users to roam freely
new services can be expected by the users. International with their handsets and have affordable access to equivalent
Telecommunication Union is slightly more specific as they products and services globally by end-users utilizing the most
define UMTS by higher capacity and enhanced network proper network despite which operator it belongs to. Today,
functionalities that would allow more demanding services and technically and efficient roaming exists for the GSM standard
applications, like multimedia. For an end-user, the major throughout Europe, parts of Africa, Asia and South America.
difference between 3G and 2G is underlined by these However, it is far from global: for instance, roaming to and
definitions to be the increased user value created by a broad from North America and Japan requires special measures and
service offering. UMTS provides key features for end-users users may even incur custom duties and taxes relating to the
provided as – Options or facilities for Internet and other use of their handsets abroad. Moreover, the complexity and
multimedia-rich applications, a good range of content services non-transparency of pricing for roaming services have been
and global roaming capability. problematic for 2G or 2.5G users [2]-[8].
UMTS will provide basic services as with GSM and they are – Some related works on UMTS can be seen in [2]-[7]. The
voice, Short Messaging Service (SMS), roaming and billing. paper is organized as follows. Section 2 discusses the features
So multimedia services would be developed by operators as of UMTS, section 3 is the architecture details, section 4 is the
561
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_98, © Springer Science+Business Media B.V. 2010
562 ISSAC
simulation of two UMTS scenarios where we use different propagation constraints or other economic reasons. In order to
network topologies and section 5 is the conclusion. enable subscribers to always use their terminal, services will be
adaptive to different data rate availability and other Quality of
II. UMTS FEATURES Service (QoS) parameters. In the early stages of deployment,
A. Fast access UMTS coverage can be limited. UMTS can enable roaming
UMTS is better than 2G or 2.5 G cellular technology for its with other networks like GSM system operated by the same
potential to support 2 Mbit/s data rates. This capability, operator or with other GSM or 3G systems of other operators,
together with the built-in IP support, merges powerfully to including UMTS compatible satellites.
allow interactive multimedia services and new wideband
multimedia-rich applications, such as video telephony, video F. UMTS Services global availability through Satellite
conferencing, network video gaming etc. Satellite technology can give enhanced global coverage and
services and it is expected to play an important role for UMTS
B. Packet transmission and data rate on demand world-wide coverage. UMTS is being standardized to ensure
Current cellular systems use circuit switched technology for an efficient and effective roaming and handover between
data. UMTS combines packet and circuit data transmission satellite and terrestrial networks.
with the advantage of virtual connectivity to the network at all
times and alternative ways of billing. For example, pay-per-bit G. Spectrum for UMTS
option, to pay per session, to pay flat rate etc as demanded by WRC-92 identified the frequency bands 1885-2025 MHz
many upcoming data services. The operation of the system can and 2110-2200 MHz for future IMT-2000 systems, with the
be made much cheaper as UMTS is also being designed to bands 1980-2010 MHz and 2170-2200 MHz intended for the
offer data rate on demand which can work in combination with satellite part of these systems. These bands can be made
packet data transmission. available for UMTS licenses to enable operators for network
deployments. Wideband Code-Division Multiple-Access (W-
C. Friendly and consistent service environment CDMA) is one of the main technologies for the
UMTS services are based on having the user face common implementation of third-generation (3G) cellular systems. In
capabilities throughout all UMTS user and radio environments. W-CDMA interface different users can simultaneously
When roaming from the user’s network to other UMTS transmit at different data rates and data rates can even vary in
operators, a personal user will experience a consistent set of time. UMTS networks need to support all current 2G/3G
services and this would give a feeling or sense of his or her services [5]-[7].
home network (termed as “Virtual Home Environment” or
VHE). VHE will ensure the delivery of the service provider’s III. UMTS ARCHITECTURE
full environment. VHE will enable terminals to handshake and A UMTS network consists of three domains that interact
agree on some functionalities with the visited network, and with each other, namely – Core Network (CN), UMTS
"home network like" services will be presented with full Terrestrial Radio Access Network (UTRAN) and User
security and transparency across a mix of access and core Equipment (UE) [8]. The main function of the core network is
networks. to provide switching, routing and transit for user traffic. Core
network also contains the databases and network management
D. Mobility and Coverage functions. The basic Core Network architecture for UMTS is
UMTS is implemented as a global system, combining both based on GSM network with GPRS. All equipment has to be
terrestrial and satellite communication network components. modified for UMTS operation and services. The UTRAN
Multi-mode terminals operating also via 2.5 or 3G systems will provides the air interface access method for User Equipment.
further extend the reach of many UMTS services. With these Base Station is referred as Node-B and control equipment for
terminals a user will be able to move around and roam from a Node-B's is called Radio Network Controller (RNC).
private network into a micro-cellular public network, then into It is important for a network to know the approximate
a wide area macro-cellular network (like a 2.5 or 3G network), location in order to be able to contact the user equipment.
and then to a satellite mobile network, with minimal break in Some of the list of system areas from largest to smallest –
communication and packet loss. UMTS systems (including satellite), Public Land Mobile
Network (PLMN), MSC/VLR (Mobile Switching
E. Radio Technology for all environments Center)/(Visitor Location Register) or SGSN (Serving GPRS
The UMTS radio interface UTRA (UMTS Terrestrial Radio Support Node), Location Area, Routing Area (PS domain),
Air interface) can help operation with high spectral efficiency UTRAN Registration Area (PS domain), Cell and Sub cell etc.
and service quality. In practical implementations, UMTS
terminals might be unable to function at the peak data rates at A. Core Network
all times, and in remote or heavily congested areas system The Core Network consists of circuit switched and packet
services might only support lower rates due to radio switched domains. Some of the circuit switched elements are
STUDY OF UMTS HANDOVER SCENARIOS 563
Mobile services Switching Centre (MSC), Visitor location services to be offered over the PS domain (like VoIP). CS
register (VLR) and Gateway MSC. Packet switched elements mode of operation: The MS is attached to the CS domain only
are Serving GPRS Support Node (SGSN) and Gateway GPRS and may only operate services of the CS domain.
Support Node (GGSN). Network elements like EIR UMTS IC card has similar physical characteristics as GSM
(Equipment Identity Register), HLR (Home Location SIM card. It has various functions as follows – support of one
Register), VLR (Visitor Location Register) and AUC User Service Identity Module (USIM) application (optionally
(Authentication Center) are shared by both domains. The more that one), support of one or more user profile on the
Asynchronous Transfer Mode (ATM) is defined for UMTS USIM, Update USIM specific information over the air,
core transmission. ATM Adaptation Layer type 2 (AAL2) security functions, user authentication, and optional inclusion
handles circuit switched connection and packet connection of payment methods and optional secure downloading of new
protocol AAL5 is designed for data delivery. The architecture applications [8]-[9].
of the Core Network may change when new services and
features are introduced. Number Portability Database (NPDB) IV. UMTS HANDOVER
will be used to enable user to change the network while There are following categories of UMTS handover (also
keeping their old phone number. Gateway Location Register referred to as handoff) that can happen:
(GLR) may be used to optimize the subscriber handling
between network boundaries. A. Hard Handover
Hard handover means that all the old radio links in the UE
B. Radio Access are removed before the new radio links are established. Hard
Wide band CDMA technology was opted for UTRAN air handover can be seamless or non-seamless. Seamless hard
interface. UMTS WCDMA uses Direct Sequence CDMA handover means that the handover is not perceptible to the
where user data is multiplied with quasi-random bits derived user. In practice a handover that requires a change of the
from WCDMA spreading codes. In UMTS, in addition to carrier frequency, known as inter-frequency handover is
channelization, codes are used for synchronization and always performed as hard handover.
scrambling. WCDMA has two basic modes of operation:
Frequency Division Duplex (FDD) and Time Division Duplex B. Soft Handover
(TDD). The functions of Node-B are stated as follows: Air Soft handover means that the radio links are added and
interface Transmission/Reception, Modulation/Demodulation, removed in a way that the UE always keeps at least one radio
CDMA Physical Channel coding, Micro Diversity, Error link to the UTRAN. Soft handover is performed by means of
Handing and Closed loop power control. The functions of macro diversity, which refers to the condition that several
RNC are – Radio Resource Control, Admission Control, radio links are active at the same time. Normally soft handover
Channel Allocation, Power Control Settings, Handover Control, can be used when cells operated on the same frequency are
Macro Diversity, Ciphering, Segmentation/Reassembly, changed.
Broadcast Signaling and Open Loop Power Control.
C. Softer handover
C. User Equipment Softer handover is a special case of soft handover where the
The UMTS standard does not hinder or cripple the radio links that are added and removed belong to the same
functionality of the User Equipment. Terminals work as an air Node B (i.e. the site of co-located base stations from which
interface counter part for Node-B and have many different several sector/cells are served). In softer handover, macro
types of identities which are taken directly from GSM diversity with maximum ratio combining can be performed in
specifications. They are International Mobile Subscriber the Node B, whereas generally in soft handover on the
Identity (IMSI), Temporary Mobile Subscriber Identity downlink, macro diversity with selection combining is applied.
(TMSI), Packet Temporary Mobile Subscriber Identity (P- Active Set is defined as the set of Node-Bs the UE is
TMSI), Temporary Logical Link Identity (TLLI), Mobile simultaneously connected to (i.e., the UMTS Terrestrial Radio
station ISDN (MSISDN), International Mobile Station Access cells currently assigning a downlink DPCH (Downlink
Equipment Identity (IMEI), International Mobile Station Dedicated Physical Channel) to the UE).
Equipment Identity and Software Number (IMEISV).
UMTS mobile station (MS) can operate in one of three V. SIMULATION DETAILS
modes: PS/CS (Packet Switched/Circuit Switched) mode of A. Scenario 1 (Soft and hard handovers between UMTS
operation: The MS is attached to both the PS domain and CS networks)
domain, and the MS is capable of simultaneously operating PS We used existing simulation models and only verified the
services and CS services. PS mode of operation: The MS is existing results and no modifications were done to the model.
attached to the PS domain only and may only operate services Our aim is to conduct a performance comparison between soft
of the PS domain. However, this does not prevent CS-like and hard handovers between UMTS networks (with different
564 ISSAC
towers). Consider fig. 1. The configurations done are given handovers compared to the soft handovers. The difference of
below. In these scenarios, the two UEs (User Equipment) transmission power is around 3 dBm roughly. Thus soft
move around and make zigzag (back and forth) transition handovers enable the UEs to perform the handovers at power
between two Node-Bs (UMTS towers) creating multiple levels that are low, which is an added benefit for other UEs
handovers from one cell to the other. Continuous upload of around when we consider the issue of interference.
large FTP files to the server was done by the two UEs during To achieve equal performance benefits with lower power,
the simulation, which makes the traffic move in the uplink soft handovers benefit from an idea that can be termed as "soft
direction in the UMTS part of the network. UMTS QoS 3 is handover gain". As a radio bearer is aided by more than one
used for the traffic associated with FTP. radio links during the soft handovers, the network can lower
The configurations done and the network topology of both the quality (BLER, Block Error Rate) of these links below the
scenarios are same except that in hard handover case, the soft requested level as long as their combined or integrated quality
handovers are disabled at the RNC (Radio Network is enough to provide the needed or requested quality for the
Controller), so that all the handovers done are hard handovers. radio bearer. Thus the transmission power can be maintained at
To study the soft handover case, they are all configured as soft lower levels during soft handovers, as a result of lowering the
handovers. radio link quality, while still providing the same quality to
higher layer as the results show us. Have a look at fig. 2 to 5.
As the UE travels around the Node B, it notes the pilot channel But during the hard handover case, only one sector receives
signal strength that emanates from all the three sectors. These traffic. See fig.7, 8 and 9 to help understand the scenarios
signal strength measurements allows it to change its active set. fully.
Let’s have a look at the softer handover events that occurred
VI. CONCLUSION
during simulation. The UE path is from sector 1 to sector 0,
followed by sector 0 to sector 2 and finally from sector 2 to The Universal Mobile Telecommunications System (UMTS)
sector 1. At the start of simulation, based on the strongest is one of the third-generation mobile telecommunications
signal from sector 1, this sector is included in the initial active technologies, which is currently worked upon to be developed
set. The UE starts moving at 115 sec. After some time when it into a 4G technology. The technology was proposed by 3GPP
reaches the edge of sector 0 at 175 sec, adding sector 0 into the (3rd Generation Partnership Project) and is part of the global
active set. The UE is now in softer-handover between sectors 0 International Telecommunication Union IMT-2000
and 1 while it remains in this position for around 60 seconds. (International Mobile Telecommunications-2000) standard.
Now the UE traverses toward the middle of sector 0 at 235 sec 3GPP specifications are based on evolved Global System for
(remaining in sector 0 for 60 seconds) and thus UE leaves Mobile Communications (GSM) specifications. A study is
sector 1 deleting it from its active set at 255 sec. At 375 sec the done on the different scenarios of UMTS network and its
UE moves to sector 2. While UE is in the overlapped area performance evaluation is done when the mobile node is on the
between sectors 1 and 2, it gets into softer-handover state once move. We were able to look into two network scenarios where
again at 369 sec. The UE continue moving along its path to the different types of handovers were investigated and the
middle portion of sector 2, deleting or removing sector 0 and performance graphs studied. This would help to understand the
ending its softer-handover state at 412 sec. In the last traversal, issues in soft and hard handover scenarios if one were to look
the UE comes back to sector 1 and enters into a softer into the details of UMTS technology.
handover state when it lies on the area shared by both sectors
(1 and 2) at 605 sec. Then the UE goes to the center of sector REFERENCES
1, dropping sector 2 from its active set at 650 sec. [1] EU, Decision No 128/1999/EC of the European parliament and of the
Thus Softer Handover takes place when the UE is council, of the 14 December 1998, on the coordinated introduction of a
third-generation mobile and wireless communications system (UMTS) in
positioned in an area where sectors are overlapped. Hard the Community, 1998, available [online]: http://europa.eu.int/eur-
handovers take place when the UE is moving from one sector lex/pri/en/oj/dat/1999/l_017/ l_01719990122en00010007.pdf
to another. In both scenarios, the results for several statistic [2] Saravanan, N.; Sreenivasulu, N.; Jayaram, D.; Chockalingam, A.,
groups are collected like: UE active set details, Node Bs "Design and Performance Evaluation of an Inter-System Handover
Algorithm in UMTS/GSM Networks", Proceedings of TENCON 2005,
throughput, UE RLC/MAC uplink transmission power etc. 2005, pp.1-6.
Let’s have a look at the hard handover events that occurred [3] Oudelaar, J., "Evolution towards UMTS", Proceedings of 5th IEEE
during simulation. At the start of simulation, based on the International Symposium on Personal, Indoor and Mobile Radio
strongest signal from sector 1, this sector is included in the Communications, 1994. Wireless Networks - Catching the Mobile
Future, 1994, pp.852-856.
initial active set. UE starts moving at 115 sec. After some time [4] Bosch, P.; Mullender, S.; Samuel, S., "Mobility in UMTS packet IP
when it reaches the edge of sector 0 at 175 sec, a hard networks", Proceedings of Wireless Communications and Networking
handover happens. Sector 0 gets added into the active set while Conference, 2006, pp. 545 - 550.
[5] Pereira, A.; Gameiro, A., "Simulation-based Evaluation of Spectrum
sector 1 is taken off. At 355 sec, the UE traverses to sector 2 Opportunities in UMTS Cellular Networks", Proceedings of Mobile and
and a hard handover happens from sector 0 to sector 2 at 369 Wireless Communications Summit, 2007, pp.1-5.
sec. Finally the UE returns to sector 1 and performing another [6] Website: http://www.cellular.co.za/umts_technical.htm, Retrieved March
hard handover state, but now from sector 2 to 1 at 605 sec. 15, 2009.
[7] Website: http://www.umtsworld.com/technology/wcdma.htm, Retrieved
Then the UE goes to the middle of sector 1 and remains there March 20, 2009.
at 650 sec. [8] Petter Nielsen, Introduction of the third generation of mobile
As we compared the UE transmission power between both communication system in Norway, Department of informatics, University
of Oslo, Norway. Website: http://heim.ifi.uio.no/~pnielsen/PHD/Ni.doc,
softer and hard handover scenarios, we found that the UE in
Retrieved March 25, 2009.
the softer handover scenario uses less transmission power, thus [9] UMTS World, Overview of The Universal Mobile Telecommunication
reducing the cell interference. But on the other hand during System, Online: http://www.umtsworld.com/ technology/overview.htm,
softer-handovers the UE is taking more network resources. Retrieved February 2009.
Note that during the softer handover scenario, the UMTS
Node B is getting traffic through two sectors at the same time.
Reduction of Energy Consumption in WSN using an
Approach Geocast
Harilton da S. Araújo Wagner L. T. de Castro Raimir Holanda Filho
Technology Sciences Center Technology Sciences Center Technology Sciences Center
Master in Applied Computer Computer Science Dept. Master in Applied Computer
University of Fortaleza University of Fortaleza University of Fortaleza
Fortaleza – Ceará – Brasil Fortaleza – Ceará – Brasil Fortaleza – Ceará – Brasil
[email protected] [email protected] [email protected]
Abstract – Wireless sensor networks have deserved special The energy consumption is the main requisite of the
attention in the last years due to their intrinsic characteristics network and it affects all of the phases of its life cycle, not
that become them attractive in various applications, such as in being dependant upon the application. The communications in
the industrial activities, in the environment monitoring and
security, among others. In general, this kind of network can be WSN consume more energy than the processing and the
utilized as powerful tools in order to monitor and, eventually, sensor coverage performed by the network nodes. This
control an environment. The routing protocols, for the WSN, characteristic requires routing protocols that make possible
must have self-configuration characteristics aimed at discovering that the sensor nodes communicate themselves in an efficient
which is the best way for transferring the information, with and effective way with minimum energy consumption.
delivery assurance and with minimum energy consumption,
among the nodes that compose the network. This article proposes The routing protocols for sensor networks must have self-
a modification of the Directed Diffusion routing protocol so as to configuration characteristics in order to discover which the
reduce the energy consumption in the network when there is an best way is for transferring the information from a node to the
occurrence of failures. The proposal utilizes a Geocast approach
other one, with delivery assurance and minimum energy
to repair broken paths by constructing a new routing tree. The
performed simulations demonstrated that our proposal is more consumption, having the purpose of extend the lifetime of the
efficient from the energy viewpoint. network. If a node fails, through which a given information
flows, a new routing will have to be made in order that the
Keywords - wireless sensor; networks; routing; geocast information to be collected is able to reach the destination
node. The communications, among the network nodes, must
I. INTRODUCTION
be made so as the energy consumption is optimized. In this
A network of sensors is composed of a large number of sense, some protocols have the essential role of increasing the
nodes that are utilized as per request at an area of interest. time of the network useful life.
Each node has one or more sensors, in addition to the capacity
for processing, storage and communication. This article proposes a modification of the Directed
Diffusion routing protocol, having the purpose of reducing the
The Wireless Sensor Networks have aroused the interest of energy consumption in the network by means of focusing on
the scientific community due to their applicability that reaches the presence of failures. A Geocast approach was used in order
various areas such as the military, environmental, medical and to reduce the number of sent and received control messages.
industrial ones. Those networks differ in several aspects from That reduction was reached through the construction of a new
the traditional networks. Some characteristics are: they are routing tree being limited and directed towards a specific
composed of a large number of sensor nodes; the sensors have region of the network, focusing the occurrence of failures.
high energy limitations, and low processing and memory
capacity. In addition to that, some WSN applications require The article is organized in the following way: the Section II
characteristics of self-organization, being able to adjust presents the linked works. The Section III describes the
themselves, in an autonomous way, to the possible structural modifications that we made in the Directed Diffusion protocol.
changes due to external interventions such as modifications of The evaluation of the algorithm is discussed on the Section IV,
the topology caused by failures, inclusion of nodes or as per a followed by conclusions on the Section V.
request submitted by an external entity (user).
II. RELATED WORKS
The purpose of the WSN is to collect data of a region being
A proposal for optimization of the energy consumption is
covered by sensors and to allow the extraction of that data, by
analyzed on [1]. The conception of this proposal is to put
an external entity, through the sink node.
some specific nodes in the sleep (disconnected) mode in order
567
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_99, © Springer Science+Business Media B.V. 2010
568 ARAÚJO ET AL.
to save energy, while keeping the connectivity of the network utilized in order to reduce the energy consumption. The
thus assuring the communications. algorithm is described in detail on the following Section.
The EAR (Energy Aware Routing) protocol is described III. PROPOSAL FOR AN ENERGY AWARE PROTOCOL
on [2]. The EAR protocol is a reactive one that looks for paths
We propose, on this article, an extension towards an
of minimum energy consumption through the network. The
existing data centric routing protocol, Directed Diffusion [6].
basic operation consists of the occasional use of a set of paths
chosen through a probability function (which is dependant The main purpose of our proposal is to develop a protocol
upon the energy consumption of each path). In this way, it being tolerant to failures and having a minimum of energy
avoids that the best path has its energy depleted, thus consumption. In order to simultaneously reach those
increasing the network utilization time. It is assumed that each requisites, an approach, based on the Geocast routing method,
node has an address and information about its localization. was utilized in our algorithm. The approach consists in
repairing broken paths by constructing a new, limited and
An algorithm is proposed on [3], which utilizes GPS in
directed routing tree. The approach used in the proposed
order to restrict the flooding area of the packages. This
routing algorithm is described as follows:
algorithm uses two approaches in order to determine which
nodes must propagate the control messages. The first one is A. Limited and Directed Geocast Approach
determined based on the position and the size of the area. The
Based on [5], [7] and [8], the limited and directed geocast
second one is established according to the distance and the
approach was created in order to perform the construction of a
localization of the destination.
new routing tree being limited and directed towards a specific
A routing protocol, that limits the research aimed at looking region of the network denominated the geocast region. In
for new routes in ad hoc networks, is described on [4]. It is order to do that, it becomes necessary the use of geographical
utilized, in that proposal, an approach based on flooding and localization information regarding the nodes, which allows to
on the localization of the nodes in order to limit the flooding direct the routing of the messages. The approach, as adopted
area. When sending a message, the emitter node defines an on this article, restricts, in the rectangular format, the region
“expected zone” of the destination node. If any node, that that contains the incidence area of events. At the time of
receives the message, is outside that zone, it is discarded. The receiving a message, each node verifies whether it belongs to
purpose is to reduce the cost of the retransmissions of each the geocast region. If it belongs, the node sends the message
message through the whole network. again to its neighbors; in the contrary case, the node disregards
the control message. It is assumed that all of the nodes know
Alternative definitions of requisition zones, being aware of
their geographical positions.
the localization, are analyzed on [5]. Those definitions have
the purpose of optimizing the performance by delimiting the The scheme that restricts the geocast region is similar to the
areas of interest. In order to do that, the areas are approached optimization scheme as quoted on [7]. This approach restricts
based on the rectangular, circular or conic format. the flooding area by the straight lines r2,r3,r4,e r5. It assumes
An algorithm for data dissemination (Directed Diffusion) is that the node A1 (Sink) knows the localization (XA2,YA2) of the
described on [6], where it is proposed a routing scheme being node A2 (Source). Being r1 the straight line that passes through
centered on data, there not being semantics of directing. It also the points A1 and A2. The distance, between the points A1 and
attempts to take care of dynamic networks through a is represented by DA1,A2.P is the mid-point between A1 and A2.
negotiation scheme with dissemination of interests and The radios R is the distance utilized to draw the parallel
reinforcements of paths, allowing the network to converge in straight lines(r2 er3)and the perpendicular straight lines (r4 er5)
the presence of any topologic modification. In addition to that, towards the straight line. The distance between and is
the Directed Diffusion utilizes, as metrics for the selection of represented by:
the route, that one that has the lower delay, which generally
leads to the route with a lower quantity of hops. A
disadvantage of this approach is the high cost of
communications for repairs of paths when there is occurrence
of failures, due to the need of periodically performing a If the distance between and is represented by ,
flooding in the network in order to reinforce other paths. then the distance between and P is indicated by:
We present, on this article, a proposal of routing protocol
for wireless sensor networks. The proposal is implemented as
a modification of the Directed Diffusion protocol, and it
differs from the solutions presented in the literature for the
fact that it uses an approach to minimize power consumption The figure 1 shows an example of the restricted geocast
during recovery from failures. Localization information is area.
REDUCTION OF ENERGY CONSUMPTION IN WSN USING AN APPROACH GEOCAST 569
to locally determine the route along which the data will be established. In order to do it, basically, the Directed Diffusion
conveyed without incurring in high cost related to the protocol restarts the reinforcement mechanism to search for
knowledge of the whole network topology. The selection other paths. However, the repair scheme of paths in the
process of routes can be represented as follows. Being Directed Diffusion has a high cost as regards to the energy
a directed graph, where is a set of vertexes consumption as it requires a flooding of the network to
(sensor nodes) and is the set of pairs , not ordered of reinforce other paths.
elements in (connections). In case that a message is In the case of occurrence of failures in the nodes that
conveyed through a path ) in a graph complete the stretches, our algorithm utilizes a Geocast
, where , ... , are vertexes and , ... , ) approach instead of performing a flooding of the network for
are edges, then each node in loses some of its energy linked reinforcement of other paths. That approach has the purpose of
to the cost for sending messages. Being the cost for repairing broken paths by constructing a routing tree being
sending the message, , the initial energy, and the limited and directed towards a specific region of the network
residual energy of after sending a message. Hence, in which the set of source nodes is located, thus minimizing
for . Consequently, the the number of control messages being sent and received. The
approach restricts, in the rectangular format, the area where
energy of the path is indicated by , where:
the new routing tree will be constructed. At the time of
receiving the package, each node verifies whether it belongs to
the geocast region. If the package belongs, the node sends the
message again to its neighbors; in the contrary case, the node
The rule, as implemented in the protocol, does not choose disregards the control message.
just the path that has the highest value of available residual
energy for all of the nodes, but also the number of nodes that
compose the path. As the solution utilizes just localized
interactions, the reason is used for getting the best path. The
reason is got by means of the division between the sum of
the residual energy ( ) and the number of hops that compose
the path ( ), according to what the equation shows .
occurrence of failures. In the first approach, the broadcast for broadcasts for the reconstruction of routes at the Directed
the construction of the tree has maximum reaching, that is to Diffusion, while in our proposal the broadcasts are limited to
say, it reaches all of the network nodes. According to the the geocast region and they are directed to the region of
second approach, the limited and directed construction of the interest.
routing tree is performed, having the purpose of reaching just
the nodes that are close to the occurrence of events.
IV. EVALUATION OF PERFORMANCE
In order to evaluate the performance of our proposal, we
utilized the Sinalgo simulator [11]. Sinalgo is a framework,
written in Java, that allows the simulation of wireless
networks and abstracting itself from the lower layers of the
network stack.
A. Simulation Scenario and Metrics Utilized
The simulation scenario was constructed in order to allow a
didactic understanding of the proposed algorithm. The
topology of our scenario allows that each node communicates Figure 5. Energy map (Directed Diffusion)
with up to 8 neighboring nodes. There are 121 nodes, in an 11
x 11 network topology.
The Figure 7 shows that the energy consumption, in the It is intended to evaluate, in future works, the cost for
whole network, is reduced by utilizing our proposal. This is receiving messages, which measures how many messages are
got due to the use of the geocast approach regarding the delivered for each energy unit (J) used for all of the network
mechanism for selection of routes, which chooses the path that nodes. In addition to that, we will also evaluate the reliability
has the highest residual energy with the lowest number of of the network, which will be measured based on the delivery
hops. rate.
ACKNOWLEDGMENT
This study was partially sponsored by CNPq (National
Counsel of Technological and Scientific Development)
process Nr. 135654/2008-4 and 112816/2009-6.
REFERENCES
[1] Chen, B., Jamieson, K., Balakrishnan, H. and Morris, R. Span: An
energy-efficient coordination algorithm for topology maintenance in ad
hoc wireless networks. In 7th Annual International Conference on
Mobile Computing and Networking (July 2001).
[2] Shah, R. C. and Rabaey, J. Energy aware routing for low energy ad hoc
sensor networks. IEEE WCNC (2002).
[3] Lee, Sung-Ju; Hsu, J.; Hayashida, R. Gerla, M. et al. Selecting a Routing
Strategy for your Ad Hoc Network. University of California, Los
Angeles. Computer Communications, Vol. 26, Issue 7, pp 723-733, May
2003, Elsevier Science.
[4] Y.-B. Ko and N. H. Vaidya. Location-aided routing (LAR) in mobile ad
hoc networks. Wireless Networks, Hingham, MA, USA, v.6, n.4, p. 307–
Figure 8. Cost of receiving messages 321, 2000.
[5] Y.-B. Ko and N. H. Vaidya. Location-aided routing (LAR) in mobile ad
The Figure 8 shows that the cost of receiving messages is hoc networks. In ACM/IEEE International Conference on Mobile
reduced by approximately 50 percent, using our proposal. This Computing and Networking (MobiCom’98), October 1998.
is due to the reduction in the number of control messages and [6] C. Intanagonwiwat, R. Govindan, and D. Estrin, “Directed Diffusion: A
scalable and robust communication paradigm for sensor networks”,
the selection of routes that have a smaller number of hops with Proceedings of ACM MobiCom'00, Boston, MA, pp. 56-67, Aug. 2000.
higher residual energy. [7] Y.-B. Ko and N. H. Vaidya. Optimizations for location-aided routing
(LAR) in mobile ad hoc networks. Technical Report 98-023, CS Dept.,
Texas A&M University, November 1998.
V. CONCLUSION [8] Y.-B. Ko and N. H. Vaidya. Location-based multicast in mobile ad hoc
We presented, on this article, a routing algorithm that aims networks. Technical Report 98-018, CS Dept., Texas A&M University,
September 1998.
at reducing the energy consumption of the network. The [9] I. F. Akyildiz, W. Su, Y. Sankarasubramaniam, and E. Cayirici, “A
proposal consists in the modification of the Directed Diffusion survey on sensor networks”, Communications of the ACM, pp. 102-114,
protocol. Two modifications were implemented in the original Aug. 2002. A.P.R. Canázio, Automatic Cork Processing System,
Reference Manual. Addinson Wesley Publishing Company, 2002.
protocol: the first one is a geocast approach and the second
[10] Teixeira, I., de Rezende, J. F., and Pedroza, A.C.P. - "Wireless Sensor
one is a mechanism for selection of routes. The modifications Network: Improving the Network Energy Consumption", in XXI
have the purpose of reducing the flooding caused by the Simpósio Brasileiro de Telecomunicações - SBT'2004, Belém, Brazil,
Directed Diffusion in order to recover broken paths and to September 2004.
[11] Sinalgo (2007). http://dcg.ethz.ch/projects/sinalgo/.
determine the best route for delivery of the collected event. [12] W. Heinzelman, A. Chandrakasan, and H. Balakrishnan,“Energy-
Efficient Communication Protocol for Wireless Microsensor Networks”.
The evaluation of the proposal, by means of simulation,
Hawaii International Conference on System Sciences, January 4-7, 2000,
demonstrated that the utilization of the geocast approach and Maui, Hawaii.
of the mechanism for selection of routes shows positive results
for reducing and balancing the energy consumption of the
network.
Conflicts in Online Social Network Privacy Policy
Analysis and Resloution
Abstract—Social network sites have the potential to OpenSocial is a set of common application programming
fundamentally change the character of our social lives, interfaces (APIs) for web-based social network applications,
both on an interpersonal and a community level. In the developed by Google along with MySpace and a number of
past few years, social network sites have become integrated other social networks. Applications that build with OpenSocial
into the daily practices of millions of users. Social APIs can be implemented into different social network sites,
applications provide a convenient approach where users which support OpenSocial APIs. Social applications are
can communicate and share their personal information. information exchange utilities, which will acquire user’s
However, privacy and information security have been information from social network sites, where users put their
information, and then provide some service base on this
concerned as a critical issue in development social
information. Social applications usually provide some privacy
applications. Policy-based management, which is a
control settings, which will be enforce into social network site
management approach, has been used to control the where the application is installed. It means if a same user
information access privilege. Users’ policies are more installs one application on two different social network sites,
flexible and ease to change. However, different policies each site will have one privacy control setting. If this
may cause conflicts that affect execution of social application is used to publish user signed information, two
applications. In this paper, we propose a semantic different sites may publish different information, which belong
temporal logic mechanism to detect policy conflict in social to a same user. This is a problem.
network environment.
Privacy concern has been a critical issue since social
Keywords-component; network sites appear. Privacy control that social network site
provide is limited to usage of social network site self. The
I. INTRODUCTION privacy control in social applications has to be set by users or
follow the social network site’s setting. Since online social
Social network sites have become increasingly popular; privacy is harder to guarantee [1], privacy concern has gotten
many people are using Facebook, Myspace, and Twitter to more attention from society. Policy-based management has
communicate with others. Social network sites are not a been used in social network sites. The privacy settings in social
platform for communication; it has become a collaboration network sites are a kind of policy rules. These policy rules are
platform, since social applications provide a great extension. easy generated into machine-readable format. Policy-based
People share their information with others on social network management is an administrative approach to manage system
sites, but the privacy control is simple and limited through a behaviors and rules within an information domain. A policy
number of selections that social network sites provide. describes the set of behaviors that a party is required or allowed
The main motivation for the users to create accounts and to take. In social network environment, these behaviors are
participate in these services is sharing their information with usually access actions. There are two main types of policies in
others for some purposes. People usually sharing two types of social network sites, one is global policy, which is used to
information: professional information and private information. control entire information set within user account, and another
In the first case, social network sites are as a utility to establish type is application policies. Application policies are used to
and entertain business task. In the second case, social network control information that will be used in application processing.
sites are used to share private information such as personal Policy conflicts can be classified into static conflicts and
contract details, personal photos and videos. However, by using dynamic conflicts. Analysis of static policy conflicts is time-
social applications, social network sites are not just used to independent. These conflicts can be resolved by simple logic
share information, but to communicate with specifically people. reasoning or compiling techniques. Dynamic policy analysis is
Sharing information has become information exchange. time-dependent, which means the applicability of policies and
information will change over time. The major issue about
573
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_100, © Springer Science+Business Media B.V. 2010
574 WU AND LIU
dynamic policy analysis is the changing information. The On online social network sites, users define policies; they
information covered by policies can be identified as different may not have ability to make sure policies on different online
types of elements. Relations between different elements are social network sites are consistent. If there are positive and
critical to conflict detection and resolution. negative authorization or obligation policies applied on the
same subjects, targets and actions, modality conflicts may arise
There are two platform-independent privacy languages: [8]. Modality conflicts can be detected through static analysis
IBM’s Enterprise Privacy Authorization Language (EPAL) [2] of the policy specifications. Once all policy-driven actions of a
and the OASIS eXtensible Access Control Markup Language system have been analyzed and these different types of
(XACML) [3] along with Privacy policy profile of XACML conflicts that can arise have been identified, it is possible to
[4]. XACML is a large, complex language, and this can make define rules that can be used to recognize conflict situations in a
formal analysis more difficult. While EPAL and XACML are policy set. Then these rules can be invoked during a conflict
very similar, and share a common authorization model, they detection process prior to policy deployment and enforcement
differ in important ways. In almost every area, the to identify potential inconsistencies. This is known as static
functionality of XACML 2.0 is a superset of EPAL 1.2 [5]. conflict detection and takes place at specification time. Most
Where the two languages differ, the EPAL differences usually existing work in detecting conflicts are related to modality
result in less functionality than XACML has. In many cases, conflicts. But sometimes application-specific conflicts cannot
the EPAL 1.2 differences from XACML make construction of be detected directly from policy specifications without
flexible or scalable privacy policies impossible or difficult. additional information. Conflicts may arise from the semantics
However, there is not a specification for social applications. In of the policies as well. Thus, a metapolicy [9] concept is
this paper, we will use a simple XML format to present introduced to represent this additional information, which is a
privacy policies. constraint or context associated with related policies. Yet there
is still a possibility of conflicts even when there is not any
II. RELATED WORK shared subject, targets or actions between two policies. There
may be implicit relationships between subjects, targets and
Privacy within online social network sites is often not
actions when policies are applied to them. Thus a metapolicy-
expected or is undefined [6]. Dwyer [1] gives us an image that
based approach cannot resolve all conflicts, because an
different online social network sites make users have different
administrator would have to specify metapolicies for all
privacy concern level. The social application, which is an add- possible runtime conflict situations. It is almost impossible to
on for social network site, also affects the reputation of social predict all implicit relationships that may cause conflicts in the
network sites. In another hand, the social application is installed system by a user.
into user’s interface (social applications will retrieve user’s
identify information and set up a connection between user and In classical logic, propositional logic is based on a set of
application server if it is needed). The social application always elementary facts by using a set of logic operators. It indicates a
has more privilege to access user’s information, like Boolean value set. First order logic (FOL) [10] is an extension
Facebook’s APIs give developers more rights to access user’s of propositional logic. Higher order logic [11] is an extension
friends list. In this case, social applications can get more of first order logic. Temporal logic [12] is an extension of
information than user allowed. Therefore, privacy concern on classical propositional logic, which is built for representing the
social applications is more than online social network sites set of domain elementary facts by using a set of logic operators
themselves. [12]. And ontology [13] is an explicit specification of a
Police-based systems are suitable solutions for automated conceptualization, which represents a set of entities and the
online collaborations across domain boundaries, but policies relationships between these entities in a specific domain.
become more and more complex for cross-domain activities, Entities are abstracted from real world objects in that domain;
which require more adaptive management. Since a request or relationships are contacts or effects between objects. The
an action from a foreign domain may not have an exact match ontology describing this type of information uses a set of
with an existing policy, several compatible policies can be formal terms and operators.
applied to it. Choosing one of these policies may lead to a
conflict with another related policy or violate the principle of III. POLICY MODEL
least privilege [7]. Even a similar request or action with
different parameters may lead to difficulties when applying a 3.1 General Policy Model
previously compatible policy. In general, whenever multiple As a policy can contain one or more rules, which describe
policies apply to an action or request, there is a potential for relations between policy executants and requester (subject and
some form of conflict, but it is essential that multiple policies object), we focus on a simple policy that contains one rule. For
should be applied in order to cover the diversity of management more complex policies, they can easily be divided into simple
needs and of different groups of users. There may be different policy segments. A policy segment describes a behavior and
policies related to security, privacy, or other management attributes about this behavior. For each behavior, there is an
requirements, which are applied to different aspects reflecting executor or a type of executors, an object or a type of objects,
different management needs. So policy conflict analysis and some constraints, which limit and describe the action. So
(detection and/or resolution) becomes an essential part in in each policy, there are four major components: subject,
management of policy-based systems. object (resource), action and context constraints. A policy
segment describes a complete behavior, which contains subject,
CONFLICTS IN ONLINE SOCIAL NETWORK PRIVACY POLICY 575
object, action, and context constraints. These elements are <?xml version="1.0" encoding="utf-8"?>
consisted of attributes, where detail information is stored. The <Policy> Policy 1
policy’s subject is a set of attributes (A) identifying the <Workspace>
action’s executor, it can be denoted as s={A}; the object is a <Update>
set of attributes (A) that describes an action’s target, it can be <Permit>
denoted as o={A}; context constraints include other attributes <Group name="Colleague"/>
within the policy, such as network contexts, execution <User name="Allen"/>
</Permit>
environments, and etc. We distinguish this type of context
<Period start="08:00:00" end="13:00:00"/>
information to other attributes that are about subjects and </Update>
objects. A policy’s action actually represents a temporary </Workspace>
binding between subjects and objects, which can be denoted as </Policy>
A(c)=s×o. Following notations show the components of a
policy segment in this general policy model:
<?xml version="1.0" encoding="utf-8"?>
A: attribute the attributes that describe the subjects and <Policy> Policy 2
objects. <Workspace>
S: {As} the subject of a segment, the set of attributes <Update>
<Permit>
that limit policy subject.
<Group name="University"/>
O: {As} the object of a object, the set of attributes </Permit>
that limit policy objects. <Period start="12:00:00" end="17:00:00"/>
A(c)=S×O the action of segment, it indicates the </Update>
relation between subject and object. </Workspace>
Context the information that hide in the network or </Policy>
environment.
In this policy model, each element is denoted in a set of In policy 1, the subject has attributes, which is role
attributes. In another words, a policy is a set of attributes with information. It indicates the request from the “Colleague”
different constrain conditions. group members and specific user “Allen” will be permitted
within a time period. In policy 2, the subject has the same type
attributes, which is also role information. It indicates the
3.2 Privacy Control Policy Model request from the “University” group members will be permitted
A privacy control policy should contain a resource (object), within a different time period compared to the first policy.
a requester (subject), an action and a set of information that Actions in these two policies are the same. The action reflects
limited these elements. The subject is the information requester, the relation between subjects and objects. In this case, action
who wants to access the information; the object is the resource “Update” is showing that resource will be changed by the
or the information; the action is usually access or other action subject. Obviously, there is a conflict between these two
that requester want to perform. Therefore, the privacy policy policies.
model can be presents as follow: Between two policies, the dynamic conflict will involve any
s = {As} s: information requester. element within those policies. We generated some situations
o = {As} o: the user’s information. that can be used in conflict analysis:
A(c) = s×o A(c): the action that requester a) Identity may be replaced by role information to act as
want to perform. the subject ;
Context: Context: the additional b) The identity information is provisional;
Additional information to describe s, o and c) The resource is provisional;
information action. d) The relationships (action) between a resource and an
Rule = Rule: a functional segment of a identity are provisional;
{s×o×A(c)×Context} policy e) The relationships (action) are constrained by network
Table 1 Privacy Control Policy Model or security domain contexts
A rule is a functional segment of a policy; it indicates a
condition, which is used to filter requests. Here are two privacy IV. SEMENTIC TEMPORAL LOGIC
control policies that are used to check requester’s role
information and time information. Here is an example scenario: 4.1 Temporal Logic and Ontology
an organization collaborates with a research group in a One major issue about dynamic policy analysis is that
university. They use online social network site as the information fluent will change over time. The dynamic policy
collaboration platform. The social application provides the analysis has to keep track of the change of information fluent.
information exchange functionality. Here are two policies about
the update action of different participants. It is easy to represent static states and values using classical
logic. However, time-dependent situations cannot be
576 WU AND LIU
states. In 2003, Event Calculus was proposed to analyze a [4] Moses, T., ed., Privacy policy profile of XACML v2.0; OASIS
Standard, February 1, 2005; http://docs.oasis-
combination of authorization and management policies [17]. open.org/xacml/2.0/access_control-xacml-2.0-privacy_profile-
But the authors do not identify the special capability of Event spec-os.pdf.
Calculus for transient properties in various activities. We apply [5] Anderson, A., Comparing Two Privacy Policy Languages:
Event Calculus in STL for analyzing transient properties and EPAL and XACML, Sun Microsystems Laboratories Technical
actions (relationships) in collaboration environment and find Report 2005-147, 2005;
its capability to detect and resolve dynamic conflicts. http://research.sun.com/techrep/2005/smli_tr-2005-
147/TRCompareEPALandXACML.html.
Meanwhile, static and dynamic conflict detection and [6] Dwyer, C. (2007) “Digital Relationships in the 'MySpace'
resolution in policy-based management for large open Generation: Results From a Qualitative Study.” Proceedings of
the 40th Hawaii International Conference on System Sciences
distributed systems are discussed in [18]. A general model of (HICSS), Hawaii, 2007.
security policies has been discussed in [19]. Detection and [7] Liang Chen, Jason Crampton, “Inter-domain role mapping and
reconciliation of security policy conflicts following that model least privilege,” Proceedings of the 12th ACM symposium on
are restrained by the complexity of the policy set to be Access control models and technologies, pp. 157-162, 2007.
reconciled. And only two-party conflict reconciliation can be [8] M.Sloman, “Policy Specification for Programmable Networks,”
tractable. Table 2 illustrates these comparisons. Proceedings of the First International Working Conference on
Active Networks, pp: 73 – 84, 1999.
Dynamic Policy Conflict detection [9] M. Sloman, E. Lupu, “Security and Management Policy
ity domain and resolution Specification,” IEEE Network, vol. 16, no. 2, pp. 10-19,
March/April 2002.
STL Dynamic Multiple Detection and
and resolution [10] Nir Friedman, Joseph Y. Halpern, Daphne Koller, “First-order
Static suggestion conditional logic for default reasoning revisited,” ACM
Transactions on Computational Logic (TOCL), Vol. 1, Issue 2,
FTPL Static Single Detection PP. 175 – 207, 2000
Conflict Static Single Detection and [11] Andrews P.B., “An Introduction to Mathematical Logic and
detection and resolution Type Theory: To Truth Through Proof, 2nd ed.” Kluwer
in [18] Partial suggestion Academic Publishers, PP: 201-203, 2002
dynamici [12] P. Bellini, R. Mattolini, and P. Nesi, ”Temporal logics for real-
ty time system specification,” ACM Comput. Surv., vol. 32, no. 1,
Ismene in N/A Multiple Detection and pp. 12-42, March 2000.
[19] resolution [13] B. Orgun, M. Dras, A. Nayak, G. James, “Approaches for
suggestion semantic interoperability between domain ontologies,”
Table 2 Comparison of Different Policy Analysis Proceedings of the second Australasian workshop on Advances
Mechanism in ontologies, Vol. 72, PP: 41-50, 2006
[14] A. Westerinen, J. Schnizlein, “RFC3198 - Terminology for
Policy-Based Management,”
VIII. CONCLUSIONS http://www.faqs.org/rfcs/rfc3198.html 2001
In this paper, we use Semantic Temporal Logic for [15] Edward N. Zalta, "Temporal Logic", the Stanford Encyclopedia
of Philosophy, fall 2008 Edition,
analysis of privacy policy conflict in an online social network
http://plato.stanford.edu/archives/fall2008/entries/logic-
collaboration environment. Through the experiments on our temporal/
prototype system, the improvement on capability and accuracy [16] Arun K. Eamani, A. Prasad Sistla, “Language based policy
of Semantic Temporal Logic for dynamic privacy policy analysis in a SPKI Trust Management System,” Journal of
analysis is confirmed. A comparison of different policy Computer Security, Vol. 14, No. 4, pp. 327-357, 2006.
analysis mechanisms shows that the ontology-based policy [17] A.K. Bandara, E.C. Lupu, and A. Russo, “Using event calculus
analysis system is more flexible, accurate, and applicable. to formalise policy specification and analysis,” Proceedings of
IEEE 4th International Workshop on Policies for Distributed
Systems and Networks, pp. 26- 39, 2003.
REFERENCES [18] Nicole Dunlop, Jadwiga Indulska, Kerry Raymond, “Methods
[1] Catherine Dwyer, Starr Roxanne Hiltz, Katia Passerini , “Trust for Conflict Resolution in Policy-Based Management Systems,”
and privacy concern within social networking sites: A Proceedings of 7th International Conference on Enterprise
comparison of Facebook and MySpace”, Proceedings of the Distributed Object Computing, pp. 98-109, 2003.
Thirteenth Americas Conference on Information Systems, [19] Patrick McDaniel and Atul Prakash, “Methods and limitations
Keystone, Colorado August 09 - 12 2007. of security policy reconciliation,” ACM Transactions on
[2] IBM, Enterprise Privacy Authorization Language (EPAL), Information and System Security, Vol. 9, No. 3, pp. 259-291,
Version 1.2, 2003; 2006.
http://www.w3.org/Submission/2003/SUBM-EPAL- 20031110/.
[3] Moses, T., ed., eXtensible Access Control Markup Language
(XACML), Version 2.0; OASIS Standard, February 1, 2005;
http://www.oasis- open.org/committees/
tc_home.php?wg_abbrev=xacml.
Developing a Feedback Model for Managing
Congestion in Packet Switched Wide Area Networks
By
579
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_101, © Springer Science+Business Media B.V. 2010
580 OGBIMI ET AL.
2.2 A new feedback scheme acknowledgement will indicate that the packet has been sent
The new scheme being proposed was called dynamic explicit to the receiver. Some of the packets could be lost, delayed.
feedback model which will keep the bandwidth stable. It will The feedback controller passes a feedback message to any of
reduce packet drops to its barest minimum. It will also reduce the routers that there is congestion somewhere along the link,
bandwidth delay in packet switched WAN. It will balance the then the router adjusts the speed of transmitting packets. The
flow of transmission through paths. A feedback mechanism is stability of flow of packets will then be regulated.
added to monitor the acknowledgement of packets from
receiver to sender. It will also make bandwidth utilisation
high, converge to fairness and minimize amplitude of
oscillations and maintain high responsiveness.
3 METHODOLOGY
A feedback approach was generated by designing a feedback
model for congestion control in packet switch WAN. The
principle adopted was control theory. Control theory was
used in this case to assess the stability of flow of bandwidth
in a packet switched wide area network, when congested and
when not congested. The stability of the bandwidth was
determined by finding the transfer function of the flow of
bandwidth of the network. The transfer function is done by
finding the feedback gain of the routers in the network. In
this paper, three major things were done. Firstly, a fluid flow
model for packet switched WAN was developed using
queuing model. Secondly the flow of bandwidth will be
optimized and the stability was checked. Thirdly, the
conditions for stability in a packet switched WAN were
established. The result was validated using MATLAB. The
model was implemented in MATLAB.
3.1 Control Theory 3.3 Queuing Theory
Control theory is defined as the mathematical study of how to Queuing theory is the mathematical study of waiting lines.
manipulate the parameters affecting the behaviour of a Queuing theory is applicable in packet switched WAN
system to produce the desired or optimal outcome [21]. because queues can be chained to form queuing networks.
The main conceptions of control theory are dynamic systems, The use of queuing in packet switched WAN allows the users
desired outcome, its state and feedback. Control theory was to wait for their request from any of the web servers to be
concentrated mostly on the problems of stabilization. completed and idle routers will be available for packets to be
The stability of the bandwidth was evaluated using feedback distributed.
control theory. This was done by finding the transfer function 3.4 Notation
of the arrival rate of packets. Transfer function is defined as Kendall’s notation was adopted in this paper. It is notated as
the ratio of the output of the system to the input of the system AפBפC; it denoted a queuing system having A as interarrival
[17]. time distribution, B as service time distribution and C as
Mathematically, Transfer function is defined as number of servers or routers [20].
H(s) = (3) A convenient notation for summarizing the characteristics of
queuing situations in a system is given by the following
Where Y(s) = Output function of the system in Laplace
format: (KפפC): (GD[ ) פ פ19], where:
domain
X(s) = Output function into the system in Laplace domain K = arrival distribution
3.2. Model Construction M = Service distribution
The figure below shows the block diagram which illustrates C = Number of servers in the network
what this study was set out to achieve. The diagram shows GD = General discipline (all types of discipline)
transmission of packets from source (s) to destination using a First = Indefinite number of network users accessing the
feedback model. A message is transmitted to the first router network
which acts as a buffer. The message remains in the buffer for Second = Capacity of resources accessed or downloaded
a while until there is an available router to send packets until from the web
the packets get to their destinations. The message is broken The steady-state measures of performance in a queuing
into packets by means of packetization. The packets are then situation are:.
distributed to the second and then third routers. There could ૃ = rate of arrival of packets
also be many routers depending on the topology of the μ = service rate (rate of transmitting packets)
network and the environment the network is used. An Ls = Average number of users on the network
DEVELOPING A FEEDBACK MODEL FOR MANAGING CONGESTION 581
y = f (ૃ) + f '(ૃ) (x – ૃ)
at equilibrium point, f(ૃ) = 0
Y(s) =
If y = 0, then (μ - )3 (17)
= ૃ(
x=ૃ
y = f(x) + f '(x)(x – x0)
(15)
(μ - ૃ)3
if x0 = 0, then
y= (16)
(μ - ૃ)3
4 RESULTS AND DISCUSSION
Y(s) = 4.1 Bode Plot
A bode plot is a graph of the logarithmic of the transfer
(μ - )3
function of a linear, time invariant system versus frequency,
sY(s) =
plotted with a log-frequency axis, to show the system’s
s(μ - )3 frequency response. It is a combination of bode magnitude
DEVELOPING A FEEDBACK MODEL FOR MANAGING CONGESTION 583
Am p litu d e
be phase-shifted , for example a signal described by: A sin
(wt) may be attenuated but also phase-shifted. -8
Bode Diagram
-60
-100
-80
System: h2
M agnitude (dB)
-160
-90 0
Phas e (deg)
System: h2
Frequency (Hz)
-90 Phase (deg): -1.
Phas e (deg)
-180
-180
-270
-1 0 1 2 3
10 10 10 10 10
-270
Frequency (Hz) 0 1 2 3 4
10 10 10 10 10
Frequency (Hz)
Figure 4.1.1: Bode plot of transfer function (h1) with a service rate of 100
and probability of number of users at 0.1. Figure 4.1.2: Bode plot of the transfer function with a
h1 = -20 s service rate of 200 and probability of 0.1.
-------------------------------------- h2 = -80 s
S3 - 300 s2 + 29980 s – 1000000 -----------------------------------------
s3 - 600 s2 + 119920 s – 8000000
584 OGBIMI ET AL.
Figure 4.2.2: The impulse response of the transfer function number of users were varied. The bode plots and impulse
with a service rate of 200 and probability of number of users responses were drawn with MATLAB 7.0.
in the network was 0.2
REFERENCES
[1] Borisenok, S.V., (2005), Modern Control Theory and Its Application;
Impulse Response Department of Physics, Herzen State Pedagogical University, Moika
0
River Embankment, 48, 191186, St Petersburg, Russia.
System: h2 [2] Bosco, H.L. and D.C. Dowden, (2000), Evolution of Wide Area
Time (sec): 0.00401
-5
Amplitude: -1 Networks, Bellab Technical Journals.
[3] Bose, S.J., (2002), Chapter 1 – An Introduction to Queuing systems:
Kluwer/Plenum Publishers.
-10 [4] Daintith, J. (1992), Dictionary of Mathematics Series; International
Book Production, London.
Amplitude
Abstract- Over the last decade, the technological environment The industry has developed solutions for connection
has traveled paths of updates and consolidated the emergence of between IP and PSTN, but the vast majority are proprietary
proposals for new architectures, aimed at integrating the various solutions and almost entirely be based on the binomial SIP-
networks in the world of telecommunications and building
ISDN. We can mention [3] [4] [5] as proposals for developing
synergy between data, voice and image. The NGN networks are
services, standards and integration. In the industrial field, we
based on the exchange of data packets, which permit a flexible
and scalable environment where the key factor is the provisioning can mention the work [6] as reference produced by a company
of new applications, enabling widespread mobility and integration that has contributed to the development of gateways aimed at
of new services with legacy networks. integrating the IP world with the traditional telephone networks
as well as the insertion of such devices to the new paradigm of
To face with this reality, we proposed a solution which is networks, the IMS (IP Multimedia Subsystem). Such
adherent to the functional architecture of NGNs. According to the implementations have proven the existence of recurrent
layered model, at the application layer, we developed five
challenges that need to be addressed, especially if the focus
applications that permit the implementation of new services; at
moves towards the core of the majors telecommunications
the control layer we use the open ip telephony platform asterisk
and finally at the transport/access layer, our solution support the operators and consequently the provision of new services to
following protocols: SIP, IAX, ISDN, R2Digital and SS7 Over IP. their customers. One of the major challenges is being able to
keep parity between the existing legacy networks and integrate
In order to validate our proposed solution, we implemented a new services from the new networks with ease and speed of
Voice Gateway that provides an integration between IP-PSTN development for these structures.
networks.
To face with this reality, we proposed a solution which is
I. INTRODUCTION adherent to the functional architecture of NGNs. According to
the layered model, at the application layer, we developed five
Over the last decade, the technological environment has
applications that permit the implementation of new services; at
traveled paths of updates and consolidated the emergence of
the control layer we use the open ip telephony platform asterisk
proposals for new architectures [1] [2], aimed at integrating the
and finally at the transport/access layer, our solution support
various networks in the world of telecommunications and
the following protocols: SIP, IAX, ISDN, R2Digital and SS7
building synergy between data, voice and image.
Over IP.
Telecommunications operators have made progress in
modernizing it's network elements, trying to keep existing II. NGN Networks
equipments and legacy networks and integrating this context
The NGN networks are based on the exchange of data
with the new paradigm of next generation networks.
packets, capable of providing telecommunications services
Small and medium-sized enterprises end up getting on making use of multiple bandwidths [7]. Through technologies
the margins of these new structures due to the high cost of that incorporate parameters of quality of service, the functions
equipment required for this deployment, as well as the value of of service delivery and transport are treated independently. The
service provided by operators is high because the need for key factors sought by next generation networks is that the
provision of dedicated access so much to voice as for data. provision of new services, enabling widespread mobility and
When physical access does not become the point of difficulty integration of services with legacy networks.
involved, the price for adding new services is an important
The work[8] is an example which we can see the
factor to be increased by the entry of small enterprises in this
different aspects and workforce for NGN. We focused our
new technological order.
solution in the ITU-T standardization.
585
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_102, © Springer Science+Business Media B.V. 2010
586 SILVA AND FILHO
The ITU-T (International Telecommunication Union - developed. This layer do the sharing of control functions of the
Telecommunications) is the international body that initiated transport layer, and control functions of the application layer.
studies on standardization of next generation networks. These Availability of network resources, call setup, user terminals
activities began in 2003 and strengthened in 2004 with the registration, accounting and authorization are examples of
formation of FGNGN (Focus Group New Generation Network) functions performed by the control layer.
[9]. Is important to stress here, studies and referrals decided,
the break occurring at seven specific areas of study, namely: The operation of applications like Charging and
capacity, architecture and functional requirements, quality of Billing are managed and controlled in this layer and more
service, control, security, migration of existing networks to details about this kind of control can be found on the
NGN and requirements for future packet-based networks. The recommendation M3060 of ITU-T.
definition of the basic architecture, interface specifications and
The main focus of application layer is to adding value
implementation guidelines make up the document that became
to architecture, through the possibility of the provision of new
the ITU-T recommendation for NGN[10].
services to existing IP networks and legacy networks.
According to the work [11], we elaborate the Fig. 1
The work [12] is an excellent example of
below, that represents a summary of the functional architecture
consolidation and standardization of the services, permitting a
of NGN and a basic description of the functionality performed
better use of the integration of the telecommunication
by each layer.
environments and patterns of development of the sphere of
information technology.
On the layered model used by the architecture of next III. OUR SOLUTION
generation networks, the transport layer has the responsibility
to perform all connections between existing different networks, In our solution, the Gateway is fully implemented
protocols and interfaces. based on open libraries and open platforms. The Fig. 2 below,
shows the adherence referred to the functional architecture of
The provisioning of these various access networks, NGNs, summarizing the resources used and the applications
management and control of information are functions that are developed.
shared directly with the control layer. The IP protocol becomes
the core technology to the world of next generation networks, For the application layer, five modules were
since most of the equipment used, regardless of its developed with focus on generating services to users of the
functionality in the layered structure, uses this pattern to Gateway.
exchange signaling.
In the control layer, we use the open source platform
The control layer is the link between the transport Asterisk, whose main objective was to perform the interfacing
layer and the features related to services and applications between the applications developed and their access networks.
VOICE GATEWAY TO IP-PSTN INTEGRATION AND NGN ARCHITECTURE 587
In the transport layer, we use standards already set and substantial amount of options that are outlined when adopting
open libraries for providing interconnection with the following this kind of platform that integrates the world of telephony with
protocols: ISDN, R2Digital, SIP, IAX2 and SS7 over IP. the world of IT. With a single application can occur a variety of
services to be provided by the corporation.
Below is the Fig. 3 that shows the flow of operation.
We mention the work [13] as an example to test the Fig. 3 Application – Automatic Dialer
use of multiple manufacturers and we showed that in our
solution the validation of several protocols used have proven In Step 01 of this flow, the Supervisor performs a load
the flexibility and integration point between the IP-PSTN on the database, with list of names, telephone numbers and ID
networks. code of each User Destination. It is noteworthy that in the
overall business where we implemented this solution, the
We developed five applications but here we will average size of these lists dial is around 3000 contacts.
describe only two applications, the libraries used on the
transport layer and the topology of the tests. In Step 02 the supervisor start the application.
The five applications developed are as follows: Agent Step 03 is the time that the Gateway will ask the
Authorization, Automatic Dialer, Charging, Callback and database a list of Target Users to be contacted.
Employee Control. We will describe Automatic Dialer and
In Step 04 the ratio is returned.
Callback applications.
In Step 05 the Gateway will hold an investigation to
The first application that we will describe is
see if there are attendants free in Call Center.
Automatic Dialer.
The sixth stage is the time when the gateway starts
Automatic Dialer is the applications developed which
dialing according to the data contained in the database and
a list with data from several users destination is passed to the
already stored in memory.
Gateway in order to be contacted. The Gateway will be
responsible for making connections and in accordance with the In Step 07 the status of the User "B" is returned to the
response, connect the attendants to the Users "B". Gateway. The possibilities for this state are: answered, busy
and unavailable. In this flow we are describing the status
This kind of application is widely used in corporations
answered.
that perform credit recovery of its delinquent customers. This
application can be used in cases of aggressive voice marketing In Step 08, the Gateway forwards the call to the
merge for existing products, polls to vote, research on customer Callcenter and in accordance with the strategy used, an
satisfaction, among others. operator will answer the call.
As regards the possibility of adding services to the When the call is established, occur the Step 09.
specific company, this application is an example of the
The next application that we will describe is Callback.
588 SILVA AND FILHO
The main function of this application is to provide the In Step 10, the User "B" responds with a state of 'not
return of the calls to the mobile phones that need to in use'.
communicate and pay small financial rates.
In Step 11, the Gateway informs to the Callback User
The provision of such service becomes profitable for the state of User "B" and waiting for attendance by the User
internal users within companies, specially here in Brazil that "B".
the mobile telephones rates still high.
Step 12 is the point in time when the Callback User
and User "B" remain connected.
In Step 07, the Gateway sends a dial tone to the The main work of correction was performed by
Callback User. removing a portion of the code that occurred redundant
definitions of functions and variables.
The Callback User receives a dial tone, sends the
number of the User "B" to the Gateway. This is Step 08. The feasibility of performing tests with SS7 Over IP
in an environment of integration with the carriers was not
In Step 09 the Gateway dial to the User "B" and possible due the unwillingness of the circuits for such
awaits response from their state. purposes.
VOICE GATEWAY TO IP-PSTN INTEGRATION AND NGN ARCHITECTURE 589
Below in Fig. 5 we show the topology used in The interconnection with telecommunications
laboratory to validate the signaling SS7 Over IP. operators through the use of various protocols and the
flexibility to develop new applications in a modular fashion
allows us to confirm the actual adherence of the Voice Gateway
to the stage of NGN.
REFERENCES
[02] Hyeong Ho Lee, "Signaling architecture and protocols for the Next
In corporate environments tests, we validate real-
Generation Network", Advanced Communication Technology, February 2009.
world scenarios, the operation of the applications developed
and the following protocols: SIP, IAX2, ISDN and R2Digital. [03] S. Mittal, C. Dipanjan, S. Goyal and M. Sougata, “SewNet – A Framework
The tests with traditional carriers were ISDN and R2Digital for Creating Services Utilizing Telecom Functionality, April 2008.
and with the current entrants in the emerging market for VoIP [04] NetIQ, “How to Succeed on Your Journey to Unified Communications”,
and IP Telephony, were the protocols SIP and IAX2. September 2008.
Following Fig. 6 aims to summarize the scenarios for [05] de Souza Pereira, J.H., Guilherme, J., Rosa, P.F, "Development of MGs in
integration in corporate environments tests. a next generation network with MEGACO/H.248 support", 12th IEEE
International Conference on, November 2004.
[08] Vaaramaki, T.; Puttonen, J.; Fekete, G. "Next Generation Network Related
Standardization - Enablers for the Convergence", Fourth Advanced
International Conference on, June 2008.
Fig. 6 Corporate Environment Test Topology [10] ITU-T Recommendation Y.2001, December 2004.
Abstract—In this paper, we propose a new algorithm for Thus, a multiple sink approach was proposed in order to
route selection on Wireless Sensor Network (WSN) which allows overcome that problem. In this new paradigm, multiple
communication between sensor nodes and multiple sinks. For sensors can communicate with multiple sinks (many-to-
that, the sensor nodes use a fuzzy inference machine in order to many).
select the most appropriate sink. This inference will take into
account some network characteristics such as: energy, number of
The main goal of the algorithm proposed in this paper is to
hops, collisions, and delay. Simulations have shown that our promote communication between sensor nodes and multiple
algorithm is feasible not only in terms of routes selection, but also sinks, where each node will select the best route to an
in terms of energy consumption. appropriate sink node, localized somewhere in the network.
This selection will be performed by a fuzzy inference
I. INTRODUCTION machine. We have chosen this scenario because of the
In the last years, there was a considerable increase in increasing popularity of applications which require networks
research involving wireless sensor networks (WSN), due to its with multiple sinks, for example, applications that use
applicability in many areas such as security, health, wireless multimedia sensor network [3], actuator nodes [4]
agriculture, among others. The WSN comprise a large number and other applications in which there may be more than one
of small and intelligent sensor nodes capable of detecting and base station. For example, a medical application [5], in which
transmitting characteristics from the physical environment [1]. inside a hospital we could have sensors attached to patients
Intelligent sensors can be defined as a small circuit board and the doctors would be equipped with portable computing
containing an integrated sensor capable of processing and data devices such as smartphones, which would act as sinks. Such
communication [1]. scenarios require more appropriate solutions, which make the
Wireless sensor networks have unique characteristics very existing proposals, for networks that have a single sink,
different from traditional networks, especially regarding the inefficient.
number of nodes present in the networks and constraints in In our solution, the sensors will be able to select, wisely,
some resources like energy, processing power, memory and the best route in a given time, based on some network
bandwidth. These variables make impractical the re-use of characteristics, such as distance to the sinks, the energy from
most of the algorithms developed for traditional computer the neighbor nodes, among others.
systems, since the design of any solution for sensor networks The remainder of this paper is organized as follows: in
must deeply consider the consumption of energy and Section II, we describe some related works; we present our
computational constraints. proposed algorithm in Section III; in Section IV, we perform
The sensor nodes are deployed in a geographical area in the simulation of our algorithm; in Section V, the results are
order to observe events, such as atmospheric pressure, depicted and finally, in Section VI we show the conclusions
temperature, level of pollution, etc. These nodes can be and future works.
randomly deployed or following some topology (which, in
most of the cases, depend of the network designer and the area II. RELATED WORKS
to be monitored). The sensors nodes, after sensing some In this section, we will describe some related works which
event, must be able to send a report to a particular node, called use a scenario of multiple sinks, as well as problems related to
sink (base station), which will analyze the received data and, the approach of using a single sink.
according to the network goal, make some kind of decision. In [2], we have a wireless sensor network with multiple
Some of the most common communication paradigms in sinks arranged in clusters, where each sink monitors a region
WSN involve communication of multiple sensors placed in an with multiple events. The authors propose the use of a
area of observation, reporting information to a single sink management system database, distributed in each sink of the
(many-to-one). In [2] it was shown that a single sink is a network and require two changes on the routing protocol: the
bottleneck in a network, especially for real-time applications. first is that the protocol should allow the creation and storage
591
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_103, © Springer Science+Business Media B.V. 2010
592 LEAL ET AL.
of multiple routes for the spreading of a message, and the A.General Aspects
second requires that the network provides QoS, because it In our proposal, we consider that the sensors are placed
works with the prioritization in the delivery of messages for uniformly in the region to be observed, and we also consider
notification of event. Once an event of interest is detected by a multiple sinks arranged in order to maintain coverage
node, notification message is sent toward the sink. Paths are throughout the area. The sinks are devices with features far
created and stored in each clusterhead (CH): each clusterhead more advanced than the sensor nodes, having no limitations of
stores information about the event and the ID of the cluster power.
from which the message was received. Each sink located in the network, will be responsible for
The research in [6] proposes a scheme based on fuzzy logic receiving the notification messages of events at any point of it,
to select a better way and increase the survival time in for this, each sensor node will, through a fuzzy inference
MANETs. In that work, it has been mentioned that the machine, select the most appropriate sink for it in that
survival of the network can be achieved when security moment, that is, which the best route is taking into account the
measures, preventive, reactive and tolerant lines of defense, network characteristics, such as energy, number of hops,
work together. This proposal presents the following collisions and delay. We should emphasize that this study
contribution: a scheme of selection of adaptive path considers only some of these characteristics, but we can add
independent of protocol. Furthermore, the correlation of new features according to the needs of the network.
characteristics of network security and discretion are made It is important to emphasize that we need to select
through the use of fuzzy logic to improve the path selection previously a routing protocol, because our solution works in
minimizing security limitations found in routing protocols. conjunction with it. The choice of the routing protocol will be
The process of selecting path involves three phases: data made according to the needs of the network designer, and it is
collecting, the process of fuzzy inference and finally the only required that this protocol be capable of providing multi-
selection of adaptive path. path.
The work in [7] investigates use of fuzzy inference in order
to make the nodes select the best path in a flat wsn with only B.The proposed solution
are sink. The fuzzy inference system selects the best candidate We will now discuss the steps used in our algorithm to
node basead on some metrics, like, for example energy. To choose the best path for data transmission over the network.
carry out this goal some fuzzy rules are implemented, which Table I shows the notations used, while Algorithm 1 presents
use as metrics: the distance from the candidate node from the the proposed algorithm.
origin, the distance from the candidate node to the shortest
TABLE I
path, energy and link usage. One of the disadvantages of this
NOTATIONS OF THE PROPOSED ALGORITHM
protocol is the fact that it does not address the use of multiple
sinks, which makes it vulnerable to the problems mentioned ͆n The beacon message generated by the base station
above, such as bottleneck, and a much higher dissipation of n
energy of the nodes near the sink. A: The operation after the ":" The node is held in
Our proposal differs from the related works, because it was A*: A passes the message ͆n through broadcast
developed for wireless sensor networks with multiple sinks ͆n
AN: A passes the message ͆n to node N
and makes it possible to choose the best routes for sending
͆n
data to a sink. We propose an algorithm that works with the x« y The value y is stored in the variable x
routing protocol in order to assist a node in the process of
selecting the best path among the possible routes in a given
time. This process of route selecting is made by the use of a After the deployment, the nodes are initialized and the
fuzzy inference system. Every route will be analized and sinks generate a list containing the neighboring nodes (line 1
classified regarding some criteria such as energy and number of Algorithm 1). This list is used by the sinks to send a beacon
of hops. Whenever a node has to send some data, it will take message n. This process is depicted in lines 1 to 4 of
the best classified route. Hence, the network life will increase Algorithm 1. The beacon message sent by the sinks is
as bad routes (route with little energy and/or much propagated to the network through the neighboring nodes.
interference) will be avoided. Upon receiving a beacon message, a sensor node A will check
the following fields:
• Energy: this field is used to store the lowest energy
III. THE PROPOSED ALGORITHM level found in the path.
In this section, we will analyze in detail the proposed • Station_base: indicates the sink from which the
algorithm. Then, we will discuss general aspects of the beacon message originated.
algorithm in sub-section A, while in sub-section B, the • Number_hops: distance in hops from the origin sink
selection phase of the routes with the aid of fuzzy inferences of the message to this sensor node.
will be presented. • Node_predecessor: identifies the node from which
the node A received the message.
• Last_sequence: identifies the beacon message. Each
time the sinks require a route rebuilding, a new beacon
message is generated and the last_sequence field will be
AN ALGORITHM FOR ROUTE SELECTION ON MULTI-SINK WIRELESS SENSOR NETWORK 593
incremented, so that the nodes can distinguish the we will explain this calculation in more detail. Finally, the
current message from the messages sent previously. node A tests if its energy is lower than the energy stored in the
received message, if so, it updates the field energy of the
Algorithm 1: beacon message with the value of its remaining energy,
1 : Base Station n: generates V list with neighboring nodes otherwise it goes to the next step and relays the message
2 : Base Station: n belonging to V through the network.
3: Base Station n: generates (n)
4: Base Station n : n
5 : A: receives ( n ) ( n ) transmitted by neighboring
nodes
6: A: energy « read field energy ( n )
7: A: base_station: « read field base_station ( n )
8: A: number_hops « read field num_hops ( n )
9: A: node_predecessor « read field node_predecessor
( n )
10 : A: last_sequence « read field sequence_broadcast(n )
11 : A: if A does not use the route predecessor
12 : A: discard n Fig. 1. Disposal of Beacon Message
13 : A: exit
14 : A: end-if
15 : A: sequence_current = recover In our work, we have considered that each sink node sends
sequence_pkg_received (node_predecessor) beacon messages in order to make the sensor nodes update
16 : A: if sequence_current >= last_sequence their fuzzy levels in network, within a window of time ,
where is a value to optimize the lifetime of the network. The
17 : A: discard ͆n
sensor nodes in the network check their own energy for each
18 : A: exit
time window . We have also set a safety limit of the amount
19 : A: end_if
of remaining energy in the node, represented by . This limit
20 : A: FL « calculate FL (energy, number_hops)
makes the node able to report a poor energy condition at any
21 : A: update_FL_route (node_predecessor, FL)
time, without the need to wait for the next beacon message
22 : A: if energy(A) < energy( ͆n ) then
from the sinks. The value of is informed by an alert
23 : A: store energy (A) in the field energy of ͆n
message propagated through the nodes with hop = 5, avoiding
24 : A: end_if
the wasting of energy by unnecessary broadcasts. All values
25 : A * : ͆n
assigned to , and hop may also be defined by the network
26 : A: end_if
designer.
If a node D receives a beacon message whose value of field C.The Fuzzy Inference Machine
node_predecessor is different from some of the neighboring Fuzzy inference can be defined as a process of mapping a
nodes it uses as a current route, it will drop this message, as given input into an output through the use of fuzzy logic. This
illustrated in Figure 1. This measure is adopted so that node D mapping is carried out by rules that define the output
does not have to calculate a fuzzy level for a node it does not associated with a particular input [6]. For example, if the
use as a route. energy condition of a node A is bad, then do not forward
Aiming to prevent old messages from keeping on traveling packets to this sensor node. These rules have as input values
through the network, some routines are set according to lines defined by the linguistic variables, which in this work are
15 to 19 of the algorithm presented. Whenever node A represented by energy, figure 2 and number of hops, figure 3
receives a beacon message, it checks the field last_sequence and they can be altered according to the need for the network
to verify whether it refers to a new beacon message, or if it is designer. Each linguistic variables has a set of linguistic
just the spread of a message received earlier. The next step is values (fuzzy values) associated to it. As an example of
the calculation of fuzzy level (FL) of the route; this parameter linguistic values of energy we have low, medium and strong.
allows the node to select the best route in a given moment, Each linguistic variable has its membership function which
according to the last beacon message received. Each route that will convert crisp values (real values measured from the
a node can use, it is associated with a FL. Routes with a environment) into fuzzy values. And both (rules and
higher value of FL are considered preferential. The calculation membership functions) are responsible for the behavior of our
is carried out using the FL and fuzzy logic, and in our algorithm regarding the routing decision. It is important to
algorithm we have considered as parameters the linguistic point out that in fuzzy sets, the values can range from 0 to 1,
values energy and number of hops. In the sub-section below, with 1 indicating the value of absolute truth and 0 absolute
false.
594 LEAL ET AL.
Fig. 5. Death Time of the First Node As mentioned previously, in our algorithm we consider
that each sink expect a time window for sends beacon
messages in order to make the sensor nodes updates its fuzzy
Fig. 6. Death Time the Network levels. It was performed simulations with different values of
and we evaluated the performance based on death time of the
We can realize from Figure 5 that the scenario 1 (without first node of the network (Figure 8). We can see that the
fuzzy logic and without multiple sinks) has been shown more values near to the optimum for this set are between 500 and
fragile as regards the life of a node which participates actively 700 seconds.
in a route in use. We can also infer that in the scenario 2
(without multiple sinks, without fuzzy logic, and with route
balancing), represented by the dark gray column, showed a Fig. 8. Values for Time Window
gain of approximately 42% compared to the scenario 1, black
column, because the route balancing. Scenario 3 (without VI. CONCLUSION
multiple sinks and with fuzzy logic), illustrated by the light This work proposed an algorithm for path selection using
gray column, compared with scenario 1 showed an increase of fuzzy logic on WSN with multiple sinks. The main goal of the
106%, because of the possibility of using alternative paths algorithm is to extend the network time life and this can be
inferred by FL. Finally, on scenario 4 (depicted by a white achieved by making the nodes avoid bad routes while sending
column), we have the representation of the proposed solution packets through the network. We used fuzzy logic to build up
with the use of fuzzy logic and multiple sinks, in which the a knowledge base that makes possible the process of finding
gain, compared to the scenario 1, is approximately 234%. This the best routes
gain was expected because of multiple sinks, increasing the Results obtained from simulations have shown that our
number of alternative routes by which the node can use for algorithm is feasible, increasing up to 479% the lifetime of the
data transmission and the fuzzy inference which make the network, since several routes can be used for message
nodes avoid the bad routes. forwarding.
In future works, we intend to analyze the behavior of
In Figure 6, we have the representation of the time of death algorithms in more complex scenarios, and select new
of the network, time which it prevents the transmission of data linguistic variables for fuzzy inference.
due to the death of some nodes in the path of routes. The black
column represents the scenario 1 without fuzzy logic and VII. REFERENCES
multiple sinks, in which the time of death of the network
occurred after the death of the first node of the network, [1]Karl, Holger and Willig, Andreas, Protocols and Architectures for
because this scenario does not have multiple routes and Wireless Sensor Networks, John Wiley & Sons, Ltd, 2005.
[2]Boukerche, Azzedine and Martirosyan, Anahit, An Energy Efficient and
multiple sinks. In the scenario 2, shown by the dark gray
Low Latency Multiple Events' Propagation Protocol For Wireless Sensor
column, the network was active for a longer time,
Networks With Multiple Sinks, In: PE-WASUN'07, 2007, Chania, Crete
approximately 222% more than the scenario 1; this gain was Island, Greece.
achieved by the use multiple routes. The light gray column [3]Akyildiz I. F., Melodia T., Chowdhury K. R., “A Survey on Wireless
depicts scenario 3, where the network was active for a longer Multimedia Sensor Networks” Computer Networks (Elsevier), Vol. 51, no. 4,
time, approximately 330% more than the scenario 1; this gain pp.921-960, March 2007.
was achieved by the use of fuzzy logic, although the feature of [4]Pietro Ciciriello and Luca Mottola and Gian Pietro Picco, Efficient Routing
multiple sinks has not been added. We have used this scenario from Multiple Sources to Multiple Sinks in Wireless Sensor Networks, In
to assess the choice of routes from the node, and thus
596 LEAL ET AL.
Proceedings of the 4th European Conference on Wireless Sensor Networks [7]Mohammad Abdul Azim and Abbas Jamalipour, Optimized Forwarding
(EWSN), 2007. for Wireless Sensor Networks by Fuzzy Inference System, Auswireless
[5]Kim, Yung Bok and Kim, Marie and Lee, Yong Joon, COSMOS: a Conference, 2006.
middleware platform for sensor networks and a u-healthcare service, SAC [8]Sinalgo (2007). http://dcg.ethz.ch/projects/sinalgo/
'08: Proceedings of the 2008 ACM symposium on Applied computing, 2008. [9]NS2 (2008). http://nsnam.isi.edu/nsnam/index.php/main_page.
[6]Michele Nogueira Lima and Helber W. da Silva and Aldri L. dos Santos [10]Perkins, C. E. and Bhagwat, P. (1994). Highly dynamic destination-
and Guy Pujolle, Survival multipath routing for MANETs, NOMS, 2008. sequenced distancevector routing (dsdv) for mobile computers. SIGCOMM
Comput. Commun. Rev., 24(4):234-244.
Mining Mobile Sequential Patterns in Wireless Cellular Networks
Abstract— In this paper, we have taken concepts of data allocate resources required by the mobile user evenly
partitioning methods and modified the well known throughout the surrounding cells. This allows for the
algorithm SPADE (Sequential PAttern Discovery using network to be readily prepared to continue the node’s data
Equivalence classes), which we have implemented over a connection if the mobile user continues using the mobile
mining model, called mining mobile sequential patterns, device and transfers to an adjacent cell. Disadvantages to
and called MobileSPADE. From this, we have devised this strategy include producing unnecessary overhead by
two innovative forms, called Fixed-MobileSPADE distributing resources to cell towers that will not be used by
algorithm and Dynamic-MobileSPADE algorithm, which the mobile user, thus not allowing the mobile user full
will mine for mobile sequential patterns from either employment of all available resources. One alternative
fixed-length or dynamic-length transactions. In mining solution that has been extensively researched has been the
for mobile sequential patterns in a mobile commerce mining for frequent sequential patterns. In this paper,
environment, we used tower ID data extracted from a frequent sequential patterns refer to the frequent movement
dataset that was constructed and managed by the patterns of a mobile user. Finding these frequent sequential
Massachusetts Institute of Technology (MIT) Media patterns could allow for a network carrier to calculate with
Lab: Reality Mining Project, which logged objective data some probability where a mobile user will go next, based
from 100 mobile device users for a period of nine upon his/her past movement history. Using this innovative
months, to represent mobile user’s movements. mobile sequential pattern model and looking at Figure 1
Prediction analyses results were used to measure the again, when a mobile user resides in cell A, the network
performance of Fixed-MobileSPADE and Dynamic- carrier could allocate more resources of a certain service
MobileSPADE algorithms on the mining mobile requested by the user to the cell most likely predicted to
sequential patterns model. Results show that as the contain the end user. For example, based upon a mobile
quantity of items that are taken into consideration for user’s past history, if the mobile user is calculated to have a
mining mobile sequential patterns increases, the 75% chance of going from cell A to cell D, then the network
prediction accuracy correspondingly increases. can allocate/distribute more of its resources to cell D as
compared to the surrounding cells, in order to provide better
service to the end-user via a more efficient use of its
I. INTRODUCTION resource allocation methods and routing protocols.
597
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_104, © Springer Science+Business Media B.V. 2010
598 BRADLEY AND RASHAD
patterns in a cellular network and its two derivatives, patterns, because this is one of the first research projects that
algorithm Fixed-MobileSPADE and algorithm Dynamic- successfully collected an extensive amount of objective data,
MobileSPADE. Section 7 presents the experiment results as opposed to subjective data. In this paper, objective data
obtained from algorithm Fixed-MobileSPADE and refers to the data collected by the mobile device logging
algorithm Dynamic-MobileSPADE and section 8 discusses software. Subjective data refers to previous methods used to
the conclusion on the experiment results. test algorithms, which include creating a virtual mobile
commerce environment and issuing survey’s to record
relative location.
There is great potential for this dataset and other similar
II. RELATED WORK AND PRELIMINARY CONCEPTS datasets to revolutionize the whole concept of social
In this section, we analyze an experiment conducted in [1] networking and advertising. Use of this dataset provides
on the mining of frequent itemsets via a data partitioning researchers with a data model that represents objective
approach which employs a modification of the algorithm in human behavior. Prospective exploits of this dataset include
[2]. We also discuss the SPADE algorithm and which the testing and modification of current cellular network
aspects were modified/ enhanced during the development of algorithms to improve their performance, re-defining current
Fixed-MobileSPADE and Dynamic-MobileSPADE. advertising techniques to make advertising targeted and
In [1], Nguyen and Orlowska enhance the data pre- more personalized to an individual mobile user, and
processing approach of the Partition algorithm [2] to developing innovative methods to find frequent sequential
increase the performance of frequent itemset computations. patterns.
A series of data pre-processing methods such as the
introduction of new clustering and similarity algorithms
were applied across the data to show that the performance of Notation used throughout the paper is as follows:
the algorithm discovering frequent itemsets increases by • Let T = {t1, t2…tn} be a set of n transactions over data set
means of data partitioning. The Partition algorithm in [2] D.
follows the concept that a fragment P D of the database is • Transactions can be of fixed or variable length, depending
a subset of transactions within the database. With reference on the algorithm used.
to an itemset, local support is defined as the fraction of • Each transaction consists of a sequentially ordered set of
transactions containing that itemset and local candidate items {i1, i2, i3…im} Tn, referenced as tower ID
itemset describes an itemset meeting the minimum support numbers in this paper.
for the whole database. Global support and global frequent
itemset are calculated similarly except with respect to the The goal of Fixed-MobileSPADE and Dynamic-
whole database. The principle behind [1] is that each MobileSPADE algorithms is to generate all the discovered
fragment is calculated in the main memory one at a time. frequent itemsets that meet a user-defined support threshold
In [3], Zaki presents SPADE, an algorithm designed for the and confidence threshold.
accelerated discovery of sequential patterns. Previous
algorithms have required multiple scans over a database for
their calculations, but SPADE utilizes combinatorics to
decompose the original problem into smaller sub-problems, IV. ALGORITHMS FOR MINING FREQUENT MOBILE
which are individually solved in the main memory and later SEQUENTIAL PATTERNS
combined. Similar to SPADE algorithm [3], algorithm MobileSPADE
is an Apriori-based sequential pattern algorithm that utilizes
characteristics of combinatorics and adapts a vertical data
III. MOBILITY DATASET AND PRELIMINARY CONCEPTS format of the original dataset. Furthermore, two defining
It is important to discuss the mobility dataset used to obtain characteristics of algorithm MobileSPADE lie within its
performance tests results on Fixed-MobileSPADE and ability to incorporate data attributes into a consecutive
Dynamic-MobileSPADE algoritms. The dataset used in this sequentially ordered list, based upon timestamp values, and
research is a result of the reality mining project carried out its capacity to only consider itemsets of length two or greater
by the MIT Media Lab [4]. This dataset was collected over a len(itemset) 2, in order to find frequent sequential itemsets.
nine month period and involved the use of 100 cellular When using the mobile dataset from the MIT Reality Mining
devices [4, 5]. As explained in [4] mobile device logging Project, the timestamp is defined by the values in the
software was designed and installed on 100 mobile devices, “starttime” column of the dataset. A mobile user’s
which recorded critical information about the mobile user, movement path is defined as the tower ID utilized for each
such as proximity, location, services used, length of calls logged event. Table 1shows a sample of the collated dataset
made and received, type of data used, etc onto the device’s with the corresponding starttime and celltower_oid data
memory card. The data for all 100 mobile users was columns for mobile user number 94 after sorting with
collected and organized into one dataset. This dataset has reference to timestamp value. We can see the mobile user
been of great interest recently to researchers involved in utilized cell_tower(38) at timestamp(2004-07-23 12:22:34)
mobile data mining and the discovery of frequent sequential
MINING MOBILE SEQUENTIAL PATTERNS IN WIRELESS CELLULAR NETWORKS 599
1) Sort Sort(D) database by sequential order, with MobileSPADE is the greater number of sequential itemsets
reference to timestamp value. *Refer to Figure 2 considered in algorithm Fixed-MobileSPADE. This occurs
due to differences found in the candidate sequential itemset
2) Sort T into tn transactions by creating non-repeating generation phase of each algorithm. Algorithm Fixed-
MobileSPADE takes all sequential subsets of each
sequences of itemsets. For example,
transaction into consideration while algorithm Dynamic-
MobileSPADE only considers each transaction as an
3) when using the database from Figure 3, T = {t1, t2, t3, itemset. Once both algorithms have generated all itemsets of
t4, t5, t6, t7} where: length two or greater that meet the minimum support and
confidence threshold, sequence rules can then be defined,
t1 = {38, 39} based upon the sequential itemset results.
t2 = {38, 39, 40}
t3 = {38, 40, 39, 41}
t4 = {40, 39, 38, 41} V. EXPERIMENT RESULTS AND PREDICTION ANALYSIS
t5 = {39, 40, 41}
We conducted multiple experiments with algorithm Fixed-
t6 = {40, 38, 41}
MobileSPADE and algorithm Dynamic-MobileSPADE,
t7 = {40}
using different parameters, on the MIT dataset. These
experiments are performed on a 2.5 GHz Intel CPU and
4) Consider each transaction as a sequential itemset,
4096 MB of memory. There were many results collected, but
where the length of each itemset must be greater because of the limited space and for the purpose of
than two len(itemset) 2. Using the current explanation, we will only use length-2 itemset results in this
example, a list of candidate sequential itemsets will paper.
be as follows: We tested Fixed-MobileSPADE algorithm over 1000
records from the MIT dataset and let transaction_length =
t1 = (38, 39) 10, confidence_threshold = 20%, and support_threshold =
t2 = (38, 39, 40) 20%. We found that 11 length-2 itemsets met these
t3 = (38, 40, 39, 41) parameters and are now considered to be frequent length-2
t4 = (40, 39, 38, 41) sequential patterns over 1000 examples by Fixed-
t5 = (39, 40, 41) MobileSPADE algorithm. We used also the same previous
t6 = (40, 38, 41) parameters over 2000 records from the dataset. The results
show that six itemsets met these specifications and are now
5) Calculate the support support(im) of each sequential considered to be frequent length-2 sequential patterns over
candidate itemset when support(im) = 2000 records by Fixed-MobileSPADE algorithm. We tested
also the Dynamic-MobileSPADE algorithm using different
frequency_count(im)/ number_of_transactions
values for confidence_threshold and support_threshold .
where frequency_count(im) = frequency of itemset After finding the frequent sequential itemsets that meet the
tn in database D and number_of_transactions = |D| / user-defined minimum support threshold and confidence
transaction_length. If support(im) < threshold, experiments were conducted on the dataset to test
support_threshold, then remove the itemset from for prediction accuracy of Fixed-MobileSPADE and
candidate sequential itemsets. Dynamic-MobileSPADE algorithms. Using a minimum
support threshold and confidence threshold of 20% for user
94 in the dataset, results for Fixed-MobileSPADE algorithm
6) Calculate the confidence confidence(im) of each
are graphed in Figure 2, where using 200 examples yielded
sequential candidate itemset when confidence(im) = the best results. Using a minimum support threshold and
frequency_count( / frequency_count confidence threshold of 5% for user 94 in the dataset, results
where A = itemset[0] the first element of the for Dynamic-MobileSPADE algorithm are shown in Figure
itemset and ( = itemset[0:] the set of all 3, where using 200 examples yielded the best results.
items in the itemset. If confidence(im) <
We can see that, overall, Fixed-MobileSPADE algorithm
confidence_threshold, then remove the itemset
performed better for user 94 due to the fact that user 94
from candidate sequential itemsets. exhibits more frequent behavior and the Fixed-
MobileSPADE algorithm considers more candidate
End sequential itemsets than Dynamic-MobileSPADE algorithm.
We can say that Fixed-MobileSPADE algorithm was
Algorithm Dynamic-MobileSPADE is similar to algorithm designed to perform best for a user who exhibits more
Fixed-MobileSPADE in that when calculating the frequent behavior, whereas Dynamic-MobileSPADE
confidence of each itemset, the last item in each transaction algorithm was designed to perform better under conditions
is not considered. One significant difference between where the user exhibits more sporadic behavior. This can
algorithm Fixed-MobileSPADE and algorithm Dynamic- especially be seen from the experiment results for the
MINING MOBILE SEQUENTIAL PATTERNS IN WIRELESS CELLULAR NETWORKS 601
REFERENCES
[1] S. N. Nguyen and M. E. Orlowska, “A Further Study in the Data
Partitioning Approach for Frequent Itemsets Mining,” in 17th
Figure 3 Prediction accuracy using Dynamic-MobileSPADE Australasian Database Conf., pp. 31-36, Hobart, Australia, 2006.
[2] A. Savasere, E. Omiecinski, and S. Navathe, “An Efficient Algorithm
usforing Support Threshold = 5% and Confidence Threshold for Mining Association Rules in Large Databases,” in Proc. Of the
= 5% 21st Very Large Databases Conf., pp. 432 – 444, Zurich, Switzerland,
1995.
602 BRADLEY AND RASHAD
[3] Zaki, M., “SPADE: An Efficient Algorithm for Mining Frequent [5] B. Congleton and S. Nainwal, “Mining the Mine - Exploratory Social
Sequences,” Machine Learning Journal, 42(1/2), pp.31-60, 2001. Network Analysis of the Reality Mining Dataset,” School of
[4] N. Eagle and A. Pentland, “Reality Mining: Sensing Complex Social Information, University of Michigan, Ann Arbor, Michigan.
Systems,” Personal and Ubiquitous Computing, Vol 10, #4, pp. 255- [6] J. Han and M. Kamber, Data Mining: Concepts and Techniques, 2nd
268, 2006. ed. San Francisco, MK, 2006, ch. 8, “SPADE: An Apriori-Based
Vertical Data format Sequential Pattern Mining Algorithm” , pp. 502-508.
Agile Software Development Methods and its Advantages
Dr.Soly Mathew Biju
University of Wollongong in Dubai
Traditional software development methods (water- “Extreme programming” is one of the commonly
fall, spiral, Iterative) were very cumbersome and used Agile software methods [7]. It is used in the
involved a lot of documentation. These methods were problem domains where the initial requirements are
very rigid, well planned and hence provided very likely to change to a large extent.
little or no flexibility. McCauley [4] states that “the
underlying philosophy of process-oriented methods is It is based on the following assumptions:
that the requirements of the software project one are
documented, they are locked and frozen before the • Be willing to do rework frequently and fast,
design and software development process takes as well as welcome customer feedback. It
place”. This process is not flexible enough to becomes a customer focused and customer
accommodate changes in the specifications at a later driven project rather than a plan driven
stage of the development. According to Highsmith project.
and Cockburn [3], “what is new about agile methods
is not the practices they use but their recognition of • Initial software is developed based on the
people as the primary drivers of project success, stories and requirements given by the user.
coupled with an intense focus on effectiveness and This software will contain the most
manoeuvrability. This yields a new combination of important, vital feature required by the user.
values and principles that define an agile world This method is found to be very effective as
view”. it involves the customer from the very
beginning. The customer writes the user
Project development would prove to be a disaster if stories comprising simple lines of text
the customer is not sure of the final product describing the features they want included in
requirements and can give the requirement only on an the first release of the working program. The
incremental basis. It follows the CCM (Capability acceptance test for the software is designed
Maturity Model), which is process based. The “Agile and based on these user stories. This will
Method” is people based rather than plan based. In keep a check on whether all the features
Agile software development the focus is on customer requested are implemented in the program.
satisfaction. It is an incremental/iterative software
development process which focuses on quick • The developer then estimates the time for
development of a working model. the development activity for which they will
have to meet the customer, reconfirm and
603
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_105, © Springer Science+Business Media B.V. 2010
604 BIJU
• Each increment will have the developer XP undertakes the design in the following manner:
working on the analysis, design code
integration and testing phases iteratively [9]. • A system metaphor is used which names the
classes and the objects in a consistent
It follows an Iterative life cycle model with 5 manner which is agreed upon by all the team
activities: namely, Planning, design, coding, testing members. It also describes the overall
and integration. architecture.
• Planning occurs at the beginning of each • It follows the test first method where the
iteration programmer writes the code for testing the
• Design, coding, and testing are done program before writing the program. These
incrementally tests are written incrementally based on the
• Source code is continuously integrated into user requirements. These are the tests the
the main branch, one contribution at the time program unit must comply with.
• Unit tests for all integrated units. Unlike in
case of traditional software development Consider the following class Employee and its
methods regression testing is carried out at corresponding Junit test case :
all the iterations
AGILE SOFTWARE DEVELOPMENT METHODS AND ITS ADVANTAGES 605
The test case in listing1 tests whether the constructor III. SCRUM
is invoked and values are initialized correctly.
Another agile software development method is
The code must be written so that it will satisfy these SCRUM.
tests.
• Refactoring reworks the design to iteratively Scrum and Extreme Programming (XP) are clearly
adapt the changes in the environment and aligned. In fact, if you walked in on a team doing one
of these processes you might have hard time quickly
requirements. It follows group design
techniques like Whiteboard design sessions, deciding whether you had walked in on a Scrum team
Class-Responsibility-Collaborator sessions
(CRC) etc.
606 BIJU
If we are adopting agile method for the first time we [3] Beck, K. & Andres, C., Extreme Programming
will face a challenge of changing the mindset of our Explained. Addison-Wesley, 2004.
team member and helping them understand the rules
of the new method [5]. A lot of rework may be [4] Cohn, M.,“User Stories Applied for Agile Software
involved which could have been avoided if a Development” Addison-Wesley, 2004.
complete requirement analysis is done initially. If a Construx, http://forums.construx.com/ [Accessed 30th
proper order of building the software is not followed, Dec 2007]
we might end up reworking from the beginning to
include another set of requirements. This is especially [5] Evan, I. “,Agile Delivery at British Telecom”
likely if we need to add certain security or CSI communications , July 2007.
performance features to the software. The complete
process may slow down if requirement analysis and [6] IBM http://www.ibm.com/developerworks/java/
design have to be performed for all iterations. library/j-xp042203/ [Accessed 9th June 2007]
Abstract
Grid Technology is the next phase in computer infrastructure (ii) During the last years computational simulations based on
that will fundamentally change the way we think about and use the atomic description of biological molecules have been re-
computing power. RO-14-ITIM is the Grid site of the National sulted in significant advances on the comprehension of biologi-
Institute for Research and Development of Isotopic and Mole- cal processes. A molecular system has a great number of con-
cular Technologies Cluj-Napoca (INCDTIM). In this paper we formations due to the number of degrees of freedom in the ro-
present the evolution and prospects of the site, his ability to tations around chemical bonds, leading to several local minima
integrate in huge Grid networks, the hierarchical structure as in the molecular energy hyper-surface. It has been proposed
well as his key in research work of the Nord-West part of Ro- that proteins, among the great number of possible conforma-
mania research field. RO-14-ITIM site support physic domains, tions, express their biological function when their structure is
HPC and Grid computing for distributed analysis. close to the conformation of global minimum energy [3].
This type of research involves a large amount of computing
I. INTRODUCTION power and fills to be a very suitable application for grid tech-
nology.
The Grid computing vision promises to provide the needed
(iii) Over the past few years, a sizable body of experimental
platform for a new and more demanding range of applications.
data on charge transport in nanoscopic structures has been ac-
To become true, a lot of work, including design and deploy-
cumulated.
ment of adequate resource management and information ser-
We face the birth of a whole new technological area: the mo-
vices, need to be overcome. In this context we are presenting
lecule-based and molecule-controlled electronic device re-
the structure of Grid computing that applies at the INCDTIM
search, often termed “molecular electronics” (ME). The sim-
institute in Cluj-Napoca, Romania.
plest molecular electronic device that can be imagined consists
In 1996 we start working at the installation of the ATLAS
of a molecule connected with two metallic nanoelectrodes.
LHC CERN Geneva detector with the construction of parts for
There is now a variety of approaches to form such nanojunc-
the Tile Calorimeter. As a next step, studying the forecast in
tions (NJ) [4] that differs in the types of electrode, materials
Grid we made the preparation to build up our own Grid site.
employed, the way in which the molecule-electrode contacts
The work seemed easy but great challenges were to be ex-
are established and the number of molecules contacted. Recent-
pected in the next years.
ly the realization of the first molecular memory was reported
Today we apply and develop the Grid system in many re-
[5].
search areas as follows:
The computation of the current-voltage characteristic of a
(i) The fundamental research in particle physics uses ad-
realistic nanodevice needs about 6-10 GB of memory and one
vanced methods in computing and data analysis. Particle phys-
week computing time, which bring to the idea of using a Grid
ics is the theory of fundamental constituents of matter and the
environment with high processing power and MPI support.
forces by which they interact and ask some of the most basic
questions about Nature. Some of these questions have far- II. DATA ABOUT ROMANIA GRID SITES
reaching implications for our understanding of the origin of the
Universe [1]. We can speak about Grid sites in Romania after 2001, but
In 2009 Large Hadron Collider accelerator (LHC) [2] with officially on 2006 when Romania through National Authority
ALICE, ATLAS, CMS and LHCb experiments, will start tak- for Scientific Research (ANCS) and The European Organiza-
ing data. LHC will collide protons at the highest energy tion for Nuclear Research (CERN) signed a “Memorandum of
¬s=14TeV and luminosity L=1034 cm-2s-1 among all the other Understanding”, for collaboration in development and exploita-
accelerators and due to these performances, high precision tion on the Worldwide LHC Computing Grid. [6] From that
measurements will be possible and the results means new phys- time, officially we begun to work and develop the Romanian
ics. Grid sites.
To fulfil these requirements a high performance distributed In Romania we have 2 Grid Consortium, RO-Grid found in
computing system is needed. 2002 and Tier 2 Federation (ROT2F) found in 2006. The
609
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_106, © Springer Science+Business Media B.V. 2010
610 FARCAS AND POPENECIU
founders of both consortiums are the most important Institutes For the improvement of our Grid site, this first step where
in Romania which were working in Grid projects years ago, continued through other projects, consisted in purchasing mi-
and we can mention the most important: National Institute for nimal technical equipment.
Physics and Nuclear Engineering “Horia Hulubei” (IFIN-HH), In parallel we had built a location that meets the highest
National Institute for R&D in Informatics ICI Bucharest, Uni- needs for processing and storing data, which became a type
versity ”Politehnica” of Bucharest, Technical University of Tier 2 Datacenter [9] with a good security and monitoring sys-
Cluj-Napoca, Western University of Timisoara and National tem as well as a backup system assured by a 220 kVA genera-
Institute for R&D of Isotopic and Molecular Technologies tor.
(INCDTIM). RO-Grid most important objective is to encour- For the final stage we continued to equip our system with a
age and facilitate the involvement of other interested and com- high capacity storage unit and many more computing servers
petent national institutions and ROT2F purpose is to provide which where included in the INCDTIM grid site, RO-14-ITIM,
the computing resources needed to process and analyze the because the grid system is needed to store and analyze big
data gathered by the Large Hadron Collider experiments. amount of information.
Among the first site’s in Romania were RO-01-ICI and RO-
02-NIPNE and the last one that have a contribution to WLCG IV. NOWADAYS GRID SITE RO-14-ITIM
is RO-14-ITIM. [7]
Through the “Memorandum of Understanding”, we assumed
III. FIRST STEPS TOWARD INCDTIM GRID SITE responsibilities to provide computing power and storage capac-
ity in collaboration with all other WLCG sites from Romania.
As part of the WLCG Romanian group, and signer of the The resources that all this system needs are: high speed con-
“Memorandum of Understanding” we started the building of nectivity, both local and wide area; huge computer throughput,
RO-14-ITIM site with 2 servers and a 10 Mbps link in 2005, great off-line data storage; human resources to manage local
through the first CEEX project between INCDTIM and IFIN- configuration and to implement GRID specific activities.
Nowadays at INCDTIM Cluj there are 3 main directions for
HH. INCDTIM site was integrated and tested as an extension
processing data:
of the IFIN-HH ATLAS-GRID site. The following tests have
• ATLAS data processing in WLCG project;
been done [8]:
• Numerical simulation of computational physics
(i) Transmission speed measurements between IFIN-HH and
with application in biophysics and nanostructures;
INCDTIM during sustained high speed data traffic, shown in
• Simulation in environment domains.
Fig. 1, was obtained using NFS copy. We can see a limited 10
(i) LCG project - LHC Computing Grid - was launched in
Mbps link, which is not satisfactory for GRID production pur- 2002 and it has the goal to make a distributed GRID to be used
poses. by ALICE, ATLAS, CMS and LHCb experiments. Till now we
(ii) Entire system functionality tests as a single grid configu- have in Romania 2 sites that is working for the ALICE project
ration, task distribution and execution real speed depreciation and 3 sites that are working for the ATLAS project. The
measurements between the remote nodes. amount of data that CERN has to process is big enough so that
only one site in the world could not work for. The computing
resources estimated to be necessary are presented in Tabel.1
[10].
Beside LCG project we intend to use our grid site for other
inner projects. Both of them are connected with the numerical
simulation of computational physics with direct application in
biophysics and nanostructures. As a general rule, the computa-
tional techniques used by numerical simulations of computa-
tional physics are based on the use of Unix-like operating sys-
tems. The codes used are mostly developed in Fortran. A vast
majority of these codes are developed under GNU license
Fig. 1. Traffic measurement between IFIN-HH Bucharest and INCDTIM (such as ABINIT) or academic license (SIESTA, GAMESS
Cluj on April 2006
TOWARD A HIGH PERFORMANCE GRID SITE RO-14-ITIM AND THE CONTRIBUTION TO WLCG 611
Traffic measurement
500
400
S peed [Mbps]
300
200
100
0
0 4 8 12 16 20 24 28 32 36 40 44 48
Time [h]
The statistics for the last 4 month are promising, there where
over 500 job done, as shown in the Fig. 7. Fig. 9. INCDTIM Datacenter
VI. CONCLUSIONS [5] J.E. Green, et all, Heath: A 160-kilobit molecular electron-
ic memory patterned at 1011 bits per square centimeter,
RO-14-ITIM site was designed and developed along almost Nature 445(7126), 414 (2007).
5 years, from a basic structure, and with a continuous effort, to [6] Memorandum of Understanding for Collaboration in the
a competitive structure, well integrated in TIER-2 Federation. Deployment and Exploitation of the Worldwide LHC
The RO-14-ITIM contribution to WLCG with his 18.567 Computing Grid, CERN-C-RRB-2005-01, November
CPU time [13] which represents 0.45% of the whole time dedi- 2005.
cated by all 5 sites to the CERN project and 3.55% for the [7] F. Farcas, I. Szabo, A. Bulinschi, L. Miclea, A. Bot ,
ATLAS project. World Wide Romania Grid Database, 3rd International
We interact with new partners in international and national Workshop on High Performance Grid Middleware, 17th In-
programs and projects, and this give us the opportunity to de- ternational conference on control system and computer
velop our specialists and infrastructure of grid system, to reach science, 26 – 29 May 2009, Proceedings, CSCS – 17 Vol
our mission: INCDTIM Datacenter, to be the most important 2, pp 461-464, Copyright © 2009, Editura Politehnica
Grid centre in Transylvania capable to sustain the Nord – West Press
research and economical area of Romania. [8] C. Alexa, V. Boldea, M. Ciubancan, . Constantinescu, S.
Di¯, D. Pantea, T. Preda, M. Jalobeanu, G. Popeneciu,
Acknowledgments IFIN-ITIM Distributed Grid System, ISPDC 2006, Pro-
ceedings 5th International Symposium on Parallel and Dis-
This work is supported by the Romanian Research and De- tributed Computing, Timisoara Romania, July 6-9, 2006,
velopment Agency through EU12/2009, EU15/2009, POS-CEE pp 215-219, IEEE Copyright 2006.
Ingrid/2008 projects. [9] TIA – 942, Data center standards overview.
[10] J. Knobloch and L. Robertson, LHC Computing Grid,
References Technical Design Report, CERN 2006.
[11] F. Farcas, G. Popeneciu, A. Bende, C. Morari, S. Belov, L.
[1] ATLAS Collaboration, Exploring the Mystery of Matter, Miclea, ITIM Distributed Grid System applied in high
Papadakis Publisher, UK, 2008. energy, biomolecular and nanotehnology physics, IEEE
[2] LHC Homepage, http://cern.ch/lhc-new-homepage/. International Conference on Automation, quality and test-
[3] S. P. Brown, S. W. Muchmore, J. Chem. Inf. Model., 46, ing, robotics (AQTR) May 22-25 2008, theta 16th edition,
999 (2006). vol III, Proceedings pp 343-346 2008.
[4] A. Salomon et all, Comparison of Electronic Transport [12] https://goc.gridops.org/site/list?id=1655004
Measurements on Organic Molecules, Adv. Mater. 15(22), [13] https://www3.egee.cesga.es/gridsite/accounting/CESGA/
1881-1890 (2003). tier2_view.html
A MFI4OR-based Approach of Semantic Annotation
for Web Service
Cheng Zeng Keqing He, Bing Li, Chong Wang, Yangfan He
1
State Key Laboratory of Software Engineering State Key Laboratory of Software Engineering
Wuhan University Wuhan University
Wuhan, P.R. China Wuhan, P.R. China
2
College of Mathematics & Computer Science E-mail: [email protected]
Hubei University
Wuhan, P.R. China
E-mail: [email protected]
Abstract—Compare to the traditional annotation is enriched with the adoption of semantic annotation to
technology for web service, a novel approach for the semantic improve the flexibility of the WSR and help users with
annotation to web service based on MFI4OR is proposed in solutions (services) that are close enough to what they
the paper. This proposal permits multi-ontology based would have liked to get (even if they do not fully match
annotation to web service in the same domain which will lead their expectations). Therefore, how to annotate these items
to the flexible annotation, and then the flexible publication becomes a key point in the issue of semantic annotation. In
and discovery of web service in Web Service Registry (WSR). recent years, a large number of research based on semantic
The result of experiment illustrates the efficiency of web techniques have been proposed, such as the efforts around
service discovery and the degree of satisfaction to users in
OWL for Services (OWL-S) [8], the Web Services
WSR is highly improved.
Modeling Ontology (WSMO) [9], WSDL-S [10] and
Keywords-Semantic Annotation, Web Service, Domain METEOR-S [12].
Ontology, MFI4OR In this paper, a novel approach for the semantic
annotation for web service based on ISO SC32 19763-3:
I. INTRODUCTION MFI4OR (Metamodel Framework for Interoperability:
With XML based standards like UDDI, WSDL and Metamodel for Ontology Registration) [4] is proposed,
SOAP, web services play a key role in universal which will make semantic annotation more accurate and
connectivity and interoperability of applications and efficient. The remainder of the paper is organized as
services [1]. With the growing popularity of Web services, follows: Section 2 presents the traditional model of
there arise issues of finding relevant services, especially semantic annotation. In Section 3, we provide some
with the possibility of the existence of thousands of Web improved methods of semantic annotation by use of domain
services. Web Service Registry (WSR), which acts as an ontology. Section 4 proposes the MFI4OR-based approach
information mediation role in the triangular SOA, is to of semantic annotation. The related work is discussed in
solve this problem. In WSR, service providers can publish Section 5, followed by the conclusion and future work in
service description information more effectively and Section 6.
service consumers can discover appropriate services more
accurately and easily. However, currently web services are II. TRADITIONAL MODEL OF SEMANTIC ANNOTATION
described as WSDL descriptions, which provide WSR is supported by domain ontology for semantic
operational information such as Inputs and Outputs etc. It is publication and discovery of the Web services. The
a type of syntactic level description and its interface domain ontology is created from concepts and
parameters do not have meaningful names, which obstructs
terminologies that are likely to be used by Web services in
service discovery and composition. For example, one
a particular domain.
service A has an output item which use specialty as name to
describe what field an expert engages in, while another For example, in the electronic commerce domain, the
service B may use research field for describing such an domain ontology can consist of concepts from the ebXml
item. As we all know, specialty and research field represent Core Component Dictionary. Another example could be
the same information item, but they are different at domain ontology for the Traffic domain which may
syntactic level, and same at semantic level. Therefore, the include concepts like Car, CarPrice, ArrivalCity and
key to solve the limitation of keyword-based searches is to DepartureCity. As a result, before the semantic annotation
add semantic annotations to Web service specifications [2]. of web service, we must choose a concrete domain
ontology. After that, we can manually use the mapping
Domain ontology is used to add semantic information between these items in WSDL file (mainly refers to WSDL
on WSDL interface parameters to overcome the limitations
interface parameters of Web Service such as Inputs,
of a syntax-based search [3]. The discovery of web services
615
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_107, © Springer Science+Business Media B.V. 2010
616 ZENG ET AL.
Outputs and Operations) and the selected domain ontology includes terms related to a given application domain. We
to capture the meaning implied by the Web service assume that this ontology can be built by domain experts
provider in that domain. Finally, the mappings are written who analyze the terms included in the Web services
to the WSDL file as annotations. The annotated WSDL file published in the registry. The latter includes all the possible
along with the original WSDL file is shown in figure 1. terms (at this stage we use Wordnet).
However the manual process of annotation can not assure We assume that programmers of Web Services have a
the accuracy of semantic annotation. Section 3 is ready to general understanding about the domain which Web
overcome this problem. Services belong to, such as the Electronic Commerce
domain or the Traffic domain. Therefore, when defining the
names of operations and parameters in Web Service, they
may use the concept in the domain, which can avoid the
naming at will. When annotating the names of operations
and parameters, our algorithm will realize the following
two steps: (1) Making a match between these names and
the concept items in the domain ontology, and producing
the candidate concepts list from the domain ontology to
annotate. (2) Sorting the candidate concepts by the priority
of (Exact, Plug-in, Subsume, Fail) according to the
algorithm introduced in the [7]. The sorted candidate
concepts list is shown in Figure 2. Then the publisher can
select the most satisfying concept from candidate concepts
to annotate the names used in Web Service.
Figure 1. Fragment of original and annotated WSDL files However, it is possible that the names of operations and
parameters may not exist in the domain ontology, or reflect
the naming convention usually adopted by programmers:
III. IMPROVED MODEL OF SEMANTIC ANNOTATION UserName or currencyExhange are more frequent than the
In this section, a semi-automatically mapping simple names directly included in the ontology. For this
algorithm is used in the process of semantic annotation reason, if the names are composite words, we will tokenize
which will improve the accuracy of semantic annotation. the word and return the similar word to replace it; if the
Figure 2 shows the improved annotation process of names are not directly included in the ontology, our
mapping WSDL concepts to the nodes in Domain during algorithm will make use of the general-purpose ontology
semantic annotation. Different from traditional model, we (i.e. WordNet) to build the synset list of the missing name.
have developed an algorithm to semi-automatically map And then each word in the synset will be made matching in
each individual concept in the WSDL description (i.e. the same way introduced above. The sorted candidate
Inputs, Outputs, and Operations) to an ontological concept. concepts list is also created for the provider to annotate.
Phone
Mobile
Phone
Mobile
Nokia Moto Battery
subClassOf partOf
algorithm in [7]. In fact, the concept Moto should also be concepts, attributes, properties, constraints and so on.
considered as the candidate concept because it has a Theoretically, the three basic rules can cover all the
siblingOf relationship with the concept Nokia. For another possible modifications on RO and LO because any
case, if one service has an output which uses Mobile battery complex evolution can be decomposed, viewed as a
as name, there is no matching concept using the algorithm sequence consisting of these three operations.
in [7]. Actually, the concept Mobile Phone, which has a
partOf relationship with Mobile battery, has a possible B. MFI4OR-based Model of Semantic Annotation
opportunity to be the candidate concept. Our algorithm has As mentioned in section 2 and section 3, the process of
considered these possible matching relationships between semantic annotation to web service is based on the single
the signatures of Web Service and the concept items in the domain ontology. However there also exist some cases that
domain ontology when annotating, which ensures to different ontologies in the same domain are used to
achieve the satisfying annotation at most. annotate web service, which will lead to the issue of the
Whatever traditional model or improved model interoperability of multi-ontology in the same Domain.
mentioned in section 2 and 3, they are both focus on the As we all know, Domain Ontology is always
single domain ontology to annotate. In section 4, the constructed and standardized by Domain Experts which use
method of multi-ontology based annotation to web service common approved concepts and relationship among them,
is proposed which will lead to the flexible annotation in but sometimes service providers maybe define their own
WSR. concepts to semantically annotate their web services for
meeting their requirements. We regard the domain ontology
IV. MFI4OR-BASED MODEL OF SEMANTIC ANNOTATION defined by experts as Domain Reference Ontology (DRO)
and the domain ontology defined by users as Domain
A. What is MFI4OR? Local Ontology (DLO). However the process of defining
the new concepts in DLO is not at will, but evolved from
Ontology provides a basic technology to support the
DRO according to three ontology evolution rules (SameAs
semantic interoperability of different systems over a
Rule, Enhancement Rule, and Contraction Rule) defined in
distributed and heterogeneous environment. If the MFI4OR. That is to say, DRO and DLO are not isolated but
ontologies have the semantic relation, they could interrelated to each other based on these evolution rules.
understand each other to some extent, which is the Figure 4 shows that DRO and DLO can both be adopted to
precondition of semantic interoperability. MFI4OR is add semantic information to Web Service. It can extend the
mainly used for providing a common framework to scope of semantic annotation in the Domain, which will
register ontology and its evolution information for enhance the flexibility during the process of service
semantic interoperation between ontologies or web-based publication and discovery.
applications based on them. It provides the solutions by
defining “Reference Ontology” (RO) and “Local
Ontology” (LO) which are both registered by a common
structure formed as Ontology-Ontology Component-
Ontology Atomic Construct [4]. The Ontology
Components defined in RO can be reused by the other
ontologies, while that defined by LO can only for its own
definition purpose. RO is always constructed by domain
experts, while LO is evolved from RO. The LOs and RO
can interoperate with each other, which provide a solid
foundation for interoperability between the different
ontologies. Besides RO and LO, the ontology evolution
rules between them are the core part in MFI4OR which
can help record the detailed changes when LOs evolve
from RO. These rules can ensure that the different
ontologies used to annotate web service will not hamper
the interoperability among them.
General speaking, there always exist three rules:
SameAs Rule, Enhancement Rule, and Contraction Rule [5, Figure.4. Multi-Ontology annotation Based on MFI4OR
6]. SameAs Rule can set up the equivalent mapping
between instances of Ontology, Ontology Component or
Ontology Atomic Construct. Enhancement Rule is special As mentioned above, DLO is evolved from DRO, and
for adding some new elements to the original ontology the evolution information between DRO and DLO should
when versioning and evolving. Contraction Rule is often be registered and recorded, which will improve the
adopted in ontology evolution management to delete interoperability between them in WSR. Although there
618 ZENG ET AL.
the future.
C. Realization of the MFI4OR-based Model Consumer Seller Buyer Seller
1.8
web services whose output is businessman or person
according to traditional semantic model. That is to say, this 1.6
model of semantic annotation will return the more web 1.4
services to satisfy the users. 1.2
TMSA
1
IMSA
Table 1. Evolution information of DLO 0.8
0.6 MMSA
Evolution information 0.4
attribute of LO Value 0.2
Name Buyer 0
Number of web service
URI URI_ buyer 100 200 300
modelType OWL Figure 7. The relationship of retrieving time and the number of web
beforeChange Consumer service in TMSA, IMSA, MMSA
afterChange Buyer
transformationRule SameAs Rule
consistentChange True
A MFI4OR-BASED APPROACH OF SEMANTIC ANNOTATION FOR WEB SERVICE 619
ACKNOWLEDGMENTS
0.1
0
Number of web service
100 200 300 This work is supported by National Basic Research
Figure 8. The relationship of precision and the number of web service in Program (973) of China under grant No. 2007CB310801,
TMSA, IMSA, MMSA National High Technology Research and Development
Program (863) of China under grant No. 2006AA04Z156,
National Natural Science Foundation of China under grant
V. RELATED WORKS No. 60873083,60803025, 60970017 and 60903034, Natural
Science Foundation of Hubei Province for Distinguished
Our work presents an approach for adding semantics to
Young Scholars under grant No. 2008CDB351, Natural
Web services based on MFI4OR. In fact, much work has
Science foundation of Hubei Province under grant No.
been done on semantic annotation to Web services. We
2008ABA379, Research Fund for the Doctoral Program of
discuss two representatives of them. A classification
Higher Education of China under grant No. 20070486065
method is described in [11]. It uses machine learning
and 20090141120022, China Postdoctoral Science
algorithms to help users simplify the search for relevant
Foundation under grant No. 20090460107, and the
ontology. Unfortunately, the method considers just names
Research Project funded by Wuhan University under grant
of WSDL concepts, ignoring the structure of them.
No. 6082005 and 6082007, Scientific and Technological
Moreover, it uses vocabularies, but does not use ontology
research projects of Hubei Province Education Department
which is more descriptive and capture domain more
under Grant No.D20091012.
accurately for classification. MWSAF [12] uses a media
structure called SchemaGraph to facilitate the matching
between XML schema and ontology to find the relevant REFERENCES
ontology. However, since the matching results are not good [1] R. Chinnici, J. Moreau, A. Ryman, and S. Weerawarana, Eds.,
if WSDL files do not have good structure or the ontology “Web Services Description Language (WSDL) Version 2.0 Part 1:
becomes much comprehensive, it is also not suitable for our Core Language”, W3C Candidate Recommendation 27 March
2006; http://www.w3.org/TR/2006/CRwsdl2020060327
context. Moreover, [13] proposes a template-based markup
[2] K. Verma and A Sheth, “Semantically Annotating a Web Service,”
tool, which can improve annotation accuracy significantly. IEEE Internet Computing 11 (2),March–April 2007, pp. 83–85
But as used for semantic web content annotation, the [3] Jian Wang, Keqing He, Wei Liu, Rong Peng, Bing Li, “A Service-
template is over complex, and has a poor reusability Oriented Domain Modeling Framework for Networked Software”,
considering the various web contents. Dynamics of Continuous Discrete and Impulsive Systems - Series
B, Special Issue on Software Engineering and Complex Networks,
The common factor in these approaches is only relating December 2007, pp. 246-252
signatures of Web service to SINGLE domain ontology, but [4] ISO/IEC 19763-3:2007. Keqing He.
our approach can use DIFFERENT ontologies in the same http://www.iso.org/iso/iso_catalogue/catalogue_tc/catalogue_detail.
domain, which are not isolated but interrelated to each other htm?csnumber=38637
based on ontology evolution rules defined in MFI4OR, to [5] Chong Wang, Keqing He, Yangfan He. MFI4Onto: Towards
annotate web service, which will extend the scope of Ontology Registration on the Semantic Web [A]. In Proceedings of
6th International Conference on Computer and Information
semantic annotation in the Domain and greatly enhance the Technology (CIT 2006) [C], Seoul, 2006, pp.40-46
flexibility during the process of service publication and [6] Cheng Zeng, Keqing He, Zhitao Tao. Towards Improving web
discovery. service Registry and Repository Model Through Ontology-Based
Semantic Interoperability, GCC2008, 2008, pp.747-752
VI. CONCLUSION AND FUTURE WORK [7] Paolucci, M, Kawamura, T, Payne, T. R, and Sycara,K, Semantic
Matching of Web Services Capabilities, In Proceedings of the 1st
At present, few academic models of semantic International Semantic Web Conference (ISWC2002).
annotation extend beyond simple subsumption relationship [8] D. Martin, M. Burstein, J. Hobbs,OWL-S: Semantic Markup for
based input and output matching in single domain ontology. Web Services”, W3C Member Submission, Nov.2004;
In this paper, in IMSA we have considered all possible http://www.w3.org/Submission/2004/SUBM-OWL-S-20041122/.
matching relationships between the signatures of Web [9] H. Lausen, A. Polleres, and D. Roman, “Web Service Modeling
Service and the concept items in the domain ontology when Ontology (WSMO)”, W3C Member Submission,
2005;http://www.w3.org/Submission/WSMO/.
annotating. More important, in MMSA we are not limited
[10] R. Akkiraju, J. Farrell, J. Miller, M. Nagarajan, M. T. Schmidt, A.
to single domain ontology to annotate. Different ontologies Sheth and K.Verma, “Web Service Semantics - WSDL-S”, W3C
in the same domain based on MFI4OR can be adopted
620 ZENG ET AL.
Member Submission, 7 November 2005; [12] Patil A, Oundhakar S, Sheth A, Verma K. METEOR-S Web service
http://www.w3.org/Submission/WSDL-S/. annotation framework. In Proceedings of the 13th Int'l World Wide
[11] A. Hess and N. Kushmerick. Automatically attaching semantic WebConference, New York, ACM Press, 2004, pp.553-562.
metadata to Web services. Proceedings of the 2nd International [13] Brian Kettler, James Starz. A Template-Based Markup Tool for
Semantic Web Conference (ISWC 2003). Semantic Web Content. Proceedings of ISWC2005, LNCS 3729,
2005.
Ethernet Node for Serial Communications
A.S. Popa1, M. Popa2, T. Slavici3, D. Darvasi4
1,2,3
“Politehnica” University from Timisoara, Romania
4
”Ioan Slavici” University from Timisoara, Romania
Abstract—Several protocols exist in serial communications, • MISO and MOSI: data lines; for a master device MISO is
such as SPI, USB, RS232, RS485, CAN etc. In order to obtain a data input and MOSI is a data output and for a slave device
flexible and fast communication, serial communication nodes the roles of the two lines are inversed; all the MOSI lines are
were developed. They are generally based on microcontrollers
and can be seen as protocol converters. Their capabilities were
connected together and all the MISO lines are connected
enriched by adding the remote command possibility, through an together; there is only one master which will command the
Ethernet interface. This paper presents a serial communication SCLK line and will generate data on its MOSI line to the
node based on the powerful 32 bit MPC5553 microcontroller. It MOSI inputs of the slaves; only the selected slave will
implements a bridge between two SPI, two RS485 and one generate data on its MISO output to the MISO input of the
Ethernet interface. It is used for collecting data from sensors and master;
sending them over the Internet to a testing station.
• /SS: the significance of this line differs; if the device is a
slave one, then /SS is an input for its validation and the master
I. INTRODUCTION must generate so many /SS lines how many slaves are
Several protocols exist in serial communications, such as therefore for a system with n slaves there are n + 3 lines; if the
I2C, SPI, USB, RS232, RS485, LIN, CAN etc. They are more device is a master one, /SS may be output or input, its
or less appropriate for certain applications. For example, the significance being software selected; if it is an output it will
CAN protocol is very used in automotive applications, due to validate a slave and if it is an input it can be used for error
its reliability, RS232 protocol is used in industrial and PC detection or to receive an external request from other device
based applications, due to its simplicity and low cost and so for the master position.
on. RS485 is a protocol for serial asynchronous communication
In order to obtain flexible and fast communication, serial between more devices directly connected to the same lines.
communication nodes were developed. They are generally The connection is single master many slaves type. The logical
based on microcontrollers and can be seen as protocol levels are represented by differential voltages which increase
converters. Their capabilities were enriched by adding the the number of lines but the protection to noises too leading to
remote command possibility, through an Ethernet interface. a maximum distance between the devices of 1200 m.
Such microcontroller based serial nodes exist, with
different performances. In reference [1] a node for wireless III. THE PROPOSED SOLUTION
sensor networks is presented. It is based on an Atmel ATmega
128L microcontroller with two UART, one SPI and one I2C A. The Hardware
interfaces and with input ports for sensors. Reference [2] Fig. 1. presents the bloc diagram of the proposed serial
describes a network node based on an Atmel ATmega 128L node. It is based on the powerful MPC5553 microcontroller,
microcontroller, which has an I2C, two UART and one [3], based on the PowerPC Book E architecture. It collects
Bluetooth interfaces. In reference [4] a powerful node is data through two internal SPI interfaces. Each such interface
presented. It is based on two microcontrollers, an 8 bit Atmel is connected to SPI lines and to RS485 lines. Both interfaces
ATmega 128L and a 16/32 bit Atmel AT91FR4081 ARM were connected in order to ensure compatibility with newer
THUMB and includes I2C, RS232, RS485, SPI and RF sensors which transfer data through the SPI bus, and with
interfaces. older sensors which transfer data through the RS485 bus. A
This paper presents a serial communication node based on differential driver was foreseen on the SPI lines because of
the powerful 32 bit MPC5553 microcontroller. It implements the high distance between the sensors and the microcontroller.
a bridge between two SPI, two RS485 and one Ethernet The RS485 lines are connected to the SPI lines of the
interface. It is used for collecting data from sensors and microcontroller through a RS485 – SPI adapter. The Ethernet
sending them over the Internet to a testing station. The next interface was implemented with a LAN8187i transceiver
section shortly describes the SPI and RS485 protocols, the connected to the internal MII interface. The MII interface
third section presents the new solution and the fourth section ensures the connection to the internal specific FEC (Fast
outlines the conclusions and future development directions. Ethernet Controller) peripheral. FEC corresponds to the MAC
layer and requires an external transceiver for being connected
II. THE SPI AND RS485 SERIAL PROTOCOLS to the physical interface. The CAN interface will be used to
connect the module to other blocks from the vehicle,
The SPI protocol describes the communication via a full
connected to the CAN bus and the SPI3 interface will be used
duplex, master – slave interface. The interface lines are:
for writing parameters into the EEPROM internal memory
• SCLK: common clock for all the devices: it is an output through an USB interface.
for a master device and an input for a slave device;
621
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_108, © Springer Science+Business Media B.V. 2010
622 POPA ET AL.
The MPC5500 microcontroller has the e200z6 core which • data reception: the bytes received from the SPI interface
implements a version of the PowerPC Book E architecture. are buffered until the reception of the whole multiframe and
This is appropriate for high performance embedded systems • data transmission: the data bytes are sent over the SPI
applications. The e200z6 core integrates an arithmetic interface.
execution unit, a branch control unit, units for fetching
instructions and for reading/ writing data and a multiport The Ethernet driver has the following roles:
register able to support up to three read operations or two • driver initialization: establishment of a connection to the
write operations in the same clock period. Most of the Ethernet and establishment of the MAC address,
instructions are executed in a single clock period. E200z6 is a • Ethernet frames reception: frames are buffered,
scalar core with 64 general purpose registers. Floating point • Ethernet frames transmission: an Ethernet frame is sent
instructions are not supported by the hardware but are over the MII interface to the external transceiver which will
emulated, through the traps. It exists also a signal processing send it on the Ethernet and the maximum length of an
extension for supporting floating point operations in real time. Ethernet frame is 1518 bytes.
Other features of the e200z6 core are: powerful interrupts/
exceptions handling unit, memory management unit with full The communication protocol between the sensors, the node
associative 32 inputs TLB, 2 – way set – associative 8 Kbytes and the Ethernet achieves the following operations:
cache, delay and watchdog functions and low consume • the frames received from the Ethernet are unpacked and
modes. the command contained in the data field is sent on the SPI
The PWM signals are obtained by program. Internal bus;
counters are used and their content (C) is continuously • the Ethernet frames with a destination address different
compared with programmed values (PW). If C < PW, 0 logic than the MAC address of the serial node are ignored;
will be sent on the PWM output and if C PW, 1 logic will be • the frame received from Ethernet is not verified; it is
sent on the PWM output. supposed as being correct;
B. The Software • the bytes collected from the sensors are packed in one or
several frames and are sent on the Ethernet;
The application was conceived in Visual C++ based on
MFC (Microsoft Foundation Class) library which permits easy • the destination address for the frames is broadcast type;
implementation of the Windows based programs. The • in case of errors at an Ethernet frame transmission, for
application has a graphical interface with dialog windows example because collisions are detected, a retransmission is
specific to the Microsoft Visual C++ environment. tried and after 10 such tries the transfer is aborted and error
The SPI driver has the following roles: situation is indicated through a led.
• initialization of the attributes of the transfer: speed, The software was conceived in two layers:
number of bits/frame, polarity and phase of the clock signal,
initial mode,
ETHERNET NODE FOR SERIAL COMMUNICATIONS 623
• the Hardware Abstraction Layer: it consists of the classes The SPI_Driver class uses the
corresponding to the microcontroller interfaces; it contains the SPI_Driver_Interface interface for implementing the
following classes: SPI_Driver, ETH_Driver, callback mechanism. There are two methods for enabling,
DMA_Driver and TimerDriver; (b_AttachCallbackHandler(SPI_Driver_Interf
• the Application Layer: contains the application, more ace*CallbackHandler)), and disabling,
exactly the protocol between the sensors, the serial node and (v_DetachCallbackHandler()), this mechanism. It
the Ethernet; it contains the SPI_EthHandler and informs the client that the transmission was successfully or
Handler_ETHDriver classes. that a new SPI frame was received and it was sent to the
This structure is implemented by a hierarchy of packets and client. Thus, the client will be informed automatically about
classes. Fig. 2. presents this hierarchy and fig. 3. presents the the occurrence of an event without the need of continuously
UML diagram of the software architecture. interrogating the class SPI_Driver about that event. The
SPI_Driver_Interface specifies the methods
Fig. 4. presents the UML diagram of the SPI_Driver
v_RxCompleted(T_UWORD received_data) and
class. The singleton design pattern was used for implementing
this class. The reason is that the microcontroller has three SPI v_TxCompleted() for the two events above mentioned.
interfaces, therefore the class must be instantiated three times. Fig. 5. presents the UML diagram of the ETH_Driver
The spi_GetInstance static method returns the instance class. It implements the singleton pattern and defines the
of the SPI channel specified by the e_spi_channel interface corresponding to the FEC peripheral. It contains
parameter. methods for initializing the FEC peripheral, the LAN8187i
transceiver, for transmitting and receiving Ethernet frames
The SPI_Driver class contains methods for initializing and for other configurations. The return of an instance of the
the internal DSPI peripheral, for transmitting and receiving class is done through the eth_GetInstance()method.
SPI frames and for configuration.
The method ub_Init() configures the pins
The method v_Init() initializes the SPI internal
corresponding to the MII interface and initializes the internal
peripheral. The terminals and the attributes of the transfers are
registers through which the origins of the memory zones
configured. The method T_VOID v_Send(T_UWORD allotted to the circular lists of the transmission and reception
transmitData, T_BOOL b_EOQFlag) has the role of descriptors and the dimension of the reception buffers are
sending commands and data on the SPI interface. specified.
624 POPA ET AL.
The method ub_Open(T_UWORD uwMode, const The parameters of the connection can be by established
T_UBYTE* pubMacAdress, T_UWORD uwPhyType) by configuration or through the auto – negotiation process.
achieves the operations necessary to the FEC peripheral to be This is a mechanism for exchanging information between two
ready for transmitting and receiving. These operations are: interconnected nodes and in which the operation mode with
initialization of the descriptor lists, specification of the MAC higher performance is selected among the operation modes
address, configuration of the control registers for transmitting supported by the two nodes. The auto – negotiation process
and receiving and initialization of the PHY external can be started automatically when the connection beween the
transceiver. For initializing the descriptor lists, two private two nodes is lost, by setting a specific bit or can be started by
methods were foreseen, v_RxBDInit() and software. The configuration information sent by the PHY
v_TxBDInit(), called from the ub_Open() method. transceiver to the partner during the auto – negotation process,
are saved in specific registers. At the end, one register
contains the partner’s parameters and an other one contains
the connection’s parameters. The results of the auto –
negotiation operation can be used by the FEC peripheral too,
for example it can set the communication type as being half –
duplex or full – duplex.
Abstract. In this paper we present one tool named accessing during the transaction process. If the
EvalTool for evaluate the partitioning scheme of fragments are closer as possible to the needs of the
distributed database. This tool could be use for transactions set, then the transactions processing cost
calculate the best partition scheme both for horizontal can be reduced.
and vertical partitioning, and then alocate the
fragments in the right places over the network. We use
For the distributed databases design, the
an algorithm studied before, which could calculate transaction cost is reduced by increasing the local
some cost for access data to the local site and the other transactions (on a site) and at the same time by
cost for access data from the remote site. The minimum reducing the amount of data views which are not
value of this costs give the best fragmentation scheme local. The aim of vertical partitioning technique (and
for each relation. in general, partitioning techniques) is to find a
1. ABOUT FRAGMENTATION OF DISTRIBUTED
partitioning scheme to meet the objective outlined
DATABASES above.
Note that the partitioning problem can be
The distributed databases design is an tackled on different levels of detail considering some
optimization process that requires to obtain solutions additional information.
to several interpenetrating problems, namely data In paper [1], we have to consider only the
fragmentation, data allocation and local optimization. information about transactions as input data to
Each problem can be solved through different manage the partitioning problem effectively. In fact,
approaches, and therefore the design of distributed the global optimization problem (which includes a
databases becomes a very difficult task. Usually the large number of parameters and a metric complex) is
design process is heuristical defined. divided into several smaller optimization problems in
The distributed databases design derived order to reduce the search space and to reduce the
from the non-distributed databases design only in complexity of each problem separately.
terms of distribution. The design involves data In speciality literature are proposed several
acquisition, database partitioning, the allocation and vertical partitioning algorithms, so it can be
replication of partitions and local optimization. The measured the affinity between pairs of attributes and
database partitioning can be performed in several try to group attributes together under the affinity
different ways: vertical, horizontal and hybrid (also between, using the algorithm BEA (bond energy
called mixed) partitioning. This aspect of a database algorithm). In an article we use an heuristical cost
design will be highlighted in this thesis and namely estimator for design a file to obtain a partitioning
the developing of an algorithm for checking some scheme “bottom-to-top”. In another article extends
various algorithms proposed in literature for the approach proposed by BEA and algorithm has
distributed databases partitioning (fragmentation). two phases for vertical partitioning. The partitioning
Basically, it will be considered the problem algorithms mentioned above use several heuristical
of vertical data partitioning (fragmentation), also methods for a relationship fragmentation.
known as the attributes partitioning. This technique is As input date for these algorithms it is used
used for the databases design to improve the the attributes matrix (AUM). This matrix has the
performance of transactions. In vertical partitioning, attributes as columns, the transactions as lines and
the attributes of a relation R are together in groups the transactions frequency access are values in the
that do not overlap and the relationship R is designed matrix. The most previous algorithms used for the
on relations fragments in accordance with these data fragmentation use of affinity matrix of attributes
attributes groups. (AAM) which is derived from AUM. AAM is a
In the distributed databases, these fragments matrix in which for each pair of attributes it is stored
are allocated on different sites. Here comes the aim the total frequency access of the transactions that
of vertical partitioning to create vertical fragments of access the pair of attributes. The results of different
a relationship to minimize the cost of the data algorithms are often different, even if it is the same
627
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_109, © Springer Science+Business Media B.V. 2010
628 RUNCEANU AND POPESCU
affinity matrix of the attributes, so it indicates that allocate the calculated fragments. It is also possible
the objective functions used are different. Most to load an external fragmentation and use it for the
vertical partitioning algorithms do not have an allocation, this might be useful as horizontal
objective basis for assessing the correctness of fragmentation.
partitions that are obtained through the application of For both the fragmentation and allocation,
those algorithms. Also, there is a common criterion using the implemented algorithms, the tool reports all
or objective function to evaluate the results of these the possible solutions and the one considered the
algorithms for vertical partitioning. best.
The fragmentation and allocation algorithms
Bellatreche et al. [3] presents an horizontal are mainly based on the book [10]. We modified the
fragmentation method and an analytical cost model to cost analysis for the allocation, the model used is
evaluate the query execution time in horizontally described in the next section.
fragmented databases. The fragmentation schema The algorithms implemented are:
with the best performance according to the cost - Bound Energy Algorithm
model is achieved through a hill-climbing algorithm, - VF algorithm
by selecting a subset of classes for primary horizontal - Best-Fit heuristic
fragmentation. - Cost Analysis (based on the model
The work by Ezeife and Barker [6] presents described in the next section)
a set of algorithms for the horizontal fragmentation The tool works with text files as input, at
of all classes in a database schema. It takes least it needs:
relationships between classes into account to propose - one file for general information (sites names,
primary and derived horizontal fragmentation. applications frequencies, network values, and other
However, this approach works at the instance level, information)
where the class instances already exist in the - one file for the information about a relationship
database to proceed with the fragmentation process. (attributes names, useMatrix, and other information)
It also assumes a storage structure for the objects Once the 2 text files are loaded it is possible
hierarchy in the database class in which an instance to vertical fragment the relationship. The tool works
of a subclass physically contains a pointer to the with the following steps:
instance of its super-class that is logically part of it. - It is calculated the affinity matrix
This assumption leads to considering inheritance - Clustering affinity matrix is calculated by the Bond
links in the horizontal fragmentation process. Energy Algorithm (in the relationship text file is also
Ceri et al. [4] proposes a methodology for possible to specify the column insertion order)
the horizontal fragmentation of all classes in a - all the solutions for splitting CA matrix into two
database schema. The choice between primary and fragments are evaluated with the VF algorithm
derived horizontal fragmentation on each class - if the best solution is to fragment the relationship,
considers its relationships, which are defined by VF algorithm is iterated for each of the fragments
analyzing only the method calls between classes in calculated
the schema. The work does not present an algorithm The tool reports the CA matrix and all the
to support the methodology. steps of both the Bound Energy Algorithm and VF
Navathe et al. [8, 9] proposes a complete algorithm. Is also possible to store this information in
fragmentation methodology for OODBMS, which is a text file.
divided in three phases. First, there is an Analysis The allocation uses:
Phase to assist distribution designers in defining the - the best fit heuristic
most adequate fragmentation algorithm (horizontal, - a cost analysis (the model used is described in the
vertical, or both) to be applied in each class of the next section)
database schema. The Analysis Phase also considers EvalTool uses 2 policies of allocation:
the case in which no fragmentation of a class is the - “one for each site” (the default one)
best alternative. Second, they present an algorithm to - “one for all sites” (can be activated selecting the
perform Vertical Fragmentation in a class. Finally, relative option button)
the authors present an algorithm to perform An example will help to understand:
Horizontal Fragmentation on the whole class or on a consider that you are designing a distributed database
vertical fragment of a class, which may result in for your university.
mixed fragmentation. The university has one central site and five
peripheral sites managing all the didactics.
2. MAIN CARACTERISTICS Each peripheral site is independent from the
others: just use the information in the database about
In few words EvalTool considers one its students, its exams, its teachers, and so one, it is
relationship for time, once loaded is possible to easy that a few of the relationships of the database
fragment it (in the first time only vertically) and
EVALTOOL FOR EVALUATE THE PARTITIONING SCHEME 629
will be fragmented in one fragment for each QPCi = PCi + TCi (4)
peripheral site (e.g. the student relationship) and, Processing component consists of three cost
because of the “independence” of the sites, the only factors: access cost (CA), the integrity constraint (IC)
solutions for allocation with sense will be to allocate and control competitor (CC):
the fragment on its site or on the central one: none of PCi = ACi + ICi + CCi. (5)
the other sites will use it; is easy that a lot of Detailed specification of these three cost
relationship will be fragmented in this way, think just factors depend on the algorithms used for these tasks.
to the students, the exams and the teachers For example, AC has the form:
relationships, each one fragmented in 5 fragments:
not considering others fragmentation 15 fragments ¦STC ≤storagecapacityat
∀Fj ∈F
jk
site Sk, ∀S∈S. (6)
are created.
Using the “one for each site” (a more correct The first two terms from the previous
name would be “one fragment for each peripheral formula calculates the number of accesses to the
site”) is possible to reduce the number of the query qi the fragment Fj and UR + RR value
fragments that the user has to manage: the input for provides the total number of accesses that executes
the allocation won't be 3 text files with 5 fragments work | retrieve and update them. Local cost of
each one, but 3 text files with 1 fragment each one. processing them is assumed to be identical, given
But there's also the possibility that a LPCk multiplying the total cost of access to station
relationship is not fragmented in one fragment for Sk and multiplying xjk show that will select only
each peripheral site (because is not possible to relate those values of the cost for stations where parts are
a tuple to a single peripheral site or simply because stored.
there's no need). In this case there aren't no-sense Access cost function requires that each
solutions for the allocation of the relationship (or the application can be divided into subqueries that run on
fragments obtained from other fragmentations), it can each fragment stored on a station, the results are
be allocating in all the sites central or peripherals; in returned to the station from which the request was
this case must be used the “one for all sites” policy sent. This is a simplistic view that ignores the
(one fragment for all sites). This policy can also be complexity of database processing. For example, it
used if the user in the example before wants to create was revealed the cost of execution join unlor or cost
3 text file with 5 fragments each one. competitiveness and integrity.
EvalTool was developed as part of a study The cost of transmission can be made
for a grant research. Maybe the tool “fits too much” analogous to the cost of access. The queries that are
our problem so that there could be problems for using updating is required to inform all stations where the
it in a more general one, but we considered useful to replicas, while the queries that are restored, it is
offer our solution to the community. sufficient to access only one of the descendants, in
addition, after the application update, no data
3. COST MODEL FOR ALLOCATION transmission back to state of origin (except a
confirmation message), while the queries that are
We present an allocation model that restored may result in transmission of data.
minimizes the total cost of storage and processing, Component update transmission function has the
trying to comply with certain restrictions imposed on form:
response time.
We consider a decision xkj variable, defined as: TCU
i
= ¦ ¦ uij *xjk *go(i), k + ¦ ¦ uij *xjk *gk, o(i),
∀Sk∈S ∀Fj ∈F ∀Sk ∈S ∀Fj ∈F
(7)
1, if fragmentFk is placed on site Sj
xkj = ® (1) where the first term is to send a message to update
¯ 0, in the other cases
the initial station 0 (0 demand all copies of fragments
Total cost function has two components: that must be updated and the second term is to
processing applications and storage. This can be confirm transmission). Component retrieve can be
expressed as: specified as follows:
The cost of transmission for qt query can be matrix. In [5], Chakaravarthy et al. computes the
described as: processing costs of distributed fragments using an
TCi = TCUi + TCRi (9) application/attribute matrix that counts the number of
Constraint functions can be analyzed times applications access attributes at distributed
similarly. For example, response time constraint can sites.
be described thus: execution time of the call qi ≤ This paper computes the partition evaluator
maximum response time of applications qi, ∀qi ∈ Q , value (PE) with this formula:
and storage constraints may be described by the PE = EL2 + ER2 (11)
relation: where
M T ª § S ·º
ACi = ¦ ¦(u * UR +r * RR )* x * LPC , (10)
∀Sk∈S ∀Fj∈F
ij ij ij ij jk k E L2 = ¦¦ «qt2 ∗ S it ¨¨1 − it ¸¸» (12)
«
i =1 t =1 ¬ © ni ¹»¼
The heuristic methods must be sought that n - Total number of attributes in a relation that is
produce sub-optimal solutions and must test the being partitioned.
closeness of the results of heuristic algorithm for T - Total number of transactions that are under
optimal allocation. Unfortunately there is insufficient consideration.
qt - frequency of transaction t for t = 1,2, …, T.
information to determine the “near” optimal
solution. M - Total number of fragments of a partition.
ni - number of attributes in fragment i.
Sit - set of relevant attributes in fragment i that the
4. PARTITION EVALUATION FUNCTION transaction t
accesses; it is empty if t does not need fragment i.
The partition evaluation function is an |Sit| - number of relevant attributes in fragment i that
objective function used to evaluate the "goodness" of the transaction t accesses.
a particular partitioning scheme. It can be used to
measure the performance of a fragmentation system ª º
( ) ∗ nR
T
E R2 = ¦ min iM=1 ¦ «¦ f tpkk
2 itk
whenever a major change in access pattern
remote » (13)
information or global conceptual schema occurs. t =1 k ≠i ¬ Pk itk ¼
However, the defining of the suitable function for the
vertical fragments and the building of EvalTool for
ni – number of attributes in fragment i.
vertical fragmentation are the main contributions of remote
this work. [1] nitk - Total number of attributes that are in
fragment k accessed remotely with respect to
5. THE PARTITIONING EVALUATION FUNCTION fragment i by transaction t.
Attributes of relations are the members of f tji - Frequency of transaction t accessing attribute j
relational vertical fragments. On the other hand, the in fragment i
methods and attributes of object classes are the Ritk - Set of relevant attributes in fragment k
member of object vertical fragments. The instances
are members of the horizontal fragments object. The accessed remotely with respect to fragment i by
set of instance objects in fragments created by transaction t; these are attributes not in fragment i but
different fragmentation schemes may be different. It needed by t
is necessary to develop a modified partition evaluator Ritk - Number of relevant attributes in fragment k
to deal with object horizontal fragmentation schemes. accessed remotely with respect to fragment i by
This partition evaluator counts the number of local transaction t
irrelevant accesses made to fragments by queries as The algorithm implements the objective
well as the numbers of remote accesses made to function defined above. For measuring the
fragments for each set of horizontal fragments performance of a partition scheme, algorithm needs
created with a scheme. two inputs. One is every horizontal fragment from
Therefore, in order to measure the system the partition scheme; the other is the application
performance each fragment is assumed placed at one object instance sets. Then produces the PE value of
of the distributed sites, while the PE function is used this partition scheme. Each PE value corresponds to
to measure the partition evaluation value (the PE the total penalty cost incurred by each scheme
value) for the system. The higher the PE value of the through both local irrelevant access costs and remote
system, the worse the system performance. relevant access costs. A lower PE value means less
Computing the PE value of horizontal fragments performance penalty and thus, a better performance.
requires an input matrix that shows for each The algorithm is below. This algorithm accepts the
application the number of times an instance object is application object instance sets of al1 applications,
accessed. This is called the application instance the set of horizontal fragments allocated to
EVALTOOL FOR EVALUATE THE PARTITIONING SCHEME 631
distributed sites and application access frequencies at fragment and that fragment haven’t one single
distributed sites. It then computes the local irrelevant attribute accessed by that transaction, then
access cost of each class as the sum of local transaction not be run on that fragment. We consider
irrelevant access cost of al1 applications for each 3 cases of matrix of attributes.
fragment of the class. Similarly, it computes the For the first test we used a matrix of
remote relevant access cost as the sum of remote attributes use with ten attributes accesed by eight
access cost made by al1 applications to fragments of transactions.
this class at remote sites. The sum of these two costs Tranzactions \ Attributes 1 2 3 4 5 6 7 8 9 10
makes up the PE value of the class. T1 25 0 0 0 25 0 25 0 0 0
We implemented the algorithm based on the T2 0 50 50 0 0 0 0 50 50 0
formula from equation (11) calculates the value of T3 0 0 0 25 0 25 0 0 0 25
PE, for a given fragmentation scheme so we used an T 4 0 35 0 0 0 0 35 35 0 0
T5 25 25 25 0 25 0 25 25 25 0
entrance date: the matrix used for attributes - A; the
T6 25 0 0 0 25 0 0 0 0 0
lots of fragments on which it calculated the value of
T7 0 0 25 0 0 0 0 0 25 0
PE, the relation -R.
T8 0 0 15 15 0 15 0 0 15 15
Algorithm
Input: A = attribute usage matrix;
R = Relation ; F = fragments set We present in Figure 1 values for each
Output: E2L: irrelevant local attribute cost; 2 2
number of fragments and values for EM , E R and
E2R: relevant remote attribute cost;
EP : partition evaluator value EP .
Begin Total number of fragments evaluates was
2 115975. Optimal value (minimum) is obtained for 3
E L=0
for i from 1 to M do fragments – fragment I (1,5,7), fragment II (2,3,8,9)
ei=0 and fragment III (4,6,10).
begin The program to generate all the
for j from 1 to T do combinations of ten attributes accessed by eight
ei=ei+qt2*|Sit| * (1 - |Sit| / ni ) transactions offers three solutions (for five
end_for fragments) and two solutions (for eight fragments),
E2M = E2M + ei having the same value for EP. However, the project
end_for of distributed database can choose which scheme of
E2R=0 partition wishes to use it.
for t from 1 to t do
begin
Number
of Partition scheme EM2 E R2 EP
fragments values values values
minim =maxint
for i from 1 to M do 1 (1,2,3,4,5,6) 24895 0 24895
for k from 1 to M do 2 (1,4) (2,3,5,6) 7565 55 7620
3 (1) (2,3,5,6) (4) 7565 276 7841
begin 4 (1) (2) (3,5,6) (4) 5063 11336 16399
if k ≠ i then 5 (1) (2) (3,6) (5) (4) 0 22492 22492
begin 6 (1) (2) (3) (4) (5) (6) 0 40913 40913
if exist attribute in matrix A who is from k Fig. 1 Results for test
fragment then
E2R= E2R+ (f tk)2 * | Rit | * | Rit | / 7. CONCLUSIONS
remote
n itk)
end_if In this paper is presented a general approach
2
if E R min < minim then of the fragmentation issue of the dates from a
minim = E2R min distributed database. Using this tool for obtained the
end_if best partitioning scheme with implementation of
end_for several clasic algorithms is one solution in designing
E2R = E2R + E2R min phase of a distributed database.
end_for
REFERENCES
end_for
2 2
PE = E L + E R
End. [1] Runceanu A. Fragmentation in distributed
databases, (2008), International Joint
6. EXPERIMENTAL RESULTS Conferences on Computer, Information, and
Systems Sciences, and Engineering (CISSE
For the execution of one transaction, we 2007) Conference, Conference Proceedings
know that if a transaction could be run at one book, December 3-12, 2007, University of
632 RUNCEANU AND POPESCU
Abstract—With the advent of 24-hour news channels and Internet However, there seems to be little research done on
news sites, the ease of finding news sources has become easier automatically creating digests.
than ever. However, with the increased number of sources
covering a greater variety of topics it is becoming increasingly This paper presents techniques for gathering and classifying
difficult for users to weed through the important and not so news articles and for creating simple digests. The system is
important news of the day. This paper presents a system that based on a keyword extraction algorithm that is applicable to
automatically creates a daily digest of the important news topics multiple languages. Currently, we are focused on creating an
of the day. News articles are mined from Internet news sites and English and Japanese version of the system.
the articles are stored for later analysis. At the end of the day, or
specified period, the articles are analyzed to determine the This paper will continue as follows. In section 2 a brief
important news topics and to create a listing of the top stories of overview of the proposed system will be given. Then, in
the day. The digest shows the top stories with their title, section 3 the news gathering and classification module will be
keywords, link and summary information. Through human discussed. In section 4 an overview of creating daily digests
judged evaluation it is shown that the created digests satisfy from the classified news articles is given. Then, in section 5 a
users. look at the current implementation details will be given.
Section 6 shows evaluation results. Finally, in section 7
Keywords-component; Digest Creation, Online News, concluding remarks will be made and future work discussed.
Personalization, Synopsis Creation
II. SYSTEM OVERVIEW
I. INTRODUCTION The system is made up of two modules: the news
In the past 10 years the number of online news sources has monitoring and article processing module and the digest
grown tremendously. With the increase in sources have also creation module. The news monitoring and article processing
come increases in the number of news articles, topics covered, module monitors a list of user defined RSS feeds and news site
etc. that are available. Keeping adequately informed, in an era pages for new news and the downloads, extracts and processes
of information overload, requires constant attention to Internet the articles. Processing the articles includes keyword
news sites, the sites' RSS feeds or backtracking at the end of extraction, category classification and summary generation.
the day to determine the important events. The digest creation module takes the downloaded and
processed articles at the end of the day and creates a digest of
Online news has become a hot topic in the information
the news by determining which topics and articles were the
retrieval world. From workshops on topic detection and
most important for the day. An overview can be seen in
tracking to various research projects, the information retrieval
of online news is of growing importance. There has been work figure 1.
done to automatically create RSS feeds of news archives [4] to
make past news more easily available. There have also been
efforts to combine news, such as the Pan-European news
distribution network [8]. However, keeping track of important
and popular stories and topics is still a problem.
In order to help the user find popular stories that may be of
interest or may be important, some online news sites offer most
recommended and most viewed sections. This allows for users
to quickly scan articles that others have found interesting. In
addition, news sites are making good use of RSS feeds that
allow users to view only specific content. Another way of Figure 1. System Overview
helping users is to offer daily digests. A digest summarizes the
most important topics and stories of the day. Some sites are The system is designed to be automatic and does not
offering daily digests, which are typically manually created. require any intervention by end users. The only requirement
from the end user is for them to set the time at which the news
monitoring and article processing module runs. For example,
This research was originally done while the author was at The University
of Tokushima.
633
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2_110, © Springer Science+Business Media B.V. 2010
634 BRACEWELL
the user could set the module to run from 6am to 5pm not been processed then the link is used to download the page
everyday. At completion, the digest creation module would containing the article. The article is then extracted from the
take over and create a digest for the time span given by the HTML web page using a rule-based extraction process.
user.
The digest creation module outputs in HTML, which could B. Article Extraction
be easily used as a digest, web page served on the Internet. The After examining many news sites we found a certain pattern
main output of the system is the digest, which is made up of the that described the prototypical article. The pattern is made up
top articles of the day with their category or genre (sports, of a title, then the article body and the footer. Certain sites,
business, politics, etc.), top keywords and a summary. however, utilize handcrafted rules in the form of regular
Additional output includes ranked lists of keywords and expressions that allow for very precise extraction. However,
categories for the day. this poses a possible bottleneck for the addition of new news
sites.
III. NEWS MONITORING AND ARTICLE PROCESSING The first step in article extraction is to find the title in the
News monitoring and article processing handles body of the HTML. This is done using the title extracted from
monitoring, downloading and processing news articles during the RSS feed or web page. Since these titles sometimes have
an assigned time span. An overview of the module can be seen extraneous information, such as web site name or provider,
in figure 2. The module monitors a list of RSS feeds and news they cannot be fully matched to a line in the HTML. Instead,
web sites for new news. After articles are downloaded and each line is examined to see if it is a substring of the title. If it
extracted they are processed by extracting keywords, assigning is, then that line is marked as the article's title.
categories and generating summaries. The next step is to find the article footer. This is done by
examining each line that comes after the article's title to see if it
contains one of the defined footer elements. Some of the
currently used footer patterns are Copyright, ©, (C), Email this,
Related Stories, etc. If no footer pattern is found then the
bottom of the HTML body is used.
After the footer is found the article can be extracted. The
article body is the text that falls between the title and the footer.
The article is then cleansed. The cleansing process removes
multiple white space and unimportant multiple punctuation
marks, such as --, **, ..., etc.
C. Article Processing
Article processing consists of keyword extraction, category
classification and summarization. Keyword extraction extracts
the most important words and phrases for describing the article
and its content. Category classification assigns to each article
one or more predefined news categories. Summarization gives
the important points of the article in a few sentences.
1) Keyword Extraction
For keyword extraction, the method proposed by Bracewell
et al. [1] was used. It is capable of extracting keywords from a
single document without the need for training data or corpus
statistics. A brief description will be given in the next couple of
Figure 2. Gathering and Classification Overview paragraphs.
The algorithm uses linguistic information in the form of
A. News Monitoring noun phrases to extract keywords. Phrases can make better
The news monitoring part of the module watches a set of keywords than single words as they can keep concepts
RSS feeds and news site web pages for new news articles. The portrayed in compound words together. For example ``New
sites are processed one at a time in sequential order. Depending York'' would be a better keyword than having ``New'' and
on the site either an RSS feed or HTML page is parsed and the ``York'' separately.
links and article titles are extracted. The algorithm is made up of three modules: morphological
After the articles are collected they are processed one at a analysis, NP extraction and scoring, and NP clustering and
time. First, the article's link and title are checked for in the scoring. Morphological analysis takes care of word
article database, which stores the article and its processing segmentation, part-of-speech tagging, and word stemming. For
results. If the link or the title is in the database then the article English, Porter's stemmer [7] and Brill's tagger [3] were used.
is treated as already processed and is ignored. If the article has For Japanese, Chasen [6] was used.
AUTOMATIC CREATION OF DAILY NEWS DIGESTS FROM INTERNET NEWS SITES 635
NP extraction and scoring extracts simple noun phrases and IV. DIGEST CREATION
gives them a score based on their frequency in the article and After the current day's news has been downloaded and
the frequency of the words that make it up in the article, see [1] processed, a digest can be created that will highlight the
for details on the score. NP clustering and scoring brings important news of the day. The digest creation module is made
together noun phrases that share some commonality and is an up of two parts: article set determination and synopsis creation.
attempt to prevent redundancy in the final set of keywords. For Article set determination uses clustering in an attempt to
example, “bank bailout” and “bailout” would be clustered determine the broad topics covered in the day's articles.
together. The clusters are scored based on the scores of the Synopsis creation takes the top topics and determines the
noun phrases in them. From each of the clusters a representative article from them.
representative is chosen and becomes a keyword for the article.
2) Category Classification A. Article Set Determination
For category classification, we used the method proposed in The first step in creating the digest is to determine which
[2], which is based on the previously mentioned keyword set of articles were the most important of the day. This is done
extraction algorithm. It is a keyword based classification by clustering the articles and ranking the resulting clusters. The
method that only requires positive examples. It has the added overview of this can be seen in figure 3.
ability that it is capable of updating the training information in
an online manner as well as easily adding new categories
without a complete retraining. It also had an F-Measure of 93%
on the test set in that paper. A brief description will be given in
the next couple of paragraphs.
Figure 3. Article Set Determination Overview
The algorithm is made up of N independent classifiers,
where N is the number of categories. Training involves First, a feature vector is created for each article. The feature
building a model from a set of articles that are in the desired vector is simply a vector space model representation of the
category. A category model is made up of a category name, article using the keywords that were extracted in the new
total number of documents counter and a keyword list. The monitoring and article processing module. The next step
keywords list contains the keywords and their raw document clusters the articles in an attempt to group articles of similar
frequency for the category. topic together.
The classifiers output the likelihood that an article is in the K-Means clustering [5] is used and the distance between
category. The likelihoods from all the categories are used to two articles is calculated as one minus the cosine similarity,
determine which of the categories should be assigned to the shown in equation 2. The value of K is automatically estimated
article. The categories that have a likelihood greater than the in the following manner. First, the range of possible K is
mean plus one standard deviation are assigned to the article. determined by using the lower bound of N/2 where N is the
The assumption is that these categories stand out among the number of articles downloaded and an upper bound N.
group and are the best choices for the article.
3) Summarization
Ai • A j
For summarizing the article, the extracted keywords are Distance( Ai , A j ) = 1 − (2)
used to assign a score to each of the sentences. The top N Ai × A j
sentences are then extracted and become the article's summary.
In this research N = 4. The idea behind the summarization
Five random, but unique, values are chosen from the range.
technique is that the sentences with a high number of keywords
Each of these values are chosen to represent K and the K-
in them are more important than sentences that have a lower
Means algorithm is allowed to run for a few steps. After
number or no keywords.
completion, the average intra-cluster and extra-cluster distance
Summarization is done in three steps. The first step is to are calculated. The value with the best combination of low
break the article up into individual sentences. This is done intra-cluster and high extra-cluster distance is chosen and the
using a rule-based approach. The second step is to score each K-Means algorithm is run until completion or until 100 steps
of the sentences using equation 1, where k1, ... ,kn are the have been completed.
keywords in sentence S and KScore(ki) is the keyword score of
ki. The final step is to create the summary. For doing this, the
sentences are first sorted by score and the top four extracted. CScore(Cluster ) = P(Category ) × Cluster (3)
Then the four sentences are sorted by position in the article and
used as the summary. After the articles are clustered and have a category
N
assigned, they are ranked. The category chosen for the cluster
is the most abundant one among the articles in the cluster. For
Score(k1 k n ∈ S ) = ¦ log(1 +KScore(k ))
i =1
i (1)
ranking each cluster is first scored using equation 3. The
probability of a category, P(Category), is estimated using the
frequencies of the classified categories for the day’s articles
and as such, will vary day to day.
636 BRACEWELL
B. Synopsis Creation
After the clusters are ranked by cluster score, the top
clusters can be chosen as the most important ones. Currently,
the top 10 clusters are used, but the user could easily change
this. For each of these clusters an article is chosen to represent
the cluster. To choose the representative, article the importance
of each article is calculated using equation 4, where A is the set
of keywords that define the article, DF(ki) is the number of
articles the keyword appeared in, and KScore(ki) is the
keyword score of keyword ki.
N
Importance( A) = ¦ DF (k ) × KScore(k )
i =1
i i (4)
Evaluation was done for English and Japanese looking at In the future, we hope to configure the system to work on
the agreement for the top 10 stories of the day. Each language the web and add more news sources. We would also like to add
had two judges. news tracking, which would track stories over time. It would
also be interesting to add named entity recognition to find the
English results can be seen in Table 1. Judge 1 and Judge 2 top people and places in the news for the day. We would like to
had very different agreements. Judge 2 leaned more in favor of also create a Chinese version of the system. Finally, we want to
political articles, while judge 1 more world events. The use of explore ways to improve the system. Personalizing the digests
personalization techniques could be used to improve the per user based on user defined topics of interest and feedback
individual results. could improve the agreement. We hope to examine a large
number of human created digests to mine features of digest-
TABLE I. ENGLISH EVALUATION RESULTS worthy articles.
Judge Agreement
1 26%
REFERENCES
2 78%
[1] D. B. Bracewell, F. Ren, and S. Kuroiwa. Multilingual single document
1∩2 83% keyword extraction for information retrieval. In Proceedings of the 2005
IEEE International Conference on Natural Language Processing and
Knowledge Engineering, Wuhan, China, November 2005.
Japanese results can be seen in Table 2. Judge 1 and Judge [2] D. B. Bracewell, F. Ren, and S. Kuroiwa. Category classification and
2 had a greater similarity in agreement than judges for English. topic discovery of news articles. In Proceedings of Information-MFCSIT
Both judges leaned towards political and business articles. 2006, Cork, Ireland, August 2006.
[3] E. Brill. A simple rule-based part-of-speech tagger. In Proceedings of
3rd Applied Natural Language Processing, pages 152–155, 1992.
TABLE II. JAPANESE EVALUATION RESULTS [4] S. Debnath. Creating rss for news articles and beyond. In Proceedings of
the 19th FLAIRS Conference, pages 275–280, 2006.
Judge Agreement
[5] J. B. MacQueen. Some methods for classification and analysis of
1 43% multivariate observations. In Proceedings of the 5th Berkeley Symposium
2 41% on Mathematical Statistics and Probability, pages 281–297, 1967.
[6] Y. Matsumoto, A. Kitauchi, T. Yamashita, Y. Hirano, H. Matsuda, K.
1∩2 49%
Takaoka, and M. Asahara. Morphological analysis system chasen
version 2.2.9 manual. Technical report, Nara Institute of Science and
Technology, 2002.
[7] M. Porter. An algorithm for suffix stripping. Program, 14:130–137,
VII. CONCLUSION 1980.
In this paper a system that creates daily digests of news [8] M. Schranz, S. Dustdar, and C. Platzer. Building an integrated pan-
articles was presented. The system monitors and gathers news european news distribution network. In Proceedings of the 6th IFIP
to download and classify. Classification is made up of keyword Working Conference on Virtual Enterprises, PRO-VE 2005, pages
26–28, 2005.
extraction, category classification and summarization. Using all
of this information a digest can be created that highlights the
important news of the day as well as the top categories and
keywords. Currently, the system works on English and
Japanese news.
Index
A Cloud model, 2
Cluster
Academic advising, 25–29 ensemble, 305–308, 311–312
Academic record, 151–153 head, 539–542, 549–552
Academic social network, 151–156 Clustering algorithm, 306, 312, 542, 549–553
Acquisition system, 243–245, 253, 396 CMOS, 339, 405, 416
Adaptive assessments, 19–23 Coarse-grained PEDT/PSS, 263, 267
Adjacency relationship, 519–522 Coaxial outputs, 357
Agile software, 603–606 Collaborative learning, 83–87, 122, 134, 171
All Share Price Index (ASPI), 269–273 Collinear and triangle equation, 351–354
Analogical-digital converter, 373 Colombo Stock Exchange, 269
Analytical Hierarchy Process (AHP), 77, 82 Command circuit, 373
Architectural education, 163–166 Common-wavelength, 499, 501
Artificial intelligence, 44, 46, 73, 75, 147, 227, Communication, 45, 55, 113, 145, 149, 243,
361, 421–422, 424 461–465, 570
Artificial neural networks, 225 Communication integration, 461–465
Assessment, engineering education, 59 Component, 33, 244, 617, 629
Augmented reality, 215–216 Computational modeling, 193–195
Automated advising, 25 Computer graphics, 114, 181, 215, 316
Automation, 73, 109, 111–112, 163, 175–179, Computer Numerical Control (CNC),
187–192, 250, 327, 345–349, 398 163–166, 169
Autonomic computing, 421, 424 Computer vision, 182, 253–255, 278, 282, 312,
Autonomous system, 517–520, 522 353, 361, 433, 437, 439
Congestion control, 493–496, 503, 579–580
B Congestion window, 493–496, 498
Construccionism, 71
Ballistics system, 327 Constructivism, 71, 73, 133–134, 137
Bandwidth allocation, 479–481, 483 Content-addressable memory (CAM), 405–410
Binary sensor network, 381–386 Context, 12, 20, 22, 43–47, 73, 79–80, 134,
Biometric identification, 303–307 139–143, 161–162, 172, 559, 574–575
Bioorganic compounds, 377–378 Conventional proportional-integral control, 203
Blow-CAST-Fish, 523–528 Converter, 243–246, 257, 259, 288, 373, 376, 393
Bologna process, 151–156 Cooperative maritime network, 449–454
Both side differential transmissions, 357 Coremaking plant, 247–251
Break-even point, 105, 107 Course scheduling, 25
Browser, 27–28, 47, 79–81, 129, 134, 154–155, Covert channel, 503–506
160, 328, 461–465, 557 Critical thinking, 55–58, 86, 121–126, 134
Curriculum management system, 89–94
C
D
Calibration techniques, 185
Cardiac cavity, 351–354 Daemon, 443, 445
CAT instrument, 55–58 Data hiding, 367
Cellular systems, 562 Data protection, 455, 458–459, 460, 560
Chaotic, 367–371 Decomposition method, 427–432
Chromaticity, 362–365 Deconvolution method, 393–398
Chromatic modulation, 361–365 Design of SOHO, 555–560
Circuit analysis, 59–60, 193 Diameters of carbon nanotubes, 275–279
Clamped joints, 417–419 Didactic application, 443, 448
Cloud Migration Strategy, 2 Diesel fuel, 377–378
639
K. Elleithy et al. (eds.), Technological Developments in Networking, Education and Automation,
DOI 10.1007/978-90-481-9151-2, © Springer Science+Business Media B.V. 2010
640 INDEX
L N