IJCSIS Vol. 8 No. 6, September 2010
ISSN 1947-5500
International Journal of
Computer Science
& Information Security
© IJCSIS PUBLICATION 2010
Editorial
Message from Managing Editor
IJCSIS is an open access publishing venue for research in general computer science
and information security.
Target Audience: IT academics, university IT faculties; industry IT departments;
government departments; the mobile industry and computing industry.
Coverage includes: security infrastructures, network security: Internet security,
content protection, cryptography, steganography and formal methods in information
security; computer science, computer applications, multimedia systems, software,
information systems, intelligent systems, web services, data mining, wireless
communication, networking and technologies, innovation technology and management.
The average paper acceptance rate for IJCSIS issues is kept at 25-30% with an aim to
provide selective research work of quality in the areas of computer science and
engineering. Thanks for your contributions in September 2010 issue and we are
grateful to the experienced team of reviewers for providing valuable comments.
Available at http:/ / sites.qooqle.com/ site/ iicsis/
IJCSIS Vol. 8, No. 6, September 2010 Edition
ISSN 1947-5500 © IJCSIS, USA.
Abstracts Indexed by (among others):
Gooole scholar = r\it*^e*r* 5>UUUa
*- ' \— m I I ^J vJ \j ul !■■■■■■ ^anrti * no in* r&r cei* nr*
b&rcMinch engine r&r science
IJCSIS EDITORIAL BOARD
Dr. Gregorio Martinez Perez
Associate Professor - Professor Titular de Universidad, University of Murcia
(UMU), Spain
Dr. M. Emre Celebi,
Assistant Professor, Department of Computer Science, Louisiana State University
in Shreveport, USA
Dr. Yong Li
School of Electronic and Information Engineering, Beijing J iaotong University,
P. R. China
Prof. Hamid Reza Naji
Department of Computer Enigneering, Shahid Beheshti University, Tehran, Iran
Dr. Sanjayjasola
Professor and Dean, School of Information and Communication Technology,
Gautam Buddha University
Dr Riktesh Srivastava
Assistant Professor, Information Systems, Skyline University College, University
City of Sharjah, Sharjah, PO 1797, UAE
Dr. Siddhivinayak Kulkarni
University of Ballarat, Ballarat, Victoria, Australia
Professor (Dr) Mokhtar Beldjehem
Sainte-Anne University, Halifax, NS, Canada
Dr. Alex Pappachen J ames, (Research Fellow)
Queensland Micro-nanotechnology center, Griffith University, Australia
Dr. T.C. Manjunath,
ATRIA Institute of Tech, India.
TABLE OF CONTENTS
1. Paper 23081018: Improvement Dynamic Source Routing Protocol by Localization for Ad hoc
Networks (pp. 1-6)
Mehdi Khazaei
Kermanshah University of Technology, Information Technology Engineering Group, Kermanshah, Iran
2. Paper 31081053: Steganalysis of Reversible Vertical Horizontal Data Hiding Technique (pp. 7-12)
Thorn Ho Thi Huong, Faculty of Information Technology, Haiphong Private University, Haiphong,
Vietnam
Canh Ho Van, Dept. of Professional Technique, Ministry of Public Security, Hanoi, Vietnam
Tien Trinh Nhat, College of Technology, Vietnam National University, Hanoi, Vietnam
3. Paper 21081012: Off-line Handwritten Signature Recognition Using Wavelet Neural Network (pp.
13-21)
Mayada Torek, Computer Science Department, Faculty of Computers and Information Sciences, Mansoura,
Egypt
Taher Hamza, Computer Science Department, Faculty of Computers and Information Sciences, Mansoura,
Egypt
Elsayed Radwan, Computer Science Department, Faculty of Computers and Information Sciences,
Mansoura, Egypt
4. Paper 21081013: A Black-Box Test Case Generation Method (pp. 22-31)
Nicha Kosindrdecha , Autonomous System Research Laboratory, Faculty of Science and Technology,
Assumption University, Bangkok, Thailand
Jirapun Daengdej, Autonomous System Research Laboratory, Faculty of Science and Technology,
Assumption University Bangkok, Thailand
5. Paper 21081014: White-Box Test Reduction Using Case-Based Maintenance (pp. 32-40)
Siripong Roongruangsuwan, Autonomous System Research Laboratory, Faculty of Science and Technology,
Assumption University, Bangkok, Thailand
Jirapun Daengdej, Autonomous System Research Laboratory, Faculty of Science and Technology,
Assumption University, Bangkok, Thailand
6. Paper 23081021: Nat Traversal for Video Streaming Applications (pp. 41-46)
Omar A. Ibraheem #1 , Omer Amer Abouabdalla * 2 , Sureswaran Ramadass n
# National Advanced IPv6 center of Excellence (NAV6), Universiti Sains Malaysia (USM), Pulau penang,
Malaysia
7. Paper 25081026: The Integration of GPS Navigator Device with Vehicles Tracking System for
Rental Cars Firms (pp. 47-51)
Omarah O. Alharaki, KICT, International Islamic University, Kuala Lumpur, Malaysia
Fahad S. Alaieri, KICT, International Islamic University, Kuala Lumpur, Malaysia
Akram M. Zeki, KICT, International Islamic University, Kuala Lumpur, Malaysia
8. Paper 29081036: Process Framework in Global extreme Programming (pp. 52-59)
Ridi Ferdiana, Lukito Edi Nugroho, Paulus Insap Santoso
Department of Electrical Engineering and Information Technology, Gadjah Mada University (UGM)
Yogyakarta, Indonesia
Ahmad Ashari, Department of Computer Science and Electronics, Gadjah Mada University (UGM),
Yogyakarta, Indonesia
9. Paper 29081038: A Hybrid PSO-SVM Approach for Haplotype Tagging SNP Selection Problem
(pp. 60-65)
Min-Hui Lin, Department of Computer Science and Information Engineering, Dahan Institute of
Technology, Sincheng, Hualien County 971, Taiwan, Republic of China
Chun-Liang Leu, Department of Information Technology, Ching Kuo Institute of Management and Health,
Keelung 336, Taiwan, Republic of China
10. Paper 07110912: PAPR Reduction Technique for LTE SC-FDMA Systems Using Root-Raised
Cosine Filter (pp. 66-71)
Md. Masud Rana, Jinsang Kim and Won-Kyung Cho
Deptartment of Electronics and Radio Engineering, Kyung Hee University
1 Seocheon, Kihung, Yongin, Gyeonggi, 449-701, Republic of Korea
11. Paper 22081017: Survey of Routing Protocols and Channel Assignment protocols in Wireless
Mesh Networks (pp. 72-77)
Vivek M Rathod, Suhas J Manangi, Satish E, Saumya Hegde
National Institute of Technology Karnataka - Surathkal
12. Paper 29071044: An Approach For Designing Distributed Real Time Database (pp. 78-87)
Dr. Dhuha Basheer Abdullah, Computer Sciences Dept./Computers Sciences and Mathematics College
/Mosul University, Mosul- Iraq
Ammar Thaher Yaseen, Computer Sciences Dept./Computers Sciences and Mathematics College /Mosul
University, Mosul- Iraq
13. Paper 31081051: Person Identification System using Static-dynamic signatures fusion (pp. 88-92)
Dr. S.A Daramola 1 and Prof.T.S Ibiyemi 2
1 Department of Electrical and Information Engineering, Covenant University Ota Ogun State, Nigeria.
2 Department of Electrical Engineering, University ofllorin, Ilorin, Kwara-State, Nigeria
14. Paper 31081061: Short term flood forecasting using RBF static neural network modeling a
comparative study (pp. 93-98)
Rahul P. Deshmukh, Indian Institute of Technology, Bombay, Powai, Mumbai, India
A. A. Ghatol, Former Vice-Chancellor, Dr. Babasaheb Ambedkar Technological University, Lonere,
Raigad, India
15. Paper 31031083: Analysis of impact of Symmetric Encryption Algorithms in Data Security Model
of Grid Networks (pp. 99-106)
N. Thenmozhi, Department of Computer Science, N.K.R. Govt. Arts College for Women, Namakkal-637
001, India.
M. Madheswaran, Department of Electronics and Communication Engg., Muthayammal Engineering
College, Rasipuram-637 408, India.
16. Paper 09071005: Low Power and Area Consumption Custom Networks-On-Chip Architectures
Using RST Algorithms (pp. 107-115)
1 P.Ezhumali 2 Dr.CArun
1 Professor, Dept of Computer Science Engineering
2Asst. Professor, Dept of Electronics and Communication
Ralalakshmi Engineering College, Thandalam-602 105, Chennai, India
17. Paper 16081005: Prediction of Epileptic form Activity in Brain Electroencephalogram Waves
using Support vector machine (pp. 116-121)
Pavithra Devi S T , M.Phil Research Scholar, PSGR Krishnammal College for Women , Coimbatore
Tamilnadu, India
Vijaya M S, Assistant Professor and Head, GRG School of Applied Computer Technology, PSGR
Krishnammal College for Women, Coimbatore Tamilnadu, India
18. Paper 16081006: Deployment of Intelligent Agents in Cognitive networks (pp. 122-127)
Huda Fatima, Dept. of CS, Jazan University, Jazan, K.S.A
Dr. Sateesh Kumar Pradhan, Dept. of Comp. Engineering, King Khalid University, Abha, K.S.A
MohiuddinAli Khan, Dept. of Comp. Networks, Jazan University, Jazan, K.S.A
Dr. G.N.Dash, Dept. of Comp. Science, Sambalpur University, Orissa, India
19. Paper 21081011: A Performance Study on AES Algorithms (pp. 128-133)
B.D. C.N. Prasad \ P.E.S.N.krishna Prasad \ P Sita Rama Murty 3 andKMadhavi 4
1. Dept. of Computer Applications, PVP Siddardha Institute of Technology, Vijayawada,
2. Dept. ofCSIT, Sri Prakash College of Engineering, Tuni,
3. Dept. ofCSIT, Sri Prakash College of Engineering, Tuni,
4. Dept. ofCSE, Dadi Institute of Technology, Anakapalli,
20. Paper 24091018: Hybrid Fingerprint Image compression and Decompression Technique (pp. 134-
138)
Dr.R.Seshadri, ,B.Tech„M.E,Ph.D, Director, S.V.U. Computer Center S.V. University, Tirupati
Yaswanth Kumar. Avulapti , M.C.A, M.Tech, (PhD), Research Scholar, Dept of Computer Science, S. V.
University, Tirupati
Dr.M.Usha Rani M.C.A, PhD, Associate Professor, Dept. of Computer Science,, SPMW, Tirupati
21. Paper 25081025: Punctured Self -Concatenated Trellis Codes with Iterative Decoding (pp. 139-144)
Labib Francis Gergis, Misr Academy, Mansoura City, Egypt
22. Paper 26081028: Application of Fuzzy Composition Relation For DNA Sequence Classification
(pp. 145-148)
Amrita Priyam, Dept. of Computer Science and Engineering, Birla Institute of Technology, Ranchi, India.
B. M. Karan , G. Sahoo +
+ Dept. of Electrical and Electronics Engineering, + + Dept. of Information Technology
Birla Institute of Technology, Ranchi, India
23. Paper 26081029: Data Security in Mobile Ad Hoc Networks using Genetic Based Biometrics (pp.
149-153)
B. Shanthini, Research Scholar , CSE Department , Anna University , Chennai, India
5. Swamynathan, Assistant Professor, CSE Department , Anna University , Chennai, India
24. Paper 28081031: Effective Multi-Stage Clustering for Inter- and Intra-Cluster Homogeneity (pp.
154-160)
Sunita M. Karad f , Assistant Professor of Computer Engineering, MIT, Pune, India
V.M.Wadhai ff , Professor and Dean of Research, MITSOT, MAE, Pune, India
M.U.Khar at fff , Principle ofPankaj Laddhad IT, Yelgaon, Buldhana, India
Prasad S.Halgaonkar ffff , Faculty of Computer Engineering, MITCOE, Pune, India
Dipti D. Patil fffff , Assistant Professor of Computer Engineering, MITCOE, Pune, India
25. Paper 28081033: A Pilot Based RLS Channel Estimation for LTE SC-FDMA in High Doppler
Spread (pp. 161-166)
M. M. Rana
Department of Electronics and Communication Engineering, Khulna University of Engineering and
Technology, Khunla, Bangladesh
26. Paper 28081034: Priority Based Congestion Control for Multimedia Traffic In 3G Networks (pp.
167-173)
ProfV.S Rathore 1, Neetu Sharma 2, Amit Sharma 3, Durgesh Kumar Mishra 4
123 Department of Computer Engineering, Rajasthan, India
12 Rajasthan College of Engineering for women, Rajasthan, India
3 Shri Balagi College of Engineering & Technology, Rajasthan, India
4 Acropolis Institute of Technology and Research, Indore, MP, India
27. Paper 31081050: Adaptive Sub-block ARQ techniques for wireless networks (pp. 174-178)
A. N. Kemkar, Member, ISTE and Dr. T. R. Sontakke, Member, ISTE
28. Paper 30071074: Trigon-based Authentication Service Creation with Globus Middleware (pp.
179-185)
Ruckmani V , Ramakrishna Engineering College, Coimbatore, India
Anitha Kumari K , PSG College of Technology, Coimbatore, India
Sudha Sadasivam G , PSG College of Technology, Coimbatore, India
Dhaarini M P , PSG College of Technology, Coimbatore, India
29. Paper 30081040: Performance Evaluation of Speaker Identification for Partial Coefficients of
Transformed Full, Block and Row Mean of Speech Spectrogram using DCT, WALSH and HAAR
(pp. 186-198)
Dr. H. B. Kekre, Senior Professor, MPSTME, SVKM's NMIMS University, Mumbai, 400-056, India
Dr. Tanuja K. Sarode, Assistant Professor, Thadomal Shahani Engg. College, Bandra (W), Mumbai, 400-
050, India
Shachi J. Natu, Lecturer, Thadomal Shahani Engg. College, Bandra (W), Mumbai, 400-050, India
Prachi J. Natu, Assistant Professor, GVAIET, Shelu, Karjat 410201, India
30. Paper 30081041: A Research Proposal for Mitigating DoS Attacks in IP-based Networks (pp. 199-
201)
Sakharam Lokhande, Assistant Professor, School of Computational Science, Swami Ramanand Teerth
Marathwada University, Nanded, MS, India, 431606.
Parag Bhalchandra **, Assistant Professor, School of Computational Science, Swami Ramanand Teerth
Marathwada University, Nanded, MS, India, 431606.
Nilesh Deshmukh, Assistant Professor , School of Computational Science, Swami Ramanand Teerth
Marathwada University, Nanded, MS, India, 431606.
Dr. Santosh Khamitkar, Assistant Professor , School of Computational Science, Swami Ramanand Teerth
Marathwada University, Nanded, MS, India, 431606.
Santosh Phulari, Assistant Professor, School of Computational Science, Swami Ramanand Teerth
Marathwada University, Nanded, MS, India, 431606.
Ravindra Rathod, Assistant Professor, School of Computational Science, Swami Ramanand Teerth
Marathwada University, Nanded, MS, India, 431606
31. Paper 30081042: An Efficient and Minimum Cost Topology Construction for Rural Wireless
Mesh Networks (pp. 202-209)
Prof. V. Anuratha, H.O.D - PG. Comp. Science, Sree Saraswathi Thyagaraja college, Pollachi, Tamil
Nadu, India
Dr. P. Sivaprakasam , Associate Professor, Sri Vasavi college of Arts & Science, Erode, Tamil Nadu,
India
32. Paper 31071081: Reinforcement Learning by Comparing Immediate Reward (pp. 210-214)
Punit Pandey, Department Of Computer Science and Engineering, Jaypee University Of Engineering And
Technology
Dr. Shishir Kumar, Department Of Computer Science and Engineering, Jaypee University Of Engineering
And Technology
Deepshikha Pandey, Department Of Computer Science and Engineering, Jaypee University Of Engineering
And Technology
33. Paper 31081043: Information realization with statistical predictive inferences and coding form
(pp. 215-220)
D. Mukherjee, Sir Padampat Singhania University, Udaipur-313601,Rajasthan,India
P.Chakrabarti* , AKhanna , V.Gupta
Sir Padampat Singhania University, Udaipur-313601,Rajasthan,India
34. Paper 31081045: Scaling Apriori for Association Rule Mining using Mutual Information Based
Entropy (pp. 221-227)
S. Prakash, Research Scholar, Sasurie College of Engineering, Vijayamangalam,Erode(DT), Tamilnadu,
India.
Dr. R. M. S. Parvathi M.E.(CSE),Ph.D., Principal, Sengunthar College ofEngg.for Women, Tiruchengode.
Tamilnadu, India.
35. Paper 31081046: Clustering of High- Volume Data Streams In Network Traffic (pp. 229-233)
M. Vijayakumar, Research Scholar, Sasurie College of Engineering, Vijayamangalam, Erode(Dt) ,
Tamilnadu, India.
Dr. R.M.S. Parvathi M.E.(CSE),Ph.D., Principal, Sengunthar College of Engg.for Women, Tiruchengode.
Tamilnadu, India.
36. Paper 31081049: A.R.Q. techniques using Sub-block retransmission for wireless networks (pp.
234-237)
A. N. Kemkar, Member, ISTE and Dr. T. R. Sontakke, Member, ISTE
37. Paper 22081015: Performance Analysis of Delay in Optical Packet Switching Using Various
Traffic Patterns (pp. 238-244)
AKavitha , IT dept, Chettinad College of Engineering & Technology, Karur, Tamilnadu, India
V.Rajamni, Indra Ganesan College of Engineering, Trichy, Tamilnadu, India
P. Anandhakumar, IT Dept, Madras Institute of Technology, Chennai, Tamilnadu, India
38. Paper 29081039: A Feedback Design for Rotation Invariant Feature Extraction in
Implementation with Iris Credentials (pp. 245-254)
M. Sankari, Department of Computer Applications, Nehru Institute of Engineering and Technology,
Coimbatore, India.
R Bremananth, School ofEEE, Information Engg. (Div.), Nanyang Technological University, Singapore -
639798.
39. Paper 31081047: Empirical Mode Decomposition Analysis of Heart Rate Variability (pp. 255-258)
C. Santhi, M.E., Assistant Professor, Electronics and Communication Engineering, Government College of
Technology, Coimbatore-641 013
N. Kumaravel, Ph.D, Professor, Head of the Department, Electronics and Communication Engineering,
Anna University, Chennai-600 025.
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
Improvement Dynamic Source Routing Protocol by
Localization for Ad hoc Networks
Mehdi Khazaei
Kermanshah University of Technology
Information Technology Engineering Group
Kermanshah, Iran .
Abstract-Ad hoc networks are temporary networks with a
dynamic topology which don't have any established
infrastructure or centralized administration. Consequently, in
recent years many researchers have focused on these networks.
These networks need efficient routing protocols in terms of
Quality of Services (QOS) metrics. Ad hoc networks suffer from
frequent and rapid topology changes that cause many challenges
in their routing. Most of the routing protocols like this proposed
protocol try to find a route between source and destination
nodes and when any route is expired, a new route would be
formed. Rapid route reconstruction may cause the network
inefficiency. Therefore, we have to decrease this processes. The
proposed protocol as DSR routing protocol build one routes
between source and destination but create backup routes during
the route reply process, route maintenance process and use local
recovery process in order to improve the data transfer and
attended to QOS. The protocol performance is demonstrated by
using the simulation results obtain from the global mobile
simulation software (Glomosim). The experimental results show
that this protocol can decrease the packet loss ratio and increase
data transfer rather than DSR that, it is useful for the
applications that need a high level of reliability.
Keywords; Protocol, Routing, Local Recovery, Mobile Ad-hoc
Networks
I. Introduction
Routing in ad hoc networks is a very challenging issue due
to nodes mobility, dynamic topology, frequent link breakage,
limitation of nodes (memory, battery, bandwidth, and
processing power), and limited transmission range of the node
and lack of central point like base stations or servers. On the
other hand, there are a lot of performance metrics and quality
services which should be satisfied in an ad hoc network like
end -to-end data throughput, average end-to-end data delay,
packet loss ratio, Normalized Routing Load, Packet Delivery
Ratio, and route optimality. Each protocol can satisfy some of
these metrics and has some drawbacks in terms of other
metrics. Furthermore, due to the nature of ad hoc networks
(distributed and cooperated routing), even for a fixed metric,
each protocol can show a different performance with different
networks features like number of mobile nodes, mobility of
nodes, pause time and.... So by comparing between different
ad hoc routing protocols we can extract very important
information about the performance of these protocols in the
Different situations. In the other hands, the nodes mobility and
the probability of links failure may cause the fault tolerance
Issues more important for routing problem in ad hoc network
therefore, each routing protocol should be fault tolerant in
probable route failures [1].
Routing protocols in conventional wired networks are
usually based upon either distance vector or link state routing
algorithms as a DSDV [2], CGSR [2] and FSR [2]. These
algorithms require periodic routing advertisements to be
broadcast by each router. These conventional routing
algorithms are clearly not efficient for type of dynamic
changes which may occur in an ad-hoc network [2, 3]. A new
class of on-demand routing protocols e.g., DSR [4, 5], TORA
[2], AODV [6, 7]) for mobile ad hoc networks has been
developed with the goal of minimizing the routing overhead.
These Protocols reactively discover and maintain only the
needed routes, in contrast to proactive protocols (e.g., DSDV
[2]) which maintain all routes regardless of their usage. The
key characteristic of an on-demand protocol is the source-
initiated route discovery procedure. Whenever a traffic source
needs a route, it initiates a Route discovery process by sending
a route request for the destination (typically via a network-
wide flood) and Waits for a route reply. Each route discovery
flood is Associated with significant latency and overhead. This
is particularly true for large networks. Therefore, for on-
demand routing to be effective, it is desirable to keep the route
discovery frequency low [8].
Single route routing allows the establishment of one route
between a source and single destination node. Because of node
mobility, the route may be broken frequently; therefore,
having replacement route in cache memory to transmit data
will improve the fault tolerance and higher aggregate
bandwidth in these networks. Beside of this, by repairing the
broken routes locally, the number of route rediscovery
processes can be decreased. This paper improves the fault
tolerance and increase reliability by obtain replacement routes
in RREP 1 and RRER 2 processes and local recovery process
together. This optimization is done on the DSR protocol.
The rest of this paper is organized as follows. In section II
the DSR protocol is explained. Section III deals with the
related works and Section IV describe the proposed protocol
mechanism in detail. Performance evaluation by simulation is
presented in section V and concluding remarks are made in
section VI.
1 Route reply
2 Route error
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
II. Dynamic Source Routing Protocol (DSR)
DSR is an on-demand routing protocol for ad hoc
networks. Like any source routing protocol, in DSR the source
includes the full route in the packets' header. The intermediate
nodes use this to forward packets towards the destination and
maintain a route cache containing routes to other nodes. In
following subsections DSR operation are briefly described [4].
A. Route discovery
If the source does not have a route to the destination in its
route cache, it broadcasts a RREQ 3 message specifying the
destination node for which the route is requested. The RREQ
message includes a route record which specifies the sequence
of nodes traversed by the message. When an intermediate
node receives a RREQ, it checks to see if it is already in the
Route record. If it is, it drops the message. This is done to
prevent routing loops. If the intermediate node had received
the RREQ before, then it also drops the message. The
intermediate node forwards the RREQ to the next hop
according to the route specified in the header. When the
destination receives the RREQ, it sends back a route reply
message. If the destination has a route to the source in its route
cache, then it can send a RREP message a long this route.
Otherwise, the RREP message can be sent along the reverse
route back to the source. Intermediate nodes may also use their
route cache to reply To RREQs. If an intermediate node has a
route to the destination in its cache, then it can append the
route to the route record in the RREQ, and send an RREP back
to the source containing this route. This can help limit
flooding of The RREQ. However, if the cached route is out-
of-date it can result in the source receiving stale routes [4].
B. Route maintenance
When a node detects a broken link while trying to forward
a packet to the next hop, it sends a RERR message back to the
source containing the link in error. When an RERR message is
received, all routes containing the link in error are deleted at
that node [4].
III. Related Works
Ad hoc routing protocols such as ADOV, DSR, DSDV and
OLSR have been investigated on the ad hoc networks in the
past few years. The investigations of the performance of these
protocols on the ad hoc networks have produced many useful
results. However, we have seen very limited findings of how
these Ad-hoc routing protocols perform on wireless ad hoc
networks. Nonetheless, we can see many attempts at
developing routing protocols for ad hoc networks under the
different deployment of ad hoc networks [8, 9 and 10]. In
following are brought some of these attempts.
SMR is an on demand routing protocol that uses
maximally disjoint routes to transmit data packets. Unlike
DSR, intermediate nodes do not allow to send RREP packets
back to the source instead, only destination nodes reply to the
RREQ packets and selects maximally-disjoint routes [11].
MP-DSR is a multi-route QOS aware extension to DSR. It
focuses on a QOS metric, end-to-end reliability. End-to-end
Route Request
Reliability is defined as the probability of sending data
successfully from the source to the destination node within a
time window. MP-DSR selects a set of routes that satisfy a
specific end-to-end reliability Requirement [12].
MSR is attempts to minimize the end-to-end delay for
sending a data from source to destination by using multi-route
routing and intelligent traffic allocation [13].
CHAMP is multi-route protocol that uses round-robin
traffic allocation to keep routes fresh. It also employs
cooperative packet caching to improve fault tolerance and
takes advantage of temporal locality in routing, where a
dropped packet is a recently sent packet [9].
The local recovery techniques have been used in some
routing protocols for route maintenance processes. This
technique aims to reduce the frequency of RREQ floods
triggered by nodes that are located in the broken routes [14].
SLR is one of these routing protocols. It modifies DSR,
using a new route recovery mechanism called bypass routing.
Bypass routing utilizes both route caches and local error
recovery techniques during failures to reduce the control
overhead [15].
LRR is also another routing protocol that uses local
recovery techniques. In this protocol the information of next-
to-next (NN) node is stored at each intermediate node along
the route. After detecting a broken link by an upstream node, it
sends out the non-propagating requests to find another node
which is in contact with itself and the NN node on the route;
therefore the routes can be repaired locally in the shortest
possible time [16].
MRFT protocol improves fault tolerance in DSR and SMR
protocols. To achieve the goal of decreasing the packet loss
ratio and increasing fault- tolerance, MRFT uses both multi-
route and local recovery techniques together [17].
IV. The Proposed Protocol
This paper proposes IM-DSR 4 protocol to improve fault
tolerance and QOS in DSR protocol. To achieve the goal of
decreasing the packet loss ratio and increasing fault-tolerance,
IM-DSR uses local recovery techniques and alternate route
during route reply and route maintenance that reliability in the
network would be increased. IM-DSR modifying the route
discovery, route reply and route maintenance processes in
DSR. The IM-DSR protocol is including route discovery,
route reply, route maintenance and local recovery processes
that discussed in the following subsection.
A. Route Discovery Process
IM-DSR is an on demand routing protocol that builds
single route using request/reply cycles. When the source node
needs to send data to the destination but no route information
is known, it floods RREQ packets over the entire network.
When an intermediate node receives a RREQ that is not a
duplicate, it appends its ID to the packet and rebroadcasts it. In
IM-DSR all of the duplicate RREQs that are received by
intermediate nodes are dropped. In IM-DSR, intermediate
4 Improvement-DSR
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
Nodes are allowed to send RREPs back to the source even
when they have route information to the destination in their
route caches.
B. Route Reply Process
When receiving the first RREQ, the destination sends a
RREP back to the source. After that, the destination node
consumes other RREQs. The Route-Number of the RREP is
one.
After receiving RREP packet by the intermediate nodes, if
it has not route with same length to destination node, they
store the routes in their route caches. The Route-Number of
this routes are zero and used in route maintenance process for
improving the break routes also sending data if there is not
main route.
Look at the Fig. 1, suppose that node (H) sends the RREP
to the source node (A), a route is found and sent to node (A)
by RREP is A-C-D-G-F-H. Now, suppose that the RREP is
received by node (C) which is middle node. Node (C) saves
the routes to destination (H) which is C-D-G-F-H, additionally
node (C) save C-D, C-D-G and C-D-G-F routes in the route
caches.
When the source node receive the RREP, it will store the
route and use that for transmit data.
C. Route Maintenance and Local Recovery Processes
During a transmission session, a problem such as node
mobility, or low battery power might be raised, which can lead
to break an existing route and lose route connectivity. This
may force a route rediscovery process by flooding RREQs
over the network. To avoid this phenomenon, IM-DSR uses
following mechanism that one of them is local recovery
techniques. Using local recovery techniques is very useful
despite they consume the limited power of each nodes.
Suppose that a node finds a broken link, while sending a
packet. At first, it seeks the route cache and deletes all routes
include the broken link, and then according to kind of the
packet one of the following items is done:
• If transitional packet would be a RREQ, the node would
not send RRER to the source node.
• If transitional packet would be a RREP, send RRER to the
node which makes the RREP.
^ G
E
Information In Cache Memory
N etwork Connections
Figure 1 . Routes Structure in an Ad-hoc Network
• If transitional packet is a RRER, it would examine how
many times the packet would be saved. If it was the first
time, the meaning is, the packet would be saved by
examining a route cache and finding alternate route. The
RRER is sent to destination through that route then the
RRER is made and it will report the broken link to the
source of RRER. If it were not first time or if alternate
route were not in the route cache of node, the RRER
would be deleted and only a RRER would be sent to the
source node. Therefore RRER is saved only for one time
by the IM-DSR protocol.
• If transitional packet would be data, it would examine
how many times the packet would be saved. If this time
were less than three, the data packet would be sent by
examining their route cache and alternate route then it
will send a RRER to the source node. If these times were
more than three or if alternate route was not in the route
cache of node the data packet would be deleted and only a
RRER would be sent to the source node. Therefore data
packet is saved for three times by the IM-DSR protocol.
If very data packet passed the same route towards
destination node and they faced the broken link (while
sending), the node which recognized an error, for every data
packet send a RRER to the source node. In order to avoid this
item every node before sending RRER to source node,
examine this is a first RRER or not. If it was not send, a new
RRER send to source node.
Every node which recognizes the broken link and makes
the RRER, examined the route cache in order to find alternate
route and put it in the RRER, which the node that received the
RRER, replaces the route in the RRER with the previous
invalid route in the route cache.
Fig. 2 shows this matter. The source node (A) sends data to
the destination node (H) through A-C-D-H. When node (D)
sends the data packet, it will find the failure in node (H). By
examine the route cache, it chooses the alternate route D-G-F-
H, hence, the data packet is sent to destination node (H)
through this route, then RRER is made and it is sent to the
node (A). This packet includes the alternate route D-G-F-H.
The node (A) receives the RRER and deletes A-C-D-H from
route cache and replaces A-C-D-G-F-H.
Every middle and source node which receives the RRER,
examine those route in route cache which includes the broken
link and should be deleted from cache and if packet included
alternate route, exploited that and saved in route cache with
number two. If in buffer, data packet waiting to send toward
alternate route destination, it will send through that route.
Such as the Fig. 2, while passing the RRER, node (C) adds C-
D-G-F-H to the route cache.
If a node who detected a broken link cannot find any
alternate route in its route cache, so it drops the data packet
and sends a RERR without any repaired route to the source.
After that, because of performing local recovery process by
the node that detects the broken link, the source node does not
trigger the rediscovery process immediately. After detecting
the broken link, node sends a RERR to the source and starts
the local recovery process simultaneously. To repair the route
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
J?
B
E
Information In Cache Memory
N etwork C onne ctions
Figure 2. Ad-hoc Network Structure, when occur break link
Locally, node triggers a local query to its neighbors. The
neighbors reply if they have any valid route to the destination.
When node receives the RRP 5 , it repairs the primary route and
then sends a RRP back to the source, like shown in the Fig. 3.
In Fig. 3 when node (D) finds the failure in node (G), First
examined the route cache for replacement route, if not found
then begun local recovery request which node (F) reply with
F-G-H as repaired route. When node (D) receives this route
update its memory cache and sends data from this route to
destination and send RRP to the source node that here is node
(A). When node (A) receives the RRP deletes A-C-D-G-H
from route cache and replaces A-C-D-F-G-H. Every middle
node which receives RRP, examine if included repaired route,
then exploited that and saved in route cache with number two.
In Fig. 3 while passing RRP, node (C) adds C-D-F-G-H to the
route cache more node (C) adds C-D, C-D-F and C-D-F-G to
route cache.
If the source node has not primary route to send data, it
will use the repaired route. If no data packet containing
repaired route was reported to the source node for a certain
amount of time, then it sends a new RREQ to the destination
node. After receiving the new RREQ by the destination node,
it performs the route rediscovery process that was described in
subsection A.
In the IM-DSR protocol, route with Route-Number one is
main route. Route with Route-number zero are obtain by route
reply process and route with Route-Number two are those
which are obtain in route maintenance process and known as
repaired route.
When the source node wants to send data to a destination,
it tries to use the primary, which has the highest priority (the
routes with route-Number one) at first, if not exist, it will use
obtain route (the route with Route-Number zero) otherwise
use repaired route (the route with Route-Number two) to send
data.
V. Performance Evaluation
A. Simulation Environment
In order to demonstrate the effectiveness of IM-DSR
protocol, we evaluate our proposed protocol and compare its
performance to the DSR (uni-route). We have implemented
IM-DSR protocol using the Global Mobile Simulation library
4f
B
^ G
E
Information In Cache Memory
N etwork C onne ctions
■ Rep aire d Route by Lo c al Re c o very Process
^ ~~ — ~~ ~ Send Repaired Route packet
I igure 3. Ad-hoc Network Structure, when Occur Local Recovery
(GLOMOSIM) [18]. The Simulation environment consists of
50 numbers of nodes in a rectangular region of size 1500
meters by 1500 meters.
The nodes are randomly placed in the region and each of
them has a radio propagation range of 250 meters. 200
constant bit rate (CBR) flows are deployed for data
transmission. Simulation time is 300 seconds. The random
waypoint model is chosen as the node mobility model. All
data packets are 512 bytes. Band width is 2 Mbps and
simulation done for 0, 1, 3, 5 and 10 second as stop time.
Minimum and maximum speed for nodes are m/s and 30
m/s. IEEE 802.1 1 selected for MAC layer protocol.
B. Simulation Results
Fig. 4 shows packet delivery ratio (PDR) in every two
protocol. It is defined as ratio of the number of data packets
delivered to the destinations generated by sources. In DSR
protocol if sending node of data has not any routes to send
data, it would start route request process by RREQ. In every
request finds a route and if it was invalid, the route request
process begins again. The IM-DSR protocol against DSR
protocol obtains routes in route reply and Maintenance process
and used them for sending data. This would lead to increase
the packet delivery ratio in IM-DSR protocol.
Fig. 5 shows the number of RREQ (NRQ) in every two
protocols. This number is sum of the RREQ in request
process. Fig. 5 shows that this number in DSR protocol is
greater than IM-DSR protocol because in every route
discovery find only one route and if this route would be
invalid begin route discovery again. But IM-DSR finds routes
in route reply, maintenance and local recovery processes,
hence reduces this number.
' Route Repaired Packet
Stop Tirne
Figure 4. Packet Delivery Ratio (PDR)
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
3:C'
250
a 200
150
100
50
3
Stop Time
5
a 5
-DSR
- IM-DSR
Stop Time
Figure 5. Number of RREQ (NRQ)
Fig. 6 shows sum of the hop-count (SHC) in the network.
This number is summation of the all routes hop count that
lower value leads to less delay in network. This number is less
in IM-DSR protocol because of less route request process and
found optimal routes.
Fig. 7 shows the number of broken links (NBL) in the
network. Problem such as node mobility, low battery power or
congestion might be raised, which can lead to break an
existing route and lose route connectivity. Number of broken
link in IM-DSR protocol is less than DSR protocol because of
used alternate routes that making in route request, reply and
maintenance processes and used local recovery process.
Fig. 8 shows an average of delay (AVD) in two protocols.
Because of sending data from route with long hop-count and
data packets waiting more time in buffer, delay in DSR
protocol is more than IM-DSR protocol. The IM-DSR
protocol which, between the source and destination, selected
optimal routes and saved multi-route node decreases the delay
in comparisons with DSR protocol.
800 i
700
600
500
ffi 400 -
-■-IM-DSR
300
200 ■
100 -
D
1
3
Stop Tune
;
10
Figure 6. Sum of Hop-Count (SHC)
4000
3500
3000
2500
I
j 2000
15 00
1000
500
-DSR
-IM-DSR
Figure 8. Average of Delay (AVD)
VI. Conclusion
In this paper, we proposed a new routing protocol called
IM-DSR to provide higher QOS in ad hoc networks. This
protocol is an extension of DSR to increase the reliability by
modifying the route discovery and route maintenance
processes in DSR also added the local recovery techniques to
DSR. The simulation results show that IM-DSR is very
effective in decreasing the packet loss and also increasing the
fault tolerance ad hoc networks. In all of the cases, IM-DSR
has the higher packet delivery ratio than the DSR protocol
while improving the overhead of route maintenance,
maintaining acceptable overhead. Therefore the proposed
routing protocol is very useful for the applications that need a
high level of reliability.
References
[1] V. Nazari Talooki, J. Rodriguez, "Quality of Service for Flat Routing
Protocols in Mobile Ad hoc Networks", Mobimedia'09, London, UK,
September, 2009.
[2] E. M. Royer, "A Review of current Routing Protocols for Ad Hoc Mobile
Wireless Networks", IEEE Personal Communications, vol. 6, no. 2, pp.
46-55, 1999.
'The handbook of ad hoc wireless networks", CRC press,
Figure 7. Number of Broken Link (NBL)
[3] M. IILyas,
2003.
[4] D. Johnson, Et. al, "The Dynamic Source Routing Protocol for Mobile Ad
Hoc Network" 1 1 I m I i I lull www.ietf.org/internetdrafts/ draft-ietf
manet-dsr-10.txt (July 2004).
[5] D. B. Johnson, D. A. Maltz, "Dynamic Source Routing in Ad Hoc
Wireless Networks", Mobile Computing, Edited by Imielinski, Korth,
Chapter 5, pp. 153-181, Kluwer Academic Publishers, 1996.
[6] C. Perkins, E. Royer, "Ad Hoc On-Demand Distance Vector (AODV)
Routing" 1 1 1 1 1 1 i l 1 1 1 1 1 www.ietf.org/internet-drafts/draft-ietfmanet-
aodv-13.txt (Feb 2003).
[7] C. Perkins, E. Royer, "Ad-hoc On-demand Distance Vector Routing", 2nd
IEEE Workshop on Mobile Computing Systems and Applications, New
Orleans, LA, United States, pp. 90-100, Feb. 1999.
[8] Q. Jiang, D. Manivannan, "Routing protocols for Sensor networks", IEEE,
pages 93-98, 2004.
[9]- S. Mueller, R.P. Tsang and D. Ghosal, "Multipath Routing in Mobile Ad
hoc Networks Issues and Challenges", Proceedings of CA
USA,27,USA,2007.
[10] K. Wu, J. Harms, "On-Demand Multipath Routing for Mobile Ad Hoc
Networks", 4th European Personal Mobile Communications Conference
(EPMCC 01), Vienna, Austria, Feb. 2001.
[11] S. Ju-Lee, M. Gerla, "Split Multipath Routing with Maximally Disjoint
Paths in Ad Hoc Networks", IEEE International Conference on
Communications, vol. 10, pp. 3201-3205, Helsinki, Jun. 2001.
[12] R. Leung, E. Poon, A. C. Chan, B.Li, "MP-DSR: A QoS-Aware Multi
path Dynamic Source Routing Protocol for Wireless Ad-Hoc Networks",
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
26th Annual IEEE International Conference on Local Computer
Networks (LCN2001), pp. 132-141, Tampa, FL, United States, Nov.
2001.
[13] X. Li, L. Cuthbert, "On-Demand Node-Disjoint Multipath Routing in
Wireless Ad Hoc Networks", 29th Annual IEEE International
Conference on Local Computer Networks (LCN2004), Tampa, FL,
United States, pp. 419-420, Nov. 2004.
[14] E. Esmaeili, P. Akhlaghi, M. Dehghan and M. Fathi, "A New Multi-Path
Routing Algorithm with Local Recovery Capability in Mobile Ad hoc
Networks", Fifth International Symposium, Patras, Greece, July, 2006.
[15] C. Sengul, R. Kravets, "Bypass Routing: an On-Demand Local Recovery
Protocol for Ad Hoc Networks", Ad Hoc Networks, In press (Available
online 2005).
[16] R. Duggirala, R. Gupta, Q. Zeng, D. P. Agrawal, "Performance
Enhancements of Ad Hoc Networks with Localized Route Repair",
IEEE Transactions on Computers, vol. 52, no. 7, pp. 854-861, 2003.
[17] M. khazaei, R. berangi "A multi-path routing protocol with fault
tolerance in mobile ad hoc networks ", Proceedings of IEEE
international CSI, 14th, tehran, iran, Oct, 2009.
[18] UCLA Parallel Computing Laboratory and wireless Adaptive Mobility
Laboratory, GloMoSim: A Scalable Simulation Environment for
Wireless and Wired Network.
1
Mehdi khazaei received the bachelor's and master's
degrees from Iran University of Science and
Technology (IUST), in 2004 and 2007, respectively.
Currently He is lectureship in Kermanshah University
of Technology (KUT) and his researches focused on
wireless networks, especially ad hoc networks.
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol 8, No. 6, September 2010
Steganalysis of Reversible Vertical Horizontal Data
Hiding Technique
Thorn Ho Thi Huong,
Faculty of Information Technology,
Haiphong Private University,
Haiphong, Vietnam
Canh Ho Van
Dept. of Professional Technique,
Ministry of Public Security,
Hanoi, Vietnam
Tien Trinh Nhat
College of Technology,
Vietnam National University,
Hanoi, Vietnam
Abstract — This paper proposes a steganalysis scheme for
detecting the reversible vertical horizontal (RVH) data hiding [1],
The RVH scheme was introduced in the IJCSIS International
Journal Vol. 7, No. 3, March 2010. In the RVH data hiding, the
message bits are embedded into cover-image by two embedding
phases: the horizontal embedding procedure HEm and the
vertical embedding procedure VEm. The pixel pairs belonging to
the horizontally embeddable and vertically embeddable pixel pair
domain are transformed to mark message bits. Through analysis,
we detect out that, the two histograms of LSB scanning
horizontally and vertically vary from a stego-image to the cover
image. Based on this observation, we design a specific steganlytic
method for attacking the RVH steganography. Experimental
results show the detection accuracies of the steganography with
various embedding rates are acceptable. The proposed technique
can be applied in detecting the misuse of steganographic
technology in malicious activities.
Keywords-Steganography, steganalysis, watermarking, cover image,
stego image, payload, reversible data hiding.
I.
Introduction
Steganography is [1, 3, 4, 5] is the art and science of
concealed communication. The basic concept is to hide the
very existence of the secret message. Digital object such as a
text, image, video, or audio segment can be used as the cover
data. To obtain acceptable hiding payload and keep fidelity of
the stego-image, the LSB replacement techniques [2,4 or other
references] are popular and widely studied in the literature.
These methods usually hide more data in image areas with
higher spatial variations. Reversible steganography [1,3-5] is
one of the interesting branches of steganographic technology in
which the original cover image can be reconstructed without
any loss.
Steganalysis is the counterpart of steganography, the goal
of the steganalysis is to detect the hidden message,
equivalently, to discriminate the stego object from the non-
stego-object. The steganalysis techniques proposed in the
literature can be classified into two categories: the universal
steganalysis which is designed to detect the hidden message
embedded with various data embedding algorithms such as a
technique proposed in [6] is used to attack the LSB
steganography, and the specific steganalysis which is designed
to attack a specific steganography technique such as a
steganalytic method was presented in [2] for detecting stego-
images using the method proposed in [3].
In this paper, we proposed a steganalytic scheme to detect
the RVH watermarking scheme introduced in brief in the
abstract. Our experimental results show the feasibility of the
proposed method. It is useful in detecting malicious activities
on stego-images and also suggests a design consideration for
future development of steganographic techniques. The rest of
this paper is organized as follows. In the next section, we
present again the RVH scheme in brief. Section III describes
the proposed steganalytic method. Experimental results are
given in section IV, and conclusions are made finally in
Section V.
II. Review of the RVH data hiding scheme
In the steganographic method proposed in [1] used the
multiple embedding strategies to improve the image quality
and the embedding capacity. Basically, this method embeds
each message bit b of the secret bit stream into each grayscale
cover pixel pair of a grayscale cover image in raster scan order.
This scheme includes two main stages, namely, the horizontal
embedding procedure HEm and the vertical embedding
procedure VEm. For the HEm procedure, the input image is
horizontally scanned in raster scan order (i.e., from left to right
and top to bottom) to gather two neighboring pixels x and y
into a cover pixel pair (x, y). If y is an odd value, then the cover
pixel pair (x, y) is defined as a horizontally embeddable pixel
pair. Otherwise, the cover pixel pair (x, y) is defined as a
horizontally non-embeddable pixel pair. For the VEm
procedure, the input image is vertically scanned in raster scan
order to group two neighboring pixels u and v into a pixel pair
(u, v). If v is an even value, then the pixel pair (u, v) is defined
as a vertically embeddable pixel pair. Otherwise, the pixel pair
(u, v) is defined as a vertically non-embeddable pixel.
The secret bit sequence S is divided into two subsequence
SI and S2. The bit stream Bl is created by concatenating the
secret subsequence SI and the auxiliary data bit stream Al
(i.e., B1=S1||A1). Similarly, the bit stream B2= S2||A2. The
generation of Al and A2 will be described latter. The overview
of the RVH embedding process is shown in Fig. 1.
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
Fig. 1. Embedding phase of RVH steganographic system [1]
Firstly, the bit sequence Bl is horizontally embedded into
O by using the HEm procedure to obtain the output image T
sized H x W pixels. Secondly, the compressed location map
CM1 whose length is LCI (will be described later), is
embedded in to T by using the least significant bit (LSB)
replacement technique to obtain the output image U with size
of H x W pixels. Thirdly, the bit sequence B2 is vertically
embedded into U by using the VEm procedure to get the output
image V with size of H x W pixels. Fourthly, the compressed
location map CM2 whose length is LC2 is embedded into V by
using the LSB replacement technique to get the final stego
image with size of H x W pixels.
Each bit b in stream Bl is horizontally embedded into each
horizontally embeddable pixel pair (x, y) at a time by using the
horizontal embedding rule HR defined below until the whole
bit stream Bl is completely marked into O to obtain the output
image T.
Each bit b in B2 is vertically embedded into each vertically
embeddable pixel pair (u, v) at a time by using the vertical
embedding rule VR defines below until the entire bit sequence
B2 is concealed into U to get the output image V.
The horizontal embedding rule HR: For each pair (x, y), we
apply the following embedding rules:
• HR1: If the tobeembedded bit b=l, then the stego
pixel pair is unchanged by (xO, yO) = (x, y).
• HR2: If the tobeembedded bit b=0, then the stego
pixel pair is changed by (xO, yO) = (x, y-1).
The vertical embedding rule VR: For each pair (u,v), we
apply the following embedding rules:
• VR1 : If the tobeembedded bit b=0, then the stego
pixel pair is unchanged by (uO, yO) = (u,v).
• VR2: If the tobeembedded bit b=l, then the stego
pixel pair is changed by (uO, yO) = (u, v+1).
It is noted that the rule HR and VR don't cause the
underflow and overflow problem. That is the changed pixel
pairs are assured to fall in the allowable range [0,255].
The auxiliary data bit sequence Al is actually the LSBs of
the first LCI (LCI is the length of the compressed location
map CM1 ended with the unique end of map indicator EOM1)
pixels in the image T and generated as follows. Initially, Bl is
(IJCSIS) International Journal of Computer Science and Information Security,
Vol 8, No. 6, September 2010
equal to SI (i.e., B1=S1). During the execution of the HEm
procedure, for the first LCI pixels in O, when each pixel has
been processed for embedding, its LSB is taken as an auxiliary
data bit of Al and appended to the end of Bl. That is, Bl is
gradually grown until the LCI auxiliary data bits in Al are
concatenated into Bl. Finally, the tobeembedded bit stream
is B1=S1||A1, which is completely embedded into O.
Similar to the generation of Al, the auxiliary data stream
A2 is actually the LSBs of the first LC2 (LC2 is the length of
the compressed location map CM2 ended with the unique end
of map indicator EOM2) pixels in the image V and generated
as follows. B2 initially equals the secret bit sequence S2.
During the execution of the procedure VEm, for the first LC2
pixels in the image U, when each pixel has been processed for
embedding, its LSB is taken as an auxiliary data bit of A2 and
append to the end of B2 until the LC2 auxiliary data bits in A2
are concatenated into B2. Finally, the information bit sequence
is B2=S2||A2, which is fully marked into the image U.
For the purposes of extracting Bl and recovering O, a
location map HL sized H x (W/2) is needed to record the
positions of the horizontally embeddable pixel pair (x, y) in O.
The location map HL is a one-bit bitmap. All the entries of HL
are initialized to 0. If cover pixel pair (x, y) is the horizontally
embeddable pixel pair, then the corresponding entry of HL is
set to be 1. Next, the location map HL is losslessly compressed
by using the JBIG2 codec (Howard et al, 1998 [8]) or an
arithmetic coding toolkit (Carpenter, 2002 [7]) to obtain the
compressed location map CM1 whose length is LCI. The
compressed location map CM1 is embedded into the image T
by using the LSB replacement technique as mentioned above.
Similarly, for the purposes of extracting B2 and recovering the
image U, we also need a location map VL sized (H/2) x W to
mark the position of the vertically embeddable pixel pairs (u, v)
in U. Then, VL is also losslessly compressed by using the
JBIG2 codec or an arithmetic coding toolkit to obtain the
compressed location map CM2 whose length is LC2. Next, the
map CM2 is concealed into the image V by using the LSB
steganography as mentioned above.
The final output of the embedding phase is the final stego
image X with size of H x W pixels.
III. The Proposed Steganalytic Scheme for the RVH
Steganography
After embedding a large message sequence M (its ratio is
about 90% of maximum embeddable capacity of image) into
the original image Baboon sized 512x512 pixels (show Fig. 2)
using the RVH scheme to obtain the stego-image Baboon, we
calculate histogram of the two images (cover Baboon image
and stego Baboon image), resulted in Fig. 3. It's very hard to
detect any difference between the two images.
However, when we separably calculate two histograms on
all pixel odd columns and all pixel even rows of the cover
Baboon image, shown in Fig. 4. Similarly, calculate two
histograms on all pixel odd columns and all pixel even rows of
the stego Baboon image, resulted in Fig. 5. It's easy to
difference between pair histogram in Fig.4 (a) and Fig. 5 (a), in
Fig. 4 (b) and Fig. 5 (b). The informality appears in Fig 5 (a)
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol 8, No. 6, September 2010
and (b) due to embedding process of RVH scheme following
description in detail below. 20
Fig. 2. The Baboon image sized 512x512 pixels
(a)
Fig. 3. Histogram of the tested two images: (a) the cover Baboon
(b) the stego Baboon image
(b)
image,
=>(a)
Fig. 4. Histogram of the cover Baboon images: (a) histogram on all
columns, (b) histogram on all pixel even columns
300 (b)
pixel odd
(a)
300 ( b )
Fig. 5. Histogram of the stego Baboon images: (a) histogram on all pixel odd
columns, (b) histogram on all pixel even columns
According to the horizontal embedding procedure HEm,
from an input image O, the pixels of the image O are
horizontally grouped into pixel pairs (x, y), these pairs are
partitioned into two sets, one set is El and other set is El, the
set El contains pixels pair which are horizontally embeddable
pixel pairs, while the set El consists of those pixel pairs which
are horizontally non-embeddable pixel pairs.
Now, we examine the migration of LSB histogram of the
image O and the image T obtained after embedding secret bit
Bl. Without loss of generality, let (x, y) and (x,y) be the
corresponding pixel pairs in the image O and the stego-image
T, respectively. In the horizontal embedding procedure HEm,
pixel pairs (x, y) E El, i.e. the LSB of pixel y be bit 1, are
selected to embed message bits. Here, We don't examine
change of LSB histogram of pixels x on pixel even-columns
because they are still remained value after embedding message
bits. In the image T, the LSB of y is changed to either or 1,
and each of them appears in the same probability. It is
obviously that the probability of bit and bit 1 are 0.5 and 0.5,
respectively. For pixel pairs (x, y) e El, i.e. the LSB of pixel y
be bit 0, after embedding secret bits, y is unchanged. So the
probability of bit and 1 are 1 and 0, respectively.
Next, the compressed location map CM1 (CM1 is a binary
stream, whose length is LCI) are marked into the image T by
the LSB replacement technique to obtained image U. That
changes a part of probability of LSB of bit 1 and bit on all
pixel even-columns in the image T. Assume that the bits are
randomly distributed, so the probability of bit and bit 1 are
Pmapl(O) = Pmapl(l)-
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
Base on above discussions, the probabilities of bit and bit
1 of all pixels on even-column in the image U can be
calculated. Assume the probability of pixel pairs belonging to
El and the probability of pixel pairs belonging to El be P E1
and Pet, respectively. After marking the location map CM1,
P E i and Pet are changed to P' E1 and P'et • Let P R . H is the
embedding ratio defined by dividing the number pairs actually
used to hide data by the total number of pairs the image O. The
probability of bit b={0,l} of LSB of the image U can be
calculated using the following equation
p rM _ [Pr-h x (0-5 x P' E1 + P'ei) + P^h x 0.5 if b =
n S B_HW - [ Prh x (q 5 x p , Ei) +P _ x0i5 ifb = l ( }
(IJCSIS) International Journal of Computer Science and Information Security,
Vol 8, No. 6, September 2010
used to embed message, i.e. the embedding ration of P R _
H =0.45 X 0.9=0.405. From (1), we have P L sb-h(0)
=0.405x(0.5x0.45+0.55) + 0.595x0.5 = 0.611375 and P LSB -
H (l)=0.405 X (0.5 X 0.45)+0.595 x 0.5 = 0.388625. Next,
calculated the probability of bit and the probability of bit 1 of
the output image X. We know that the probability of E2 equals
to the probability of the LSB of bit of all pixels on even -
rows, i. e. P E2 = P LSB (0)/2 + P LSB h(0)/2 = (0.5+0.61 1375)/2 =
0.5556875 and ?^2 =0.4443125. After covering a part of the
LSB of image V by the location map CM2 with a probability
0.05 (assumed), the two probability P E2 and ?^2 change to
P' E2 =0.5056875 and P'e2= 0.4943 125, respectively.
For the vertical embedding procedure VEm, vertically scan
the output image U in raster scan order to group pixel pairs (u,
v), we classify the pairs into two sets, one is E2 and other is
E2, the set E2 contains all pixel pairs which are vertically
embeddable pixel pairs, the set E2 consisting pixel pairs are
vertically non-embeddable pixel pairs. Let (u, v) and (u, v) be
pixel pairs of the image U (before using the procedure Vem)
and the stego-image V (after embedding secret message using
the procedure Vem). In the procedure VEm, only pixel pairs (u,
v) E E2, i.e. the LSB of v is bit 0, are embedded message bits.
After embedding message bits, the LSB of the pixels v
(obtained from E2) is either or 1. So, the probabilities of bit
and bit 1 of (u, v) are equals to 0.5. For pixel pairs (u, v) e E2,
i.e. the LSB of pixel v be bit 1, after embedding secret bits, v is
unchanged. So the probability of bit and 1 are and 1,
respectively.
Next, the compressed location map CM2 are marked into the
image V by the LSB replacement technique to obtained image
X. That changes a part of probability of LSB of bit 1 and bit
on all pixel even-rows in the image V. Assume the bits of the
map CM2 are randomly distributed, so the probability of bit
and bit 1 are P map2 (0) = P map 2(l)-
From above discussions, the probabilities of bit and bit 1
in the LSB of image X after using the vertical embedding
procedure VEm can be calculated. Assume the probability of
pixel pairs belonging to E2 and the probability of pixel pairs
belonging to E2 be P E2 and P^l, respectively. After marking
the location map CM2, P E2 and Pe2 are changed to P' E2 and
P'e2- Let P R _ V be the embedding ratio defined by dividing the
number pairs actually used to hide data by the total number of
pairs the image V. The probability of bit b={0,l} of LSB of
the image V can be calculated using the following equation
Ab) = \ P p R -
P R . V x (0.5 x P' E2 ) + P R - =V x 0.5
ifb =
x (0.5 x P' E2 + p' m ) + P-^y x 0.5 ifb = l
(2)
For a natural image, assume that the LSB is randomly
distributed, then the expected probability of bit and the
probability bit 1 of all pixel on pixel even-columns are the
same, i.e. P LS b(0) = P L sb(1)=0.5. So probability P E i=0.5,
Pet = 0.5. After covering a part of the LSB of image T by the
location map CM1 with a probability 0.05 (assumed), the two
probability P E1 and Pet change to P' E i=0.45 and P'et=0.55,
respectively. Consider the Baboon stego-image from the
Baboon cover image, the probability of the embeddable pairs
(i.e. those pixel pairs belonging to the procedure HEm) of an
input image T is P' E1 , and 90 % of the embeddable pairs are
The embedding ratio of 90 % of the embeddable pairs are
used to embed message, i.e. the embedding ration of P R _
v=0.5056875 x0.9=0.45511875. So probability of LSB of bit
and bit 1 of the output image X from (2) we obtain P LSB -
v(0)=0.4551 1875 X (0.5 X 0.5056875) + 0.54488125 X 0.5
^0.3875, P LSB _v(l)^0.61248.
Now, we check again the probability of LSB of bit 1 and the
probability of LSB of bit of all pixels on pixel even-columns,
PLSB even column(0)=P LS B H (0)/2 + P LSB V (0)/2 =(0.61 1375+
0.3875)/2=0.4994375,~ P LSB even coiumnO) = (Plsbh(I) +
Plsb_v(0))/2=(0.388625+0.61248)/2=0.5005525. We found out
that the probability of bit P L sv_even coiumn(O) and probability of
bit 1 P LSB even coiumn(l) are the same, that is after completing the
vertical procedure VEm, it make the value of these
probabilities be balanced. However, the probability of LSB of
bit and bit 1 of all pixels on pixel odd-columns don't equal
based on the following calculating:
PLSB_odd_column(0) = (PLSB_org_odd_column(0)/2+P LSB _v(0)/2)=(0 .5/2 +
0.3875/2)=0.44375, PLSB_odd_column(l)KPLSB_org_odd_column(iy2 +
Plsb_v(1)/2)=(0.5/2+0.612"48/2)=0.55624. Where
PLSB_or g _odd_coiumn(0) and P L sB_or g _odd_coiumn( 1 ) be the probabilities
of the LSB of bit and bit 1 of all pixels on the pixel odd-
column of the image X. A half of them isn't changed during
process of the RVH scheme, so P L sB_or g _odd_coiumn(0)/2 and
PLSB_or g _odd_coiumn(l)/2 equal to 0.5/2 and 0.5/2, respectively. We
can see obvious difference of the occurrences of bit 1 and bit
in the LSB on all pixel oddcolumns and all pixel
evencolumn of the stego-image of the RVH scheme with
respect to a standard natural image. Based on the problem, the
following rule is given to discriminate a stego-image of the
RVH steganography from a nature image.
W(X)
(true,if\P LSB (0)-
\ false, otherwise
Plsb(X)\>T
(3)
From equation (3), an image is detected be stego-image
marked by the RVH scheme if one of the measured values
Plsb(O) - Plsb(1)| on all pixel odd columns (or pixel even
columns) or pixel even rows (or pixel odd rows) is greater than
threshold T (0 < T < 1). The threshold T is used to control the
decision boundary of nature images and stego images, its value
depends on specific applications.
IV. EXPERIMENTAL RESULT
To show the reliability of the proposed method, we take
500 image from USC-SIPI Image Database [9] and content
based image retrieval (CBIR) image database [10] and convert
them into 8-bit grayscale images. The images are used to test
10
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol 8, No. 6, September 2010
proposed classification. The hidden messages used in our test
are made by the pseudo random number generator. We embed
different amount of message using the RVH scheme, and
measure the migration of LSB histogram in the stego images.
Five embedding ratios 0%, 25%, 50%, 75% and 100% are used
in the test, and the obtained |Plsb(0)-Plsb(1)| values on all pixel
even-rows of the stego-images are depicted in Fig. 6-10,
respectively. We also measure the accuracy of the proposed
method in detecting the RVH scheme in different embedding
ratio and likelihood threshold value T, shown in table 1.
0.2
§ 0.15
al
i
|Plsb(0) -
o
[(
^
h i
ft
v
f
^
0.05
f\
-
H
1
1
; ii
f
u
f
gb-.
r
:
50 100 150 200
250 300
350 400 450 500
Number of images
Fig
6. The distribution of |P L sb(0) - Plsb(1)| value of the 500 cover images on
all pixel even-rows
0.35 c
0.3
0.25
If 0.2
Q_
- T
«f °- 15
t * t t
1
K^
\\w
-^HhH
0.1
4
0.05
4^^
^ ^£#^
I A. ^jfa^J" ^ Jast^^
r i
v -.-■-■
i
-
C
50 100 150 200 250 300
350
400 450 5(
DO
Number of images
Fig.
7. The distribution of |P L sb(0) - Plsb(I)] value of the 500 stego images on
all pixel even-rows with embedding ratio 25%
S' 0.2 H
50 100 150 200 250 300 350 400 450 500
Number of images
Fig. 8. The distribution of |Plsb(0) - Plsb(1)| value of the 500 stego images on
all pixel even-rows with embedding ratio 50%
200 250 300
Number of images
350 400 450
Fig. 9. The distribution of |Plsb(0) - Plsb(1)| value of the 500 stego images on
all pixel even-rows with embedding ratio 75%
0.5
0.45
-
0.4
-
0.35
I 1
1 03
t
^ 0.25-
-Q
J2 0.2
Q-
I
0.15
-
0.1
-
0.05
-
150 200 250 300
Number of images
Fig. 10. The distribution of |P L sb(0) - Plsb(1)| value of the 500 stego images on
all pixel even-rows with embedding ratio 100%
TABLE I. The detection accuracy of the proposed method with
VARIOUS EMBEDDING RATIOS AND THRESHOLD VALUES
\. Embedding
^^atio (%)
ThresholdT\
25
50
75
100
0.01
Cover
70.6 %
2%
0.6%
0%
0%
Stego
29.4 %
98%
99.4 %
100%
100%
0.02
Cover
80.6 %
5.8 %
0.6 %
0%
Stego
19.4 %
94.2 %
99.4 %
100%
100%
0.03
Cover
84.4 %
7.4 %
2.2 %
Stego
15.6%
92.6 %
97.8 %
100%
100%
0.04
Cover
86.8 %
11%
2.4 %
Stego
13.2%
89%
97.6 %
100%
100%
0.05
Cover
89%
13.8 %
3.6%
0.2 %
Stego
11%
86.2%
96.4 %
99.8%
100%
From Fig. 6 , we can see that most of the value of |P L sb(0) -
Plsb(1)| approach zero for natural images, while the higher
value of |P L sb(0) - Plsb(1)| is obtained with embedding ratio
25%, 50%, 75% and 100% shown from Fig. 7 to Fig. 10. From
table 1, we see that when the likelihood threshold value T is set
11
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
0.035, we can obtain an acceptable result in detecting stego
images used the RVH steganography.
V. Conclusion
The paper presents a method to break the RVH
steganography based on the observation of the distribution of
and 1 bits of the LSBs on pixel odd-columns (pixel even-
columns) or pixel even-rows (pixel odd-rows) of the RVH
stego-images. The experimental results are shown that the
proposed method can detect stego - images reliably with
embedding ratio being greater 25%. On the other hand, we
show a problem of security of the RVH scheme in the data
embedding process.
Acknowledgment
Our special thanks to Haiphong Private University (HPU)
for their financial support to our research and College of
Technology, Vietnam National University, Hanoi for their
support to good working environment. We would like to
extend our thanks to my guide, our friends and family
members without whose inspiration and support our efforts
would not have come to success.
References
[1] P. Mohan Kumar, K. L. Shunmuganathan, A reversible high embedding
capacity data hiding technique for hiding secret data in images,
International Journal of Computer Science and Information Security,
Vol.7, No. 3, March 2010, pp. 109-115.
[2] Yeh-Shun Chen, Ran-Zan Wang, Yeuan-Kuen Lee, Shih-Yu Huang,
Steganalysis of reversible contrast mapping water marking, Proceedings
of the world congress on Engineering 2008 Vol I, WCE2008, July 2-4,
2008, London, U.K., pp. 555-557.
[3] D. Coltuc and J. M. Chassery, " Very fast watermarking by reversible
contrast mapping," IEEE Signal Processing Lett., vol. 14, no. 4, pp.
255- 258, Apr. 2007.
[4] J. Tian, " Reversible Data embedding using a difference expansion,"
IEEE Trans. Circuits Syst. Video technol., vol. 13, no. 8, pp. 890- 896,
Aug. 2003.
[5] Z. Ni, Y. Q. Shi, N. Ansari, and W. Su, " Reversible Data Hiding,"IEEE
Trans. Circuits Syst. Video technol., vol. 16, no. 3, pp. 354- 362, Mar.
2006.
[6] J. Fridrich, M. Goljan, and R. Du, " Reliable detection of LSB
steganography in color and grayscale images,"Proceedings of the ACM
International Multimedia Conference and Exhibition, pp. 27- 30, 2001.
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
[7] Carpenter, B., 2002. Compression via Arithmetic Coding
http://www.colloquial.com/ArithmeticCoding/
[8] P.G. Howard, F. Kossentini, B. Martins, S. Forchhammer, W. J.
Rucklidge, 1998. The emerging JBIG2 standard. IEEE Transactions on
Circuits and Systems for Video Technology 8 (7), pp. 838-848.
[9] USC-SIPI Image Database,
"http://sipi.usc.edu/services/database/Database.htm"
[10] CBIR Image Database, University of Washington,
http ://www. cs .Washington, edu/research/imagedatabase/groundtruth/. .
AUTHORS PROFILE
Ho Thi Huong Thorn received the B.S. degree of
Information Technology department from Haiphong
Private University and the M.S. degree in Information
Systems from College of Technology, Vietnam National
University in Vietnam, in 2001 and 2005, respectively.
She has started her career as Lecturer in Department of Information
Technology in Haiphong Private University, Vietnam and served for 9 years.
Currently, she is pursuing Doctor of Information Systems from College of
Technology, Vietnam National University, Hanoi, Vietnam. Her research
interests includes Image processing, Information Security, Information Hiding.
Ho Van Canh received the B.S. degree in Mathematics
from Hanoi City University in Vietnam in 1973, the Dr.
Sci. degree in Faculty of statistology from KOMENSKY
University in Czechoslovakia in 1987. Currently, he has
been working as a cryptologist in Dept. of Professional
Technique, Ministry of Public Security, Vietnam. His
research interests include cryptography, information security, information
hiding.
Trinh Nhat Tien received the B.S degree from University
of Prague in Czechoslovakia in 1974, and the Dr. degree
from University of Prague, Czechoslovakia and University
of Hanoi, Vietnam in 1984. He has started as Lecturer in
Department of Information Technology of College of
Technology, Vietnam National University, Hanoi, Vietnam since 1974. His
research interests include algorithm, complexity of algorithm, information
security.
12
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, 2010
Off-line Handwritten Signature Recognition
Using Wavelet Neural Network
Mayada Tarek
Computer Science Department,
Faculty of Computers and Information
Sciences,
Mansoura, Egypt
Taher Hamza
Computer Science Department,
Faculty of Computers and Information
Sciences,
Mansoura, Egypt
Elsayed Radwan
Computer Science Department,
Faculty of Computers and Information
Sciences,
Mansoura, Egypt
Abstract^Automatic signature verification is a well-
established and an active area for research with numerous
applications such as bank check verification, ATM access,
etc. Most off-Line signature verification systems depend
on pixels intensity in feature extraction process which is
sensitive to noise and any scale or rotation process on
signature image. This paper proposes an off-line
handwritten signature recognition system using Discrete
Wavelet Transform as feature extraction technique to
extract wavelet energy values from signature image
without any dependency of image pixels intensity. Since
Discrete Wavelet Transform suffers from down-sample
process, Wavelet Neural Network is used as a classifier to
solve this problem. A comparative study will be illustrated
between the proposed combination system and pervious
off-line handwritten signature recognition systems.
Conclusions will be appeared and future work is proposed.
Keywords -Discrete Wavelet Transform (DWT); Wavelet
Energy; Wavelet Neural Network (WNN); Off-line
Handwritten Signature.
I. INTRODUCTION
In the field of personal identification, two types of
biometrics means can be considered; first, physiological
biometrics, which involves data derived from the direct
measurement of some part of the human body; for-
example fingerprint-, face-, palm print-, retina-based
verification. Second, behavioural biometrics, which
involves data derived from an action taken by a person,
or indirectly measures characteristics of the human
body; for-example: speech-, keystroke dynamics and
signature-based verification [1].
In the last few decades, researchers have made great
efforts on off-line signature verification [1] for-
example; using the statistics of high grey-level pixels to
identify pseudo-dynamic characteristics of signatures;
developing technique based on global and grid features
1 Corresponding Author
Mail: mayaatarek @ yahoo . com
Tel : 020108631688
in conjunction with a simple Euclidean distance
classifier; proposing a system for off-line signature
verification consists of four subsystems based on
geometric features, moment representations, envelope
characteristics and wavelet features; applying wavelet
on signature verification [2,3,4,5].
Although these methods achieved a good results, they
still suffer from the exchangeability of signature
rotation and the distinguish-ability of person signature
size. Most of these feature extraction methods depend
on signature shape or pixels intensity in specific region
of signature. However, pixels' intensity are sensitive to
noise and also the signature shape may vary according
to translation, rotation and scale variations of signature
image [6].
Two types of feature can be extracted from signature
image; first, global features which are extracted from
the whole signature, including block codes [7]; second,
local features which are calculated to describe the
geometrical and topological characteristics of local
segments [8]. Because of the absence of dynamic
information in offline verification system, global
features extraction are most appropriate [9]. One of the
most appropriate global features extraction techniques is
wavelet transform, since it extracts time-frequency
wavelet coefficients from the signature image [8].
Wavelet Transform is especially suitable for processing
an off-line signature image where most details could be
hardly represented by functions, but could be matched
by the various versions of the mother wavelet with
various translations and dilations [10]. Also, wavelet
transform is invariant to translation, rotation and scale
of the image. Because of the advantage of wavelet
transform, this paper uses it in feature extraction stage.
Since one of problems that face wavelet is the huge size
of its coefficients, statistical model can be introduced to
represent them. This paper uses wavelet energy as
statistical model to represent all wavelet coefficients in
efficient way. Another problem is down-sample process
which can lose some important extracted feature from
signature image[ll]. This paper proposes a Wavelet
Neural Network (WNN) technique for off-line signature
recognition to overcome the disadvantages of Discrete
Wavelet Transform (DWT) down- sample process.
13
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, 2010
WNN takes full advantages of the partial-resolution
characteristic of the wavelet transform and the nonlinear
mapping behaviour of Artificial Neural Networks
(ANN) [15].
This paper proposes a combination model between
DWT and WNN techniques for off-line handwritten
signature recognition system. DWT technique will
analysis signature image to extract wavelet detail
coefficients. To reduce the huge number of these
coefficients with the same accuracy, a statistical model
is represented by wavelet energy. Because of the
problem of down sample, WNN technique will be used
as a suitable classifier technique to overcome this
problem. Also, a modified back-propagation technique
is used in learning WNN. A testing stage examines the
unseen signature. Moreover, a comparative study will
be illustrated between the proposed combination system
and pervious off-line handwritten signature recognition
systems. Conclusions will be appeared and future work
is suggested.
The rest of this paper organized as; in Section 2,
Handwritten signature, wavelet transform (WT),
Wavelet Neural Network (WNN) are mentioned.
Methodology and applications using a combination
between DWT and WNN techniques is described in
Section 3. Section4, consists of the result of the
proposed combination system and a comparative study
between three strategies (signature image pixels
intensity value as input to ANN , signature wavelet
energy values as input to ANN and signature wavelet
energy values as input to WNN). Finally section 5
concludes the paper.
II. PRELIMINARIES
A. Handwritten Signature
Handwritten signatures are widely accepted as a means
of document authentication, authorization and personal
verification. For legality most documents like bank
cheques, travel passports and academic certificates need
to have authorized handwritten signatures. In modern
society where fraud is rampant, there is the need for an
automatic Handwritten Signature Verification system
(HSV) [6]. Dependency on automation is due to the
difficulty faced in visual assessment for different types
and different sizes of signatures. Simple, cursive,
graphical and not a connected curve pattern are some of
the different types of signatures and machines are far
superior when it comes to processing speed and
management of large data sets with consistency [12].
Automatic HSV systems are classified into two types:
offline HSV and online HSV: static or off-line system
and dynamic or on-line system .Static off-line system
gain data after writing process has been completed .In
this case the signature is represented as a grey level
image. Dynamic systems use on-line acquisition devices
that generate electronic signals representative of the
signature during the writing process [1].
It is well known that no two genuine signatures of a
person are precisely the same and some signature
experts note that if two signatures written on paper were
same, then they could be considered as forgery by
tracing .Unfortunately, off-line signature verification is
a difficult discrimination problem because of dynamic
information regarding the signing velocity, pressure and
stroke order are not available also an off-line
handwritten signature is depend for instance on , the
angle at which people sign may be different due to
seating position or due to support taken by hand on the
writing surface and all this information can't be extract
from static image [12].
B. Wavelet Transform :
Wavelet Transform (WT) [13] is become a powerful
alternative analysis tool to Fourier methods in many
signal processing applications. The main advantages of
wavelets is that they have a varying window size, being
wide for slow frequencies and narrow for the fast ones,
thus leading to an optimal time -frequency resolution in
all the frequency ranges. Furthermore, owing to the fact
that windows are adapted to the transients of each scale,
wavelets lack the requirement of stationary. There are
two types of Wavelet Transform; Continous Wavelet
Transform(CWT), Discrete Wavelet Transform (DWT).
The Continuous Wavelet Transform [14] of a 1-D signal
x(t) is defined as in equation (1):
^(a,b)(t)= ^^(t) V{±) dt
(1)
Where t//(t) is the mother wavelet or the basis function
which, in a form analogous to sins and cosines in
Fourier analysis. All the wavelet functions used in the
transformation are derived from the mother wavelet
through translation (shifting) b and scaling (dilation or
compression) a.
The Discrete Wavelet Transform [14], which is based
on sub -band coding is found to yield a fast computation
of wavelet transform. It is easy to implement and
reduces the computation time and resources required.
In CWT, the signals are analyzed using a set of basis
functions which relate to each other by simple scaling
and translation. In the case of DWT, a time-scale
representation of the digital signal is obtained using
digital filtering techniques. The signal to be analyzed is
passed through filters with different cut off frequencies
at different scales [14].
In DWT, the extension to 2-D is usually performed by
using a product of 1-D filters. The transform is
computed by applying a filter bank as shown in
Figure 1. L and H to denote the 1-D low pass and high
14
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, 2010
pass filter, respectively. The rows and columns of image
are processed separately and down sampled by a factor
of 2 in each direction which may cause losing important
feature. Resulting in one low pass image LL and three
detail images HL, LH, and HH. Figure 2a shows the
one-level decomposition of Figure 7 in the spatial
domain. The LH channel contains image information of
low horizontal frequency and high vertical frequency,
the HL channel contains high horizontal frequency and
low vertical frequency, and the HH channel contains
high horizontal and high vertical frequencies. Three-
level frequency decomposition is shown in Figure 2b.
Note that in multi-scale wavelet decomposition only the
LL sub-band is successively decomposed [13].
rows
Image .
-©-
columns
-*©- HH
r+ H
rj->@= HL
L -©
H_}-^> LH
^^2> LL
Figure 1 : A one-level wavelet analysis filter bank.
LL
HL
LH
HH
(a) L-level Decomposition. (b) 3 -level Decomposition.
Figure 2 : Wavelet frequency decomposition.
C. Wavelet Neural Network :
neuron parameters .The output of WNN is therefore a
linear combination of several multidimensional
wavelets [15].
Figure 3 : The structure of the Wavelet Neural Network
In this WNN model, the hidden neurons have wavelet
activation functions y/ and have two parameter a t ,b t
which represent dilation and translation parameter of
wavelet function and V is the weight connecting the
input layer and hidden layer and U is the weight
connecting the hidden layer and output layer.
Let X n ={ x { },i=l, ,L and n=l, N be the WNN
input to no. n sample ; T ={ j k },k=l, ,S represents
the output of WNN ; D={ d k },k=l, ,S represents the
expected output ; Vy represents the connection weight
between no. i node (input layer) and . j node (hidden
layer) ; U jk represents the connection weight between
no. j node (hidden layer) and k node (output layer) .
Where N is the number of Sample ; S is the number of
output node ; L is the number of input node ; M is the
number of hidden layer.
III. WAVELET NEURAL NETWORK FOR
OFF-LINE HANDWRITTEN SIGNATURE
RECOGNITION
WNN is a combination technique between neural
network and wavelet decomposition .The advantages of
the WNN are a high-speed learning and a good
convergence to the global minimum [15]. The reason for
the application of WNN in case of such a problem as
classification is that the feature extraction and
representation properties of the wavelet transform are
merged into the structure of the ANN to further extend
the ability to approximate complicated patterns [16].
The WNN can be considered an expanded perceptron
[17]. The WNN is designed as a three-layer structure
with an input layer, a wavelet layer, and an output layer.
The topological structure of the WNN is illustrated in
Figure 3.
In WNN, both the position and dilation of the wavelets
as well as the weights are optimized. The basic neuron
of a WNN is a multidimensional wavelet in which the
dilation and translation coefficients are considered as
According to the fact that there aren't two genuine
signatures of one person are precisely the same, many
efforts have been done in order to comprehend the
delicate nuances of person signatures [12]. Especially
off-line signature recognition needs more effort because
of the absence of dynamic information that can't be
extracted from static image [12]. Also, the problems of
translation, rotation and scale variation of signature
image are still found when dealing with signature image
pixels' intensity [6].
This paper presents an implementation for off-line
handwritten signature recognition system using DWT
technique in feature extraction phase and WNN in
classification phase to overcome all the above problems
with off-line handwritten signature recognition system.
DWT technique depends on analyzing all signature
shapes (continuous case) instead of analyzing the pixels
intensity or segmentation part of signature (discrete
case). Because of the problem of down-sample caused
15
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, 2010
by DWT technique, WNN technique will be used in
classification stage to overcome this problem.
The proposed Off-line Handwritten Signature
recognition system as depicted in Figure 4 involves
four stages:
♦ Scan and removing noise stage.
♦ Feature extraction stage.
♦ Classification stage.
♦ Test stage.
First stage, Scan and removing noise stage, each off-
line handwritten signature is scanned due to creating
signature image. Because of the scanning process,
removing noise from signature image is an important
task. In this paper, the median filter [18] is used to
remove noise for two reasons. First, it preserves the
structural shape of the signature without removing small
strokes. Second, the absence of dealing with median
filter in wavelet transform technique, which work to
analysis image with low/high-pass filters corresponding
to its wavelet function.
The median filter is a nonlinear digital filtering
technique which is often used to remove noise. Noise
reduction is a typical pre-processing step that improves
the results. The median filter considers each pixel in the
image in turn and looks at its nearby neighbours to
decide whether or not the pixel intensity value is
representative of its surroundings. The median filter
replaces the pixel with the median of its neighbouring
pixel intensity values. The median is calculated by first
sorting all the pixel intensity values from the
surrounding neighbourhood into numerical order and
then replacing the pixel being considered with the
middle pixel intensity value [19].
Second stage, Feature extraction stage is the most
important component for designing the intelligent
system based on pattern recognition. The pattern space
is usually of high dimensionality. The objective of the
feature extraction is to characterize the object by
reducing the dimensionality of the measurement space
(i.e., the original waveform). The best classifier will
perform poorly if the features are not chosen well [20].
According to the fact that there aren't two genuine
signatures of one person are precisely the same, the
differences in the same person signature may exist in
details. Because of the details of an image will access
by high pass filter, DWT is used to access high pass
information of person's signature images. This
information is fused to obtain pattern of each person's
signatures that contains all details information of his/her
signatures [21]. Details information extracted by DWT
technique must be extracted using suitable wavelet
function to off-line handwritten signature recognition
application. According to the previous work in off-line
handwritten signature recognition have apply
Daubechies 4, 12 and 20 wavelets functions as depicted
in Figure 5 [5] as a mother wavelet function, which can
preserve maximum details of the original image, reflect
outline of the image objectively and decrease the FRR.
Figure 5: Daubechies 4, 12 and 20 wavelets functions
After DWT is applied on the image, wavelet
coefficients from the approximation sub-band is discard
and interested in wavelet coefficients from the details
sub-bands of all the decomposition levels . This entire
coefficient is very large to be used as feature extraction
model from an image. These wavelet coefficients can be
represented as statistical features such as mean, median,
standard deviation, energy and entropy [22]. In this
paper, wavelet energy values for details wavelet sub-
band is the reduced vector that contain the main
information that represent person signature from the
huge wavelet decomposition values.
While off-line handwritten signature image is sensitive
to translation, rotation and scale changes; the same
images with different scale or rotational may have
different wavelet coefficients. The main reason is that
the efficient implementation of 2D -DWT requires
applying a filter bank along the rows and columns of an
image [23].
}
Feature
extraction using
Wavelet
Transform
Wavelet energy
coefficient
Neuron of 1 person
Neuron of 2 person
Neuron of 9 person
Figure 4: Proposed off-line Handwritten signature Recognition System
16
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, 2010
Due to the separability of the filters, the separable 2D-
DWT is strongly oriented in the horizontal and vertical
directions. This makes it hardly possible to extract
translation, rotation and scale invariant features from
the wavelet coefficients. Wavelet energy can keep the
main characteristic of these wavelet coefficients and
make the same images with different translation,
rotation and scale having the same wavelet energy
values[23]. Wavelet energy values can be computed
after analysis signature image to it's wavelet sub-image
coefficient at three level analysis (LLx, HLx, LHx,
HHx). The percentages of energy of these high
frequency sub-images at the &-level wavelet
decomposition is defined in equation (2,3,4)[24]:
The back-propagation algorithm seems to be superior in
this handwritten signature verification environment
[25]. In a back-propagation neural network[26], the
learning algorithm has two phases. First, a training input
pattern is presented to WNN input layer. The WNN
propagates the input pattern from layer to layer until the
output pattern is generated by the output layer. If this
pattern is different from the desired output, an error is
calculated and then propagated backward through the
WNN from the output layer to the input layer. The
weights and both the position and dilation of the
wavelets layer are modified as the error is propagated.
The modified back-propagation training algorithm in
WNN [27] as shown in Figure 6.
ehlw :
100 * Y*(HL decomposition vector at level K) 2
^(decomposition vector) 2
(2)
100 * 2(L// decomposition vector at level K) 2
ELH<& = — — t^ — (3)
EHH
^(decomposition vector) 2
100 * £(//// decompositionvector at level K) 2
^(decomposition vector) 2
(4)
Third stage, Classification stage, after we get the
suitable wavelet energy values that represent signature
image, we take this values as input to WNN and train
this network with a modified Back-propagation (BP)
training algorithm to get efficient off-line signature
recognition. Using WNN for two reasons; first,
traditional ANN has many trade-off because of complex
computations, huge iterations and learning algorithms
are responsible for slowing down the recognition rate
using ANN; second ,recover losing important
information from signature image in DWT technique
because of down-sample process as depicted in
Figure 1 .
In this work, The input layer represents wavelet energy
values feature vector to neural network. The output
layer represents the ability to recognize the human
signature. The middle layer determined the ability to
learn the person signature recognition. Because of the
ability of Morlet function to deal with big input domain
[28] and represents its wave form in equation, Morlet
function will be the suitable wavelet activation
functions y/ in WNN to recognize offline handwritten
signature application. Morlet function equation and it's
derivation in equation (5,6) [27]:
ip(x) = cos(1.75x) exp I
(5)
Then.
dib(x) ( x 2 ,
-j^- = ~[xcos(1.7Sx) + 1.75 sin(1.75x)] exp ( - — ) (6)
17
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, 2010
Input : wavelet energy values extracted from signature image
Output : Class of recognized Signature
> Stepl: Initialize weights and offsets.
Set all weights and node offsets to small random values.
Initialize position and dilation parameter for each wavelet neuron in wavelet layer.
To choose centre point (p)between interval [z!,z 2 ] (input domain),then
t>!=p , ai=0.5(z r z 2 )
Interval [z { ,z 2 ] is divided into two parts by point p
In each sub-interval, we recursively repeat the same procedure which will
initialize b 2 , a 2 and b 3 , a 3 and so on, until all the wavelet are initialize.
> Step2: Present input and desired outputs
Present a continuous valued input vector Xj , X 2 X L
and specify the desired output Dj ,D 2 , . . . .D s .
If the net is used as a classifier then all desired outputs are typically set to zero except for that corresponding
to the class the input is from. That desired output isl. The input could be new on each trial or samples from a
training set could be presented cyclically until stabilize.
> Step 3: Calculate Actual Output
't*=YjJv*
7=1
T^VjjXr-bj
> Step 4: Calculate Error function
n=lk=l
> Step 5: Propagate error to weights and position and dilation parameter
\si =1 Vtjxr-b,
dE lv
J n=1
Where
T = YUv M xf z b l
N S
dE lvv drb(T)
J n=lk=l
Yi^VyXT-b,
af
dE
db,
(7)
e = 1nYL^- d ^ 2 (8)
(9)
l J n =l k=\ J
(11)
r«H w - D?) ^V(-^-J (i2)
J n=lk=l x J/
> Step 6: Update weights and position and dilation parameter
dE
where :a is learning rate
s is momentum factor
> Step 7: Repeat by going to step 2
Figure 6. B ack-propagation training algorithm in W avelet Neural Network
1 8 http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, 2010
Feature
extraction using
Wavelet
Transform
^>
Wavelet energy
coefficient
Second Person Signature
Neuron of 1 person
1 Neuron of 2 person
Neuronof9 person
Figure 7: Testing stage of off-line Handwritten signature Recognition System
Finally, Test stage, after learning WNN, we can
examine the ability of WNN to verify the signature of
any person as shown in Figure 7. In this stage our goal
is to input signature image and recognize the person
signature. After scanning and removing noise from
person signature image, wavelet coefficients produce
after analysis image with DWT technique and then
compute wavelet energy value from wavelet detail
coefficients, finally, this wavelet energy values are
taken as input to test WNN classifier to find result in
output layer with only 1 value in only one neuron.
Number of neurons in the output layer represents the
number of person that system recognize.
IV. RESLUT :
This section summarizes the results of using DWT
technique (wavelet energy values) as feature extraction
technique and WNN as classifier to off-line handwritten
signature recognition system. This paper uses nine
person handwritten signatures as show in Figure 8,
each person has twenty image of his handwritten
signature ,ten for train stage and ten for test stage .
GspbrYWn
4) Liu
HUM**
Figure 8: Sample Signature images
In feature extraction stage, wavelet detail coefficients
are extracted from signature image using (db4 or db 12
or db20) wavelet function. To determine the suitable
wavelet function to our database, WNN is used as a
classifier to evaluate the suitable one. Wavelet detail
coefficients (at one level analysis) of signature image
according to one wavelet function is taken as trained
data to WNN.
Three WNN will be found to compare the recognition
rate between tree wavelet function. Modified BP
training algorithm as in Figure 6 is used to train WNN.
Finally, testing WNN with trained signature . Figure 9
shows the recognition rate to (db4,dbl2,db20) wavelet
detail coefficients using WNN as mention above. As a
result from Figure 9, Db20 is recognizing to be the
suitable wavelet function which have high recognition
rate in our database to offline handwritten application.
Figure 9: Offline handwritten signature recognition rate using
(db4,dt>12,db20) wavelet detail coefficients
After determine the suitable extracting wavelet
function, wavelet energy from each signature image is
computed using equation 2,3,4 with db20 as wavelet
function at three level analysis. Nine wavelet energy
coefficients are represented each signature image.
Table 1: WNN architecture and training parameters
The number of layers
3
The number of neuron on the
layers
Input:9
Hidden: 18
Output: 9
The initial weights and biases
Random
Wavelet Activation functions
Morlet function
Learning rule
Back-Propagation
MSE
0.0001
Learning rate
0.1
Momentum factor
0.009
19
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, 2010
In classification stage, WNN is used with parameters
shown in Tablel. These parameters are selected for
WNN structure after several different experiments. In
these experiments, the WNN is employed with different
parameters such as the number of hidden layers, the size
of the hidden layers, value of the moment constant and
learning rate, and type of the activation functions.
Wavelet energy values for each signature image are the
input features to WNN input layer. Each neuron in
WNN output layer represent a person.
In the test stage , Appling the test wavelet energy
values of the test signature to trained WNN. Evaluating
the proposed off-line handwritten signature recognition
system by recognition rate to each person as shown in
Figure 10.
Figure 10: Proposed off-line handwritten signature recognition
system result
All system evaluation is made by two concept False
Acceptance Rate(FAR) which indicates how many
forgeries were incorrectly classified as genuine
signatures ,and False Rejected Rate(FRR) which
indicates how many genuine signatures were incorrectly
rejected by the system. To the training signatures FAR
and FRR is 0.01% and to the testing signatures FAR
and FRR is 0.07% .
To evaluate our proposed system a comparative study
between three off-line handwritten signature systems is
made:
1 -signature image pixels intensity value as input to
ANN(ANN)
2- signature wavelet energy values as input to ANN
(WE+ANN)
3- Our proposed system signature wavelet energy
values as input to WNN (WE+WNN).
Figure 11 represent the recognition rate to each
training person data and Figure 12 represent the
recognition rate to each testing person data. Figure 11
and Figure 12 concluded that our proposed system has
the highest recognition rate.
lUZ/o
inn% -
. QR% -
"■--.llxf
n
>&</
r— ^
\y7
1
j 96%-
I 94% -
8
1 Q7°/n -
\
\
/
\
/
--■ANN
l
i
-■-WE+ANN
90% -
.i. U/F+U/MM
Wt+WNIN
SR%
PI P2 P3 P4 P5 P6 P7 P8 P9
Person
Figure 11: Comparative Study between signature image pixels
intensity value as input to ANN (ANN)and signature wavelet energy
values as input to ANN (WE+ANN)and signature wavelet energy
values as input to WNN (WE+WNN)with training data.
102%
100%
« 98%
j 9696
I 94%
S 92%
Sv S*'?\\. <
\ ,-■""''' /
(/
Y
if
Y\
s
(
Y
/v
-■-WE+ANN
PI P2 P3 P4 P5 P6 P7 P8 P9
Figure 12: Comparative Study between signature image pixels
intensity value as input to ANN (ANN)and signature wavelet energy
values as input to ANN (WE + ANN)and signature wavelet energy
values as input to WNN (WE + WNN) testing data.
V. CONCLUSIONS AND FUTURE WORK
Handwritten signature recognition plays an important
role in our daily life especially in any bank and any
ATM system. Off -Line Handwritten Signature
recognition is a difficult task than On-line one because
of absence of dynamic information in off -Line signature
image such as angle of written style of written and so
on. This paper proposed an off-Line handwritten
recognition system with Four stages. First stage is
scanning signature image and removing noise using
median filter. Second stage, extract feature from each
signature image using DWT technique with the
advantage of multi-scale and with respect the
translation, rotation and scale variations of signature
image. Computing wavelet energy values from DWT
details sub-bands coefficient to all person signature
images using the suitable wavelet function to our
database. Daubechies 20 (db20) is recognize as a
suitable wavelet function with three levels analysis after
a comparative study with other wavelet function. Third
stage, taking the wavelet energy values as input to
WNN with Morlet function as activation function in
hidden layer. Finally, testing trained WNN with
seen/unseen signature to evaluate our proposed system
8
20
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, 2010
recognition rate. A comparative study between three
off-line handwritten signature systems is made
( signature image pixels intensity as input to ANN ,
signature wavelet energy values as input to ANN and
signature wavelet energy values as input to WNN). The
conclusion will found that our proposed system
(wavelet energy values as input to WNN) has high
recognition rate.
[12] K R Radhika, M K Venkatesha and G N Sekhar," Pattern
Recognition Techniques in Off-line hand written signature
verification - A Survey", PROCEEDINGS OF WORLD
ACADEMY OF SCIENCE, ENGINEERING AND
TECHNOLOGY ,vol. 36, ISSN 2070-3740 , 2008.
[13] Engin Avci , Abdulkadir Sengur, Davut Hanbay, " An optimum
feature extraction method for texture classification", Expert
Systems with Applications: An International Journal, Published
by Elsevier Ltd, Volume 36 , Issue 3,2009,p.p 6036-6043.
To improve our system recognition rate, each person
signature should have its own wavelet function in
feature extraction stage. Genetic algorithm will be used
as a searching strategy in the future work to found the
suitable wavelet function to each person signature.
ACKNOWLEDGEMENTS :
The authors would like to thank .Prof. Albert Swart for
making his signature database available to us.
The first author would like to thank Sarah El.metwally
and Eslam Foud for their encouragements.
REFERENCES:
[1] S. Impedovo, G. Pirlo," Verification of Handwritten Signatures: an
Overview", 14th International Conference on Image Analysis and
Processing , 2007, pp. 191-196 .
[2] M. Ammar, Y. Yoshida, T. Fulumura, " A New Effective
Approach for Off-line Verification of Signatures by Using
Pressure Features",. Proceedings of the 8th International
Conference on Pattern Recognition, 1986, p.p 566-569.
[3] Y. Qi , B.R. Hunt, "Signature Verification Using Global and Grid
Features," Pattern Recognition, vol.27, Issue. 12, 1994, p.p 1621-
1629.
[4] V.E. Ramesh, M.N. Murty, "Off-line Signature Verification Using
Genetically Optimized Weighted Features", Pattern Recognition,
vol.32, Issue.2, 1999, p.p 217-233.
[5] P.S. Deng, H.-Y. M. Liao, C.W. Ho, and H.-R. Tyan,: "Wavelet-
Based Off-line Handwritten Signature Verification", Computer
Vision and Image Understanding, vol.76, no. 3, 1999, p.p 173 -
190.
r61 http://dspace.mak.ac.ug/bitstream/123456789/599/3/karanja-
evanson-mwangi-cit-masters-report.pdf , 17-8-2010
[7] M. Kalera, S. Srihari, and A. Xu. "Offline signature verification
and identification using distance statistics", International Journal
of Pattern Recognition and Artificial Intelligence, vol. 18, no.7,
2004, pp. 1339-1360.
[8] V. Nalwa." Automatic on-line signature verification", Lecture
Notes In Computer Science, Proceedings of the Third Asian
Conference on Computer Vision,1998, p.p 10-15 .
r91 http://research.microsoft.com/pubs/69437/handwritingregistration
cvpr07.pdf , 17-8-2010. ~~ "
[10]Sing-Tze Bow, "Pattern recognition and
preprocessing" ,Marcel Dekker,Inc, chapter 15,2002.
image
[14] http://www.dtic.upf.edu/~xserra/cursos/TDP/referencies/Park-
DWT.pdf , 17-8-2010.
[15]S.Sitharama Lyengar,E.C.Cho,Vir V.Phoh /'Foundations of
Wavelet Networks and Application",
chapman&Hall/CRC Press LLC , chapter 4, 2002.
[16] Xian-Bin Wen, Hua Zhang, and Fa-Yu Wang," A Wavelet
Neural Network for SAR Image Segmentation", Sensors ,
Vol.9,No.9,2009,p.p 7509-7515 .
[17] Zhang Q. and Benveniste A,"Wavelet networks"
On Neural Networks ,Vol.3, ,1992,p.p 889-898.
IEEE Trans.
[11] G.Y. Chen, T.D. Bui, A. Krzyzak," Contour-based handwritten
numeral recognition using multi-wavelets and neural networks",
Pattern Recognition , Vol.36 ,2003,p.p 1597 - 1604.
[18] ri Gross and Longin Jan Latecki,"Digital geometric methods in
document image analysis",Pattern Recognition, Vol. 32,No.3,
1999,pp. 407.
[19] http://homepages.inf.ed.ac.uk/rbf/HIPR2/median.htm , 24-6-2010.
[20] Avci, E., Turkoglu, I., & Poyraz, M., "Intelligent target
recognition based on wavelet packet neural network", Experts
Systems with Applications: An International Journal, vol 29, Issue
1 ,2005 ,p.p 175-182.
[21] Samaneh Ghandali, Mohsen Ebrahimi Moghaddam, "Off-Line
Persian Signature Identification and Verification Based on Image
Registration and Fusion", JOURNAL OF MULTIMEDIA, VOL.
4, NO. 3, 2009.
[22] A. Wahi, E. Thirumurugan "Recognition of Objects by
Supervised Neural Network using Wavelet Features", First
International Conference on Emerging Trends in Engineering and
Technology,2008.
[23] Chi-Man Pun and Moon-Chuen Lee ,"Log-Polar Wavelet Energy
Signatures for Rotation and Scale Invariant Texture
Classification", IEEE TRANSACTIONS ON PATTERN
ANALYSIS AND MACHINE INTELLIGENCE, VOL. 25, NO.
5, 2003
[241 http://www.mathworks.com/access/helpdesk/help/toolbox/wavele
tAyenergy2.html , 17-8-2010.
[25] Alan McCabe, Jarrod Trevathan and Wayne Read," Neural
Network-based Handwritten Signature Verification", JOURNAL
OF COMPUTERS, VOL. 3, NO. 8, 2008.
[26] Insung Jung, and Gi-Nam Wang," Pattern Classification of Back-
Propagation Algorithm Using Exclusive Connecting Network",
World Academy of Science, Engineering and Technology, VOL.
36,2007.
[27] Ming Meng , Wei Sun," Short-term Load Forecasting Based on
Rough Set and Wavelet Neural Network ", International
Conference on Computational Intelligence and Security,2008.
[28] Mohd Fazril, Zaki Ahmad, Hj. Kamaruzaman, "The Performance
of Two Mothers Wavelets in Function Approximation", Joural of
Mathematical Research, Vol. 1 ,No.2,2009.
21
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
A Black-Box Test Case Generation Method
Nicha Kosindrdecha
Autonomous System Research Laboratory
Faculty of Science and Technology, Assumption University
Bangkok, Thailand
Jirapun Daengdej
Autonomous System Research Laboratory
Faculty of Science and Technology, Assumption University
Bangkok, Thailand
Abstract — Test case generation techniques have been researched
over a long period of time. Unfortunately, while many
researchers have found methods of minimizing test cases, there
are still a number of important related issues that need to be
researched. The primarily outstanding research issue is a large
single test suite containing a huge number of test cases. Our study
shows that this can lead to other two problems: unable to identify
suitable test cases for execution and those test cases are lack of
ability to cover domain specific requirement. Therefore, we
proposed an additional requirement prioritization process during
a test case generation process and an automated method to
generate multiple test suites while minimizing a number of test
cases from UML Use Case diagram 2.0. Our evaluation result
shows that the proposed method is the most recommendation
method to minimize size of test cases while maximizing ability to
cover critical domain specific requirements.
Keywords-component; Test generation, testing and quality, test
case generation, test generation technique and generate tests
I. Introduction
Software testing is known as a key critical phase in the
software development life cycle, which account for a large part
of the development effort. A way of reducing testing effort,
while ensuring its effectiveness, is to generate a minimize
number of test cases automatically from artifacts used in the
early phases of software development. Many test case
generation techniques have been proposed [2], [4], [10], [11],
[12], [15], [21], [22], [42], [47], [50], mainly random, path-
oriented, goal-oriented and model-based approaches. Random
techniques determine a set of test cases based on assumptions
concerning fault distribution. Path-oriented techniques
generally use control flow graph to identify paths to be covered
and generate the appropriate test cases for those paths. Goal-
oriented techniques identify test cases covering a selected goal
such as a statement or branch, irrespective of the path taken.
There are many researchers and practitioners who have been
working in generating a set of test cases based on the
specifications. Modeling languages are used to get the
specification and generate test cases. Since Unified Modeling
Language (UML) 2.0 is the most widely used language, many
researchers are using UML diagrams such as UML Use Case
diagram, UML Activity diagram and UML Statechart diagram
to generate test cases and this has led to model-based test case
generation techniques. The study shows that model-based test
generation methods (or also known as black-box test
generation) are widely-used for generating test cases in the
commercial industry.
Moreover, the study [2], [4], [10], [11], [12], [15], [21],
[22] shows that the primary research issue is that existing
black-box test case generation methods generate a huge single
test suite with a number of possible tests. The number of
possible black-box tests for any non-trivial software application
is extremely large. Consequently, it is unable to identify
suitable test cases for execution.
Also, the study shows that the secondary research issue is
that the existing black-box test case generation methods ignore
critical domain specific requirements [5] during a test case
generation process. These requirements are one of the most
important requirements that should be addressed during test
activities.
Therefore, we propose a new black-box test case
generation, with requirement prioritization approach, from
requirements captured as use cases, 2.0, [23], [24], [33]. A use
case is the specification of interconnected sequences of actions
that a system can perform, interacting with actors of the
system. Use cases have become one of the favorite approaches
for requirements capture. Our automated black-box approach
aims to generate a minimize number of suitable test cases while
reserving critical domain specific requirements. Additionally,
we introduce an automated test generation method derived
from UML Use Case diagram, 2.0. Our approach is developed
to automatically generate many test suites based on notions
announced in the latest version of UML.
The rest of the paper is organized as follow. Section 2
discusses an overview of test case generation techniques.
Section 3 describes motivated research issues. Section 4
introduces a new test generation process with requirement
prioritization step. Also, section 4 proposes a new black-box
test generation method. Section 5 describes an experiment,
measurement metrics and results. Section 6 provides the
conclusion and research directions in the test case generation
field. The last section represents all source references used in
this paper.
II. LITERATURE REVIEW
The literature review is structured into two sections. The
first section gives an overview of previous studies. The second
section provides the related works
22
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
A. An Overview of Recent Researches
Model-based techniques are popular and most researchers
have proposed several techniques. One of the reasons why
those model-based techniques are popular is that wrong
interpretations of complex software from non-formal
specification can result in incorrect implementations leading to
testing them for conformance to its specification standard [43].
A major advantage of model-based V&V is that it can be easily
automated, saving time and resources. Other advantages are
shifting the testing activities to an earlier part of the software
development process and generating test cases that are
independent of any particular implementation of the design [7].
The model-based techniques are method to generate test
cases from model diagrams like UML Use Case diagram [23],
[24], [33], UML Sequence diagram [7] and UML State diagram
[5], [43], [22], [2], [21], [15], [32], [4]. There are many
researchers who investigated in generating test cases from
those diagrams. The following paragraphs show examples of
model-based test generation techniques that have been
proposed for a long time.
Heumann [23] presented how using use cases, derived from
UML Use Case diagram 1.0, to generate test cases can help
launch the testing process early in the development lifecycle
and also help with testing methodology. In a software
development project, use cases define system software
requirements. Use case development begins early on, so real
use cases for key product functionality are available in early
iterations. According to the Rational Unified Process (RUP), a
use case is used to describe fully a sequence of actions
performed by a system to provide an observable result of value
to a person or another system using the product under
development." Use cases tell the customer what to expect, the
developer what to code, the technical writer what to document,
and the tester what to test. He proposed three-step process to
generate test cases from a fully detailed use case: (a) for each
use case, generate a full set of use-case scenarios (b) for each
scenario, identify at least one test case and the conditions that
will make it execute and (c) for each test case, identify the data
values with which to test.
Ryser [24] raised the practical problems in software testing
as follows: (1) Lack in planning/time and cost pressure, (2)
Lacking test documentation, (3) Lacking tool support, (4)
Formal language/specific testing languages required, (5)
Lacking measures, measurements and data to quantify testing
and evaluate test quality and (6) Insufficient test quality. They
proposed their approach to resolve the above problems. Their
approach is to derive test case from scenario / UML Use Case
diagram 1.0 and state diagram 1.0. In his work, the generation
of test cases is done in three processes: (a) preliminary test case
definition and test preparation during scenario creation (b) test
case generation from Statechart and from dependency charts
and (c) test set refinement by application dependent strategies.
Whereas all requirements are mandatory, some are more
critical than others. For example, failure to implement certain
requirements may have grave business ramifications that would
make the system a failure, while others although contractually
binding would have far less serious business consequences if
they were not implemented or not implemented correctly (b)
Help programs through negotiation and consensus building to
eliminate unnecessary potential "requirements" (i.e., goals,
desires, and "nice-to-haves" that do not merit the mandatory
nature of true requirements) and (c) schedule the
implementation of requirements (i.e., help determine what
capabilities are implemented in what increment). Additionally,
these researches in 1980-2008 [8], [27], [28], [29], [30], [38]
reveal that there are many requirement prioritization methods
such as Binary Search Tree (BST), 100-point method and
Analytic Hierarchy Process (AHP)
III. RESEARCH PROBLEM
This section discusses the details of research issues related
to test case generation techniques and research problems,
which are motivated this study. Every test case generation
technique has weak and strong points, as addressed in the
literature survey. In general, referring to the literature review,
the following lists major outstanding research challenges.
The first research problem is that existing test case
generation methods are lack of ability to identify domain
specific requirements. The study [5] shows that domain
specific requirements are some of the most critical
requirements required to be captured for implementation and
testing, such as constraints requirements and database specific
requirements. Existing approaches ignore an ability to address
domain specific requirements. Consequently, software testing
engineers may ignore the critical functionality related to the
critical domain specific requirements. Thus, this paper
introduces an approach to priority those specific requirements
and generates an effective test case.
The second problem is that existing black-box test case
generation techniques aim to generate a large single test suite
with all possible test cases which maximize cover for each
scenario. Basically, they generate a huge number of test cases
which are impossible to execute given limited time and
resources. As a result, those unexecuted test cases are useless
and it is unable to identify suitable test cases for execution.
IV. PROPOSED METHOD
A. Test Case Generation Process
This section presents a new high-level process to generate a
set of test cases introduced by using the above comprehensive
literature review and previous works [43].
B. Related Works
This section provides the related works used in this paper,
prioritize requirement methods. Donald Firesmith [16]
addressed the purpose of requirement prioritization as follows:
(a) Determine the relative necessity of the requirements.
Autonomous System Research Laboratory, Faculty of Science and
Technology, Assumption University.
23
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
Existing Test Case Generation Process Proposed Test Case Generation Process
1.
Requirement
1.
Requirement
*
1.1
Requirement
Prioritization
< j
Design
Test Case V
Generation f
Design
*
2.1
Test Case
Generation
▼
3 - I
Development
3-
Developme^B
-J
4.
Testing
f
J
4.
Testing
v
5.
Maintenance
5.
Maintenance
Figure 1 . A Proposed Process to Generate Test Cases
From the above figure, there are two test case generation
processes: existing and proposed process. The left-hand side
shows an existing process to generate test cases directly from
diagrams. Meanwhile, the right-hand side proposes to add an
additional requirement prioritization process before generating
test cases. The requirement prioritization process aims to be
able to effectively handle with a large number of requirements.
The objective of this process is to prioritize and organize
requirements in an appropriate way in order to effectively
design and prepare test cases [16], [25], [37]. There are two
sub-processes: (a) classify requirements and (b) prioritize
requirements.
Our study [51], [52], [53], [54] shows that a marketing
perspective concentrates on two factors: customer's needs and
customer satisfaction. We apply that perspective to the
requirement prioritization and propose the following:
Customer Satisfaction
Performance
Delight
(Nice to Have)
From the above figure, the horizontal axis presents a
customer's need while the vertical axis represents a customer
satisfaction. There are four groups of requirements based on
those two factors: delight, attractive, indifferent and basic.
First, the delight requirement is known as 'nice-to-have'
requirement. If this requirement is well fulfilled, it will increase
the customer satisfaction. Otherwise, it will not decrease the
satisfaction. Second, the attractive requirement is called as
'surprise' or 'know your customer' requirement. This
requirement can directly increase the customer satisfaction if it
is fulfilled. Marketers and sales [53] believe that if we can
deliver this kind of requirement, it will impress customers and
significantly improve the customer satisfaction. Third, the
indifferent requirement is a requirement that customer does not
concentrate and it will not impress customers at all. In the
competitive industry, this requirement may be fulfilled, but
there are no any impacts to the customer satisfaction. Last, the
basic requirement is a mandatory requirement that customers
basically expect. Therefore, if this requirement is well
delivered, it will not increase the customer satisfaction.
Furthermore, our study reveals that the requirement can be
simply divided into two types: functional and non-functional
requirement. Our study also presents that functional
requirements can be categorized into two groups: domain
specific requirement [5] (or known as constraints requirement)
and behavior requirement. The following shows the
requirement classification used in this paper:
Constraints/
Business
>
Domain Specific/
/ 1
Functional
Requirement
Rules >^
System \
Behavior
Constraints Req.
Behavioral Req.
Non-Functional
Requirement
{ )
Dissatisfk \
Figure 2. Classify Requirement on Marketing's Perspective
Figure 3. Classify Requirement on Software Engineer
From the above figure, functional requirement is a
requirement that customers directly are able to provide. The
non-functional requirement is a requirement that is given
indirectly. The domain specific or constraints requirement is a
requirement relative to any constraints and business rules in the
software development. Meanwhile, the behavior requirement is
a requirement that describes a behavior of system. Once the
requirement is classified based on previous two perspectives,
the next process is to prioritize requirements based on return on
investment (or ROI) [51], [52], [53]. From business
perspective, ROI is the most important factor to assess the
important of each requirement. The following presents a
ranking tree by combining those two perspectives.
24
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
ornain
Specific
H
Rank#l
Behavioral
Reg-
Rank #2
Rank #3
Domain
Specific
R e q, —
Rank hase<
on ROI
Behavioral
Rank based
on ROI
Rank based
on ROI
Domain
Specific
R e q.
Rank based
on ROI
Behavioral _[ Rank based
Req. on ROI
Rank based
on ROI
UiJHU i n ,
Specific Rank #10
R e q . J j
Behavioral
Reg.
Rank #11
Rank #12
Figure 4. Requirement Prioritization Tree
From the above figure, we give the highest priority for all
'basic requirements due to the fact that they must be
implemented even they do not increase the customer
satisfaction. We rank the lowest priority for all 'indifferent'
requirements, because customers do not concentrate on.
Additionally, we prioritize both of all 'delight' and 'attractive'
requirement based on ROI. In this paper, we propose to use a
cost-value approach to weight and prioritize requirements. This
paper proposes to use the following formula:
P(Req) = (Cost * CP)
(1)
Where:
• P is a prioritization value.
• Req is a requirement required to be prioritized.
• Cost is a total estimated cost of coding and testing for
each requirement.
• CP is an user-defined customer priority value. This value
is in the range between 1 and 10. 10 is the highest priority
and 1 is the lowest priority. This value aims to allow
customers to identify how important of each requirement
is from their perspective.
To compute the above cost for coding and testing, this
paper proposes to apply the following formula:
Cost = (EffCode*CostCode)+(EffTest*CostTest) (2)
Where
• Cost is a total estimated cost.
• EffCode is an estimated effort of coding for each
requirement. The unit is man-hours.
• CostCode is a cost of coding that is charged to customers.
This paper applies the cost- value approach to identify the
cost of coding for each requirement group (e.g. "Must-
Have", "Should-Have", "Could-Have" and "Wish"). The
unit is US dollar.
• EjfTest is an estimated effort of testing for each
requirement. The unit is man-hours.
• CostTest is a cost of testing that is charged to customers.
The approach to identify this value is similar to
CostCode' 's approach. The unit is US dollar.
In this paper, we assumed the following in order to
calculate CostCode and CostTest. Also, this paper assumes that
a standard cost for both activities is $100 per man-hours.
• A value is 1.5 of ("Must-Have", "Should-Have") - this
means that "Must-Have" requirements have one and half
times cost value than "Should-Have" requirements.
• A value is 3 of ("Must-Have", "Could-Have") - this
means that "Must-Have" requirements have three times
cost value than "Could-Have" requirements.
• A value is 2 of ("Should-Have", "Could-Have") - this
means that "Should-Have" requirements have two times
cost value than "Could-Have" requirements.
• A value is approximately 3 of ("Could-Have", "Wish") -
this means that "Could-Have" requirements have three
times cost value than "Wish" requirements.
Therefore, the procedure of requirement prioritization
process can be shortly described below:
1 . Provide estimated efforts of coding and testing for each
requirement.
2. Assign cost value for each requirement group based on
the previous requirement classification (e.g. "Must-
Have", "Should-Have", "Could-Have" and "Wish").
3. Calculate a total estimated cost for coding and testing,
by using the formula (2).
4. Define a customer priority for each requirement.
5. Compute a priority value for each requirement by
using the formula (1).
6. Prioritize requirements based on the higher priority
value.
Once the requirements are prioritized, the next proposed
step is to generate test scenario and prepare test case.
B. Test Case Generation Technique
This section presents an automated test scenario generation
derived from UML Use Case diagram 2.0. The big different
between UML Use Case diagram 1.0 and 2.0 is a package
notion that can group each use case into each package.
The following shows an example of UML Use Case
diagram 2.0.
25
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
O
Customer
^Jnquiry
-d(^Withdraw
Release 1
Release 2
S*\t\\bx Bank
./"^Own Acct.
^^X^TYansfor
NjX^TG Airlirm
^v^Support
i
Release 3
Figure 5. An Example of UML Use Case Diagram 2.0
From the above figure, the new notion in UML Use Case
diagram 2.0 is a package that is used for grouping each
function. There are three packages or releases. Each release
contains different functional requirement. The first release
contains two functions: inquiry and withdraw. The second
release is composed of: transfer own account and transfer to
other banks. The last release has only one function to support
Thai (TG) airline tickets.
Our approach aims to generate three test suites to cover the
above three packages while existing test case generation
techniques do not concentrate on. The first test suite is
developed for: inquiry and withdraw functions. The second test
suite is used for transferring own banks and other banks. The
last suite aims to a TG airline ticket support.
The approach is built based on Heumann's algorithm [23].
The limitation of our approach is to ensure that all use cases are
fully dressed. The fully dressed use case is a use case with the
comprehensive of information, as follows: use case name, use
case number, purpose, summary, pre-condition, post-condition,
actors, stakeholders, basic events, alternative events, business
rules, notes, version, author and date.
The proposed method contains four steps, as follows: (a)
extract use case diagram (b) generate test scenario (c) prepare
test data and (d) prepare other test elements. These steps can be
shortly described as follows:
1. The first step is to extract the following information
from fully dressed use cases: (a) use case number (b)
purpose (c) summary (d) pre-condition (e) post-
condition (f) basic event and (g) alternative events.
This information is called use case scenario in this
paper. The example fully dressed use cases of ATM
withdraw functionality can be found as follows:
Case
Case
y
Event
ve Events
ss
Id
Name
Rules
UC-
Withdra
To allow
1 . Insert
1. Select
(a)
001
w
bank's
Card
Inquiry
Input
customer
2. Input
2. Select
amount
s to
PIN
A/C Type
<=
withdraw
3. Select
3. Check
Outstan
money
Withdraw
Balance
ding
from
4. Select
Balanc
ATM
A/C Type
e
machines
5. Input
(b) Fee
anywher
Balance
charge
e in
6. Get
if using
Thailand
Money
7. Get Card
differen
tATM
machin
es
uc-
Transfer
To allow
1 . Insert
1. Select
Amoun
002
users to
Card
Inquiry
t<=
transfer
2. Input
2. Select
50,000
money to
PIN
A/C Type
baht
other
3. Select
3. Check
banks in
Transfer
Balance
Thailand
4. Select
from all
bank
ATM
5. Select
machines
"To"
account
6. Select
A/C Type
7. Input
Amount
8. Get
Receipt
9. Get Card
The above use cases can be extracted into the following use
case scenarios:
TABLE II.
Extracted Use Case Scenarios
Scenario Id
Summary
Basic Scenario
Scenario-001
To allow bank's
1 . Insert Card
customers to
2. Input PIN
withdraw money from
3. Select Withdraw
ATM machines
4. Select A/C Type
anywhere in Thailand.
5. Input Balance
6. Get Money
7. Get Card
TABLE I.
Example Fully Dressed Use Case
Use
Use
Summar
Basic
Alternati
Busine
26
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
Scenario-002
To allow bank's
customers to
withdraw money from 3
ATM machines
anywhere in Thailand
Scenario-003
Scenario-004
1 . Insert Card
2. Input PIN
Select Inquiry
4. Select A/C Type
5. Check Balance
6. Select Withdraw
7. Select A/C Type
8. Input Balance
9. Get Money
10. Get Card
To allow users to
transfer money to
other banks in
Thailand from all
ATM machines
To allow users to
transfer money to
other banks in
Thailand from all
ATM machines
1 . Insert Card
2. Input PIN
3. Select Transfer
4. Select bank
5. Select "To"
account
6. Select A/C Type
7. Input Amount
8. Get Receipt
9. Get Card
1 . Insert Card
2. Input PIN
3. Select Inquiry
4. Select A/C Type
5. Check Balance
6. Select Transfer
7. Select bank
Select "To"
account
9. Select A/C Type
10. Input Amount
11. Get Receipt
12. Get Card
TS-002
To allow bank's
1 . Insert Card
customers to
2. Input PIN
withdraw money from 3. Select Inquiry
ATM machines
4. Select A/C Type
anywhere in Thailand.
5. Check Balance
6. Select Withdraw
7. Select A/C Type
8. Input Balance
9. Get Money
10. Get Card
TS-003
To allow users to
1 . Insert Card
transfer money to
2. Input PIN
other banks in
3. Select Transfer
Thailand from all
4. Select bank
ATM machines
5. Select "To"
account
6. Select A/C Type
7. Input Amount
8. Get Receipt
9. Get Card
TS-004
To allow users to
1 . Insert Card
transfer money to
2. Input PIN
other banks in
3. Select Inquiry
Thailand from all
4. Select A/C Type
ATM machines
5. Check Balance
6. Select Transfer
7. Select bank
8. Select "To"
account
9. Select A/C Type
10. Input Amount
11. Get Receipt
12. Get Card
2. The second step is to automatically generate test
scenarios from the previous use case scenarios [23].
From the above table, we automatically generate the
following test scenarios:
3 . The next step is to prepare test data. This step allows to
manually prepare an input data for each scenario.
4. The last step is to prepare other test elements, such as
expected output, actual output and pass / fail status
TABLE III.
Generated Test Scenarios
Test Scenario
Summary
Basic Scenario
Id
TS-001
To allow bank's
1 . Insert Card
customers to
2. Input PIN
withdraw money from
3. Select Withdraw
ATM machines
4. Select A/C Type
anywhere in Thailand.
5. Input Balance
6. Get Money
7. Get Card
V. EVALUATION
The section describes the experiments design, measurement
metrics and results.
A. Experiments Design
1. Prepare Experiment Data. Before evaluating the
proposed methods and other methods, preparing
experiment data is required. In this step, 50
requirements and 50 use case scenarios are randomly
generated.
2. Generate Test Scenario and Test Case. A comparative
evaluation method has been made among the proposed
test scenario algorithm, Heumann's technique Jim [23],
Ryser's method [24], Nilawar's algorithm [33] and the
proposed method presented in the previous section. It
27
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
is included a prioritization requirement algorithm prior
to generate a set of test scenarios and test cases.
3. Evaluate Results. In this step, the comparative
generation methods are executed by using 50
requirements and 50 use case scenarios. These methods
are also executed for 10 times in order to find out the
average percentage of critical domain requirement
coverage, a size of test cases and total generation time.
In total, there are 500 requirements and 500 use case
scenarios executed in this experiment.
The following tables present how to randomly generate data
for requirements and use case scenarios respectively.
basic 7, basic 2 , . . . , basic n
TABLE IV.
Generate Random Requirements
Attribute
Approach
Requirement
ID
Randomly generated from the following
combination: Req + Sequence Number.
For example, Reql, Req2, Req3, ...,
ReqN.
Description
Randomly generated from the following
combination: Des + Sequence Number
same as Requirement ID.
For example, Desl, Des2, Des3, ...,
DesN.
Type of
Requirement
Randomly selected from the following
values: Functional AND Non-Functional.
MoSCoW
Criteria
Randomly selected from the following
values: Must Have (M), Should Have (S),
Could Have (C) and Won't Have (W)
Is it a critical
requirement
(Y/N)?
Randomly selected from the following
values: True (Y) and False (N)
TABLE V.
Generate Random Use Case Scenario
Attribute
Approach
Use case ID
Randomly generated from the
following combination: uCase +
Sequence Number. For example,
uCase h uCase 2 , ..., uCase n .
Purpose
Randomly generated from the
following combination: Pur +
Sequence Number same as Use case
ID. For example, Pur h Pur 2 , . . . ,
Pur n .
Pre-condition
Randomly generated from the
following combination: pCon +
Sequence Number same as Use case
ID. For example, pCon h pCon 2 , . . . ,
pCon n .
Basic Scenario
Randomly generated from the
following combination: uCase +
Sequence Number. For example,
B. Measurement Metrics
The section lists the measurement metrics used in the
experiment. This paper proposes to use three metrics, which
are: (a) size of test cases (b) total time and (c) percentage of
critical domain requirement coverage. The following describe
the measurement in details.
1. A Number of Test Cases: This is the total number of
generated test cases, expressed as a percentage, as follows:
% Size = (# Size I # of Total Size)*\W (3)
Where:
• % Size is a percentage of the number of test cases.
• # of Size is a number of test cases.
• # of Total Size is the maximum number of test cases in the
experiment, which is assigned 1,000.
2. A Domain Specific Requirement Coverage: This is an
indicator to identify the number of requirements covered in
the system, particularly critical requirements, and critical
domain requirements [5]. Due to the fact that one of the
goals of software testing is to verify and validate
requirements covered by the system, this metric is a must.
Therefore, a high percentage of critical requirement
coverage is desirable.
It can be calculated using the following formula:
% CRC = (# of Critical I # ofTotat)*100 (4)
Where:
• % CRC is the percentage of critical requirement coverage.
• # of Critical is the number of critical requirements
covered.
• # of Total is the total number of requirements.
3. Total Time: This is the total number of times the
generation methods are run in the experiment. This metric
is related to the time used during the testing development
phase (e.g. design test scenario and produce test case).
Therefore, less time is desirable.
It can be calculated using the following formula:
Total = PTime + CTime + RTime (5)
Where:
• Total is the total amount of times consumed by running
generation methods.
• PTime is the total amount of time consumed by
preparation before generating test cases.
• CTime is the time to compile source code / binary code in
order to execute the program.
• RTime is the total time to run the program under this
experiment.
28
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
C. Results and Discussion
This section discusses an evaluation result of the above
experiment. This section presents a graph that compares the
above proposed method to other three existing test case
generation techniques, based on the following measurements:
(a) size of test cases (b) critical domain coverage and (c) total
time. Those three techniques are: (a) Heumman's method (b)
Ryser's work and (c) Nilawar's approach. There are two
dimensions in the following graph: (a) horizontal and (b)
vertical axis. The horizontal represents three measurements
whereas the vertical axis represents the percentage value.
Compare Percentage of Size of Test Cases. Critical Domain Coverage and Total Time among Four
Test Case Generation Techniques
« ©0 00%
Critical Domain Coverage
Measurements
Figure 6. An Evaluation Result of Test Generation Methods
The above graph shows that the above proposed method
generates the smallest set of test cases. It is calculated as
80.80% where as the other techniques is computed over 97%.
Those techniques generated a bigger set of test cases, than a set
generated by the proposed method. The literature review
reveals that the smaller set of test cases is desirable. Also, the
graph shows that the proposed method consumes the least total
time during a generation process, comparing to other
techniques. It used only 30.20%, which is slightly less than
others. Finally, the graph presents that the proposed method is
the best techniques to coverage critical domains. Its percentage
is much greater than other techniques' percentage, over 30%.
From the above figure, this study determines and ranks the
above comparative methods into five ranking: 5 -Excellent, 4-
Very good, 3-Good, 2-Normal and 1-Poor. This study uses a
maximum and minimum value to find an interval value for
ranking those methods.
For a number of test cases, the maximum and minimum
percentage is 98% and 80.80%. The different between
maximum and minimum value is 17.2%. An interval value is
equal to a result of dividing the different values by 5. As a
result, the interval value is approximately 3.4. Thus, it can be
determined as follows: 5-Excellent (since 80.80% to 84.2%), 4-
Very good (between 84.2% and 87.6%), 3-Good (between
87.6% and 91%), 2-Normal (between 91% and 94.4%) and 1-
Poor (from 94.4% to 97.8%).
For an ability to cover critical domain specific requirement,
the maximum and minimum percentage is 53.20% and 19%.
The different value is 34.2%. The interval value is 6.84.
Therefore, it can be determined as follows: 5 -Excellent (since
46.36% to 53.2%), 4-Very good (between 39.52% and
46.36%), 3-Good (between 32.68% and 39.52%), 2-Normal
(between 25.84% and 32.68%) and 1-Poor (from 19% to
25.84%).
For a total time, the maximum and minimum percentage is
31.82% and 30.20%. The different between maximum and
minimum value is 1.62%. An interval value is equal to a result
of dividing the different values by 5. As a result, the interval
value is 0.324. Thus, it can be determined as follows: 5-
Excellent (since 30.2% to 30.524%), 4-Very good (between
30.524% and 30.848%), 3-Good (between 30.848% and
31.172%), 2-Normal (between 31.172% and 31.496%) and 1-
Poor (from 31.496% to 31.82%).
Therefore, the experiment result of those comparative
methods can be shown below:
TABLE VI.
A Comparison of Test Case Reduction Methods
Algorithm
A
Number of
Test Cases
Cover
Critical
Domain
Specific Req.
Total
Time
Heumann's
Method
1
1
5
Ryser's Method
1
1
1
Nilawar's
Method
1
1
1
Our Proposed
Method
5
5
5
In the conclusion, the proposed method is the best to
generate the smallest size of test cases with the maximum of
critical domain coverage and the least time consumed in the
generation process.
VI. CONCLUSION
In this paper, we introduced a new test case generation
method and process, with an additional requirement
prioritization process. The approach inserts an additional
process to ensure that all domain specific requirements are
captured during the test case generation. Also, the approach is
developed to minimize a number of test cases in order to be
able to select suitable test cases for execution. Additionally, we
proposed an automated approach to generate test cases from
fully described UML use cases, version 2.0. Our generated
method can generate many test suites derived from UML Use
Case diagram, 2.0. Existing test case generation methods
generate only a large single test suite that contains a lot of
numbers of test cases.
Furthermore, we conducted an evaluation experiment with
a random requirements and fully described use cases. Our
29
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
evaluation result reveals that the proposed method is the most
recommendation automated test case generation methods for
maximizing critical domain requirement coverage. Also, the
result present that the proposed method is one of best methods
to minimize a number of test cases.
The future research, we plan to enhance an ability to
prioritize requirements and conduct a large experiment for a
large system development
References
[I] Ahl, V. "An Experimental Comparison of Five Prioritization
Methods" Master's Thesis, School of Engineering, Blekinge
Institute of Technology, Ronneby, Sweden, 2005.
[2] Alessandra Cavarra, Charles Crichton, Jim Davies, Alan
Hartman, Thierry Jeron and Laurent Mounier. "Using UML for
Automatic Test Generation." Oxford University Computing
Laboratory, Tools and Algorithms for the Construction and
Analysis of Systems, TACAS'2000, 2000.
[3] Amaral. "A.S.M.S. Test case generation of systems specified in
Statecharts." M.S. thesis - Laboratory of Computing and
Applied Mathematics, INPE, Brazil, 2006.
[4] Annelises A. Andrews, Jeff Offutt and Roger T. Alexander.
"Testing Web Applications. Software and Systems Modeling.",
2004.
[5] Avik Sinha, Ph.D and Dr. Carol S. Smidts. "Domain Specific
Test Case Generation Using Higher Ordered Typed Languages
fro Specification." Ph. D. Dissertation, 2005.
[6] A. Bertolino. "Software Testing Research and Practice." 10th
International Workshop on Abstract State Machines
(ASM'2003), Taormina, Italy, 2003.
[7] A.Z. Javed, PA. Strooper and G.N. Watson "Automated
Generation of Test Cases Using Model-Driven Architecture."
Second International Workshop on Automation of Software Test
(AST'07), 2007.
[8] Beck, K. & Andres, C. "Extreme Programming Explained:
Embrace Change", 2nd ed. Boston, MA: Addison-Wesley, 2004.
[9] Boehm, B. & Ross, R. "Theory- W Software Project
Management: Principles and Examples." IEEE Transactions on
Software Engineering 15, 4: 902-916, 1989.
[10] B.M. Subraya, S.V. Subrahmanya "Object driven performance
testing in Web applications." in: Proceedings of the First Asia-
Pacific Conference on Quality Software (APAQS'00), pp. 17-26,
Hong Kong, China, 2000.
[II] Chien-Hung Liu, David C. Kung, Pei Hsia and Chih-Tung Hsu
"Object-Based Data Flow Testing of Web Applications."
Proceedings of the First Asia-Pacific Conference on Quality
Software (APAQS'00), pp. 7-16, Hong Kong, China, 2000.
[12] C.H. Liu, D.C. Kung, P. Hsia, C.T. Hsu "Structural testing of
Web applications." in: Proceedings of 11th International
Symposium on Software Reliability Engineering (ISSRE 2000),
pp. 84-96, 2000.
[13] Davis, A. "The Art of Requirements Triage." IEEE Computer
36, 3 p: 42-49, 2003.
[14] Davis, A. "Just Enough Requirements Management: Where
Software Development Meets Marketing." New York: Dorset
House (ISBN 0-932633-64-1), 2005.
[15] David C. Kung, Chien-Hung Liu and Pei Hsia "An Object-
Oriented Web Test Model for Testing Web Applications." In
Proceedings of the First Asia-Pacific Conference on Quality
Software (APAQS'00), page 1 1 1, Los Alamitos, CA, 2000.
[16] Donald Firesmith "Prioritizing Requirements. Journal of Object
Technology", Vol.3, No8, 2004.
[17] D. Harel "On visual formalisms." Communications of the ACM,
vol. 31, no. 5, pp. 514-530, 1988.
[18] D. Harel. "Statecharts: A Visual Formulation for Complex
System." Sci.Comput. Program. 8(3):232-274, 1987.
[19] Flippo Ricca and Paolo Tonella "Analysis and Testing of Web
Applications." Proc. of the 23rd International Conference on
Software Engineering, Toronto, Ontario, Canada, pp.25-34,
2001.
[20] Harel, D. "Statecharts: a visual formalism for complex system."
Science of Computer Programming, v. 8, p. 231-274, 1987.
[21] Hassan Reza, Kirk Ogaard and Amarnath Malge "A Model
Based Testing Technique to Test Web Applications Using
Statecharts." Fifth International Conference on Information
Technology, 2008.
[22] Ibrahim K. El-Far and James A. Whittaker "Model-based
Software Testing", 2000.
[23] Jim Heumann "Generating Test Cases From Use Cases."
Rational Software, 2001.
[24] Johannes Ryser and Martin Glinz "SCENT: A Method
Employing Scenarios to Systematically Derive Test Cases for
System Test", 2000.
[25] Karl E. Wiegers "First Things First: Prioritizing Requirements."
Published in Software Development, 1999.
[26] Karlsson, J. "Software Requirements Prioritizing." Proceedings
of the Second International Conference on Requirements
Engineering (ICRE'96). Colorado Springs, CO, April 15-18,
1996. Los Alamitos, CA: IEEE Computer Society, p 110-116,
1996.
[27] Karlsson, J. "Towards a Strategy for Software Requirements
Selection." Licentiate. Thesis 513, Linkoping University, 1995.
[28] Karlsson, J. & Ryan, K. "A Cost-Value Approach for
Prioritizing Requirements." IEEE Software September/October,
p67-75, 1997.
[29] Leffingwell, D. & Widrig, D. "Managing Software
Requirements: A Use Case Approach", 2nd ed. Boston, MA:
Addison-Wesley, 2003.
[30] Leslie M. Tierstein "Managing a Designer / 2000 Project."
NYOUG Fall'97 Conference, 1997.
[31] L. Brim, I. Cerna, P. Varekova, and B. Zimmerova
"Component-interaction automata as a verification oriented
component-based system specification." In: Proceedings
(SAVCBS'05), pp. 31-38, Lisbon, Portugal, 2005.
[32] Mahnaz Shams, Diwakar Krishnamurthy and Behrouz Far "A
Model-Based Approach for Testing the Performance of Web
Applications." Proceedings of the Third International Workshop
on Software Quality Assurance (SOQUA'06), 2006.
[33] Manish Nilawar and Dr. Sergiu Dascalu "A UML-Based
Approach for Testing Web Applications." Master of Science
with major in Computer Science, University of Nevada, Reno,
2003.
30
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
[34] Moisiadis, F. "Prioritising Scenario Evolution." International
Conference on Requirements Engineering (ICRE 2000), 2000.
[35] Moisiadis, F. "A Requirements Prioritisation Tool." 6th
Australian Workshop on Requirements Engineering (AWRE
2001 ). Sydney, Australia, 2001 .
[36] M. Prasanna S.N. Sivanandam R.Venkatesan R.Sundarrajan "A
Survey on Automatic Test Case Generation." Academic Open
Internet Journal, 2005.
[37] Nancy R. Mead "Requirements Prioritization Introduction."
Software Engineering Institute, Carnegie Mellon University,
2008.
[38] Park, J.; Port, D.; & Boehm B. "Supporting Distributed
Collaborative Prioritization for Win-Win Requirements Capture
and Negotiation." Proceedings of the International Third World
Multi-conference on Systemics, Cybernetics and Informatics
(SCI'99) Vol. 2. 578-584, Orlando, FL, July 31-August 4, 1999.
Orlando, FL: International Institute of Informatics and Systemic
(HIS), 1999.
[39] Rajib "Software Test Metric." QCON, 2006.
[40] Robert Nilsson, Jeff Offutt and Jonas Mellin "Test Case
Generation for Mutation-based Testing of Timeliness.", 2006.
[41] Saaty, T. L. "The Analytic Hierarchy Process." New York, NY:
McGraw-Hill, 1980.
[42] Shengbo Chen, Huaikou Miao, Zhongsheng Qian "Automatic
Generating Test Cases for Testing Web Applications."
International Conference on Computational Intelligence and
Security Workshops, 2007.
[43] Valdivino Santiago, Ana Silvia Martins do Amaral, N.L.
Vijaykumar, Maria de Fatima, Mattiello-Francisco, Eliane
Martins and Odnei Cuesta Lopes "A Practical Approach for
Automated Test Case Generation using Statecharts", 2006.
[44] Vijaykumar, N. L.; Carvalho, S. V.; Abdurahiman, V. "On
proposing Statecharts to specify performance models."
International Transactions in Operational Research, 9, 321-336,
2002.
[45] Wiegers, K. "E. Software Requirements", 2nd ed. Redmond,
WA: Microsoft Press, 2003 .
[46] Xiaoping Jia, Hongming Liu and Lizhang Qin "Formal
Structured Specification for Web Application Testing". Proc. of
the 2003 Midwest Software Engineering Conference
(MSEC'03). Chicago, IL, USA. pp.88-97, 2003.
[47] Yang, J.T., Huang, J.L., Wang, F.J. and Chu, W.C.
"Constructing an object-oriented architecture for Web
application testing." Journal of Information Science and
Engineering 18, 59-84, 2002.
[48] Ye Wu and Jeff Offutt "Modeling and Testing Web-based
Applications", 2002.
[49] Ye Wu, Jeff Offutt and Xiaochen "Modeling and Testing of
Dynamic Aspects of Web Applications, Submitted for
publication." Technical Report ISE-TR-04-01,
www.ise.gmu.edu/techreps/, 2004.
[50] Zhu, H., Hall, P., May, J. "Software Unit Test Coverage and
Adequacy." ACM Comp. Survey 29(4), pp 366-427, 1997.
[51] Kano Noriaki, Nobuhiku Seraku, Fumio Takahashi,Shinichi
Tsuji. "Attractive Quality and Must-Be Quality." Journal of the
Japanese Society for Quality Control. 14(2), pp 39-48, 1984.
[52] Cadotte, Ernest R., Turgeon, Normand "Dissatisfiers and
Satisfiers: Suggestions from Consumer Complaints and
Compliments." Journal of Consumer Satisfaction,
Dissatisfactions and Complaining Behavior. 1, pp 74-79, 1988.
[53] Brandt, D. Randall "How service marketers can identify value-
enhancing service elements." Journal of Services Marketing.
2(3), pp 35-41, 1988.
[54] Herzberg, Frederick, Mausner, B., Snyderman, B.B. "The
motivation to work." New York: Wiley, 2 nd edition, 1959.
31
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
White-Box Test Reduction Using Case-Based
Maintenance
Siripong Roongruangsuwan
Autonomous System Research Laboratory
Faculty of Science and Technology, Assumption University
Bangkok, Thailand
Jirapun Daengdej
Autonomous System Research Laboratory
Faculty of Science and Technology, Assumption University
Bangkok, Thailand
Abstract — Software testing has been proven that it takes around
50-70% of the costs associated with the large development of
commercial software systems. Many reduction techniques have
been proposed to reduce costs. Unfortunately, the cost is usually
over budget and those techniques are failed to reasonably control
costs. The primarily outstanding research issues, motivated this
study, are a large number of redundancy test cases and a
decrease of ability to detect faults. To resolve these issues, this
paper proposes new deletion algorithms to minimize a number of
white-box test cases, while maximizing an ability to reveal faults,
by using a concept of case-based maintenance. Our evaluations
have shown that the proposed techniques can significantly reduce
a number of unnecessary test cases while preserving the
capability of fault detection.
Keywords-component; Test reduction, test case reduction,
deletion method, case based maintenance and test case deletion
I.
Introduction
Software Testing is an empirical investigation conducted to
provide stakeholders with information about the quality of the
product or service under test [13], with respect to the context in
which it is intended to operate. Software Testing also provides
an objective, independent view of the software to allow the
business to appreciate and understand the risks of
implementation of the software. The software testing
techniques include the process of executing a program or
application with the intent of finding software bugs. It can also
be stated as the process of validating and verifying that
software meets the business and technical requirements that
guided its design and development, so that it works as
expected. Software Testing can be implemented at any time in
the development process; however, the most test effort is
employed after the requirements have been defined and coding
process has been completed.
Many researchers [6], [7], [8], [9], [10], [19], [24], [25],
[26], [28], [30], [36], [37], [39] have proven that these test case
reduction methods can reserve the ability to reveal faults.
However, there are many outstanding research issues in this
area. The motivated research issues are: a large number of test
cases particularly redundancy test cases and a decrease of
ability to detect faults. The study shows that test case reduction
methods have been researched over a long period of time, such
as test case prioritization, random approach and coverage-based
test case reduction techniques. Also, the study reveals that
coverage-based approaches are wildly used and researched.
Therefore, we concentrate on an approach to reduce test cases
based on the coverage factor. Many coverage factors have been
proposed over a long period of time. Unfortunately, existing
factors and test case reduction methods ignore the complexity
and impact of test cases. Thus, we propose to reduce a number
of test cases by considering both of test case complexity and
impact.
Our study [5] shows that one of effective approaches that
significantly reduce a number of redundancy test cases is to
apply the concept of artificial intelligent. There are many
artificial intelligent concepts, such as neutral network, fuzzy
logic, learning algorithms and case-based reasoning (CBR).
CBR is one of the most popular and actively researched areas
in the past. The researches [4], [5], [26], [35] show that CBR
has identical problems as same as software testing topic, such
as a huge number of redundancy cases and a decrease of
system's ability to resolve problems.
Fundamentally, there are four steps in the CBR system,
which are: retrieve, reuse, revise and retain. These steps can
lead to a serious problem of uncontrollably growing cases in
the system. However, the study shows that there are many
proposed techniques in order to control a number of cases in
the CBR system, such as add algorithms, deletion algorithms
and maintenance approaches. CBR have been investigated by
CBR researchers in order to ensure that only small amounts of
efficient cases are stored in the case base.
The previous work [27] shows that deletion algorithms are
the most popular and effective approaches to maintain a size of
the CBR system. There are many researchers have proposed
several deletion algorithms [4], [20], [35], such as random
method, utility approach and footprint algorithm. These
algorithms aim to: (a) remove all redundancy or unnecessary
cases (b) minimize a number of cases and (c) maintain the
ability of solving problems. Nevertheless, each technique has
strength and weakness. Some methods are suitable for
removing cases. Some methods are perfectly suitable for
reducing time. Some may be used for reserving the problem
solving capability. Eventually, the previous work [27]
discovered several effective methods (e.g. confidential case
filtering method, coverage value algorithm and confidential
coverage approach) to remove those cases, minimize size of
32
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
CBR and reduce amount of time, while preserving the ability of
CBR system's problem solving skill. Therefore, this paper
applies those effective deletion techniques to resolve the
problems of software testing.
In the light of software testing, the proposed techniques
focus on how to maintain the test case while maintaining the
capability of fault detection. It is appear that test cases in this
paper are treated as cases in the CBR system. Also, there is an
assumption that a given set of test cases are generated by a
path-oriented test case generation technique. The path-oriented
technique is widely used for a white-box testing that derives
test cases from available source code.
Section 2 discusses an overview of test case reduction
techniques and approach to maintain CBR. Section 3 provides
a definition of terminologies used in this paper. Section 4 lists
the outstanding research issues motivated this study. Section 5
proposes deletion algorithms using the concept of CBR.
Section 6 describes an evaluation method and discusses an
evaluation result. The last section represents all source
references used in this paper.
II. Literature Review
This section describes an overview of test case reduction
techniques and the concept of case based maintenance. The
following describes those two areas in details.
A. Test Case Reduction Techniques
This section discusses and organizes test case reduction (or
TCR) techniques researched in 1995-2006. This study shows
that there are many researchers who proposed a method to
reduce unnecessary test cases (also known as redundancy test
cases), like Offutt [2], Rothermel [8], McMaster [25] and
Sampth [31]. These techniques aim to remove and minimize a
size of test cases while maintaining the ability to detect faults.
The literature review [6], [7], [8], [9], [10], [11], [19], [24],
[25], [36], [37], [39] shows that there are two types of
reduction techniques, which are: (a) pre-process and (b) post-
process. First, the pre-process is a process that immediately
reduces a size of test cases after generating. Typically, it is
occurred before regression testing phase. Second, the post-
process is a process that maintains and removes unnecessary
test cases, after running the first regression testing activities.
Although these techniques can reduce the size of test cases, but
the ability to reveal faults seems slightly to be dropped.
However, Jefferson Offutt [5] and Rothermel [6], [7], [8], [9],
[10], [19], [20], [21], [32] has proven that these test case
reduction techniques have many benefits, particularly during
the regression testing phase, and most of reduction techniques
can maintain an acceptable rate of fault detection. The
advantages of these techniques are: (a) to spend less time in
executing test cases, particularly during the regression testing
phase (b) to significantly reduce time and cost of manually
comparing test results and (c) to effectively manage the test
data associated with test cases. This study proposes a new "2C"
classification of test case reduction techniques, classified based
on their characteristics, as follows: (a) coverage-based
techniques and (b) concept analysis-based techniques.
Autonomous System Research Laboratory, Faculty of Science and
Technology, Assumption University.
B. Case-Based Maintenance (CBM)
Due to the CBR's life cycle [16], the case base size grows
rapidly. That is caused a serious problem directly, for instance,
duplicate data, inconsistency data, incorrect data, and an
expense of searching for an appropriate case in a large case
base size. CBR can be classified as one of the Artificial
Intelligence algorithms. CBR solves new problem by retrieving
the similar case from the existing case base and then adapts the
existing case according to the target problem. Over the time,
CBR is growing. When the uncontrollable case-based growth is
occurred, the performance of CBR is decreasing. Therefore, the
maintenance process is required in order to preserve or improve
the performance of the system. The process of maintaining
CBR is called CBM. David C. Wilson [8] presented the overall
concepts of CBR and case based maintenance. This paper
focused on the case based maintenance (CBM) approach in
term of the framework. In other words, this paper described the
type of data collection and how the case based maintenance
works. There were so many policies for CBM, for example,
addition, deletion, and retain.
"CBM was defined as the process of refining a CBR system's
case-base to improve the system 's performance. It implements
policies for revising the organization or contents
(representation, domain content, accounting information, or
implementation) of the case-base in order to facilitate future
reasoning for a particular set of performance objectives. "
These studies [2], [3], [4], [16], [17], [27], [35] reveal that
several deletion algorithms have been proposed. For example, a
random approach (RD), utility deletion algorithm (UD),
footprint deletion algorithm (FD), footprint utility deletion
algorithm (FUD) and iterative case filtering algorithm (ICF).
RD is the simplest approach, which removes the case
randomly. UD deletes the case that has minimum utility value.
Footprint algorithm uses the competence model and removes
the auxiliary case from the system. FUD is a hybrid approach
between Utility algorithm and Footprint algorithm, and is
concerned with the competence model and the utility value.
Finally, ICF focuses on the case, which the reachability set is
greater than the coverage set [16], [27].
III. DEFINITION
This section describes a definition of CBR terminologies
used in the software testing area.
TABLE I.
Definitions of CBR for Software Testing
Element
CBR
Software Testing
Coverage
set
Coverage Set is the
set of target problems,
which it can be used
to solve successfully
[4].
Coverage set means a
set of stages /
elements, which they
can be used to test
successfully and
reveal faults.
Reachability
set
Reachability Set is the
set of case bases that
can be used to solve
Reachability set means
a set of test cases that
can be used to reveal
33
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
the target problem [4].
faults.
Competence
set
Competence is the
range of the target
problem that can be
solved successfully
[4].
Competence is the
range of ability to
reveal faults that can
be used to test
successfully.
Auxiliary
set
Auxiliary Case is a
case that does not
have a direct effect on
the competence of a
system when it is
deleted [4].
Auxiliary case is a test
case that does not have
a direct effect on the
competence of a
system when it is
removed.
Pivot set
Pivotal Case is the
case that does have a
direct effect on the
competence of a
system if it is deleted
[1], [29].
Pivotal case is a test
case that does have a
direct effect on the
ability to reveal faults
if it is deleted.
IV. RESEARCH PROBLEM
This section discusses the details of research issues
motivated this study. The literature review reveals that [13],
[23], [24], [25], [31], [38] those research issues are: (a) a large
number of redundancy test cases and (b) a decrease of the
ability to reveal faults. These research issues can be elaborated
in details as follows:
First, the literature review shows that redundancy test cases
are test cases tested by multiple test cases. Many test cases that
are designed to test the same things (e.g. same functions, same
line of code or same requirements) are duplicated. Those
duplicated tests are typically occurred during testing activities,
particularly during regression testing activities [13], [23], [24],
[25], [31], [38]. Those duplicated tests can be eventually
removed in order to minimize time and cost to execute tests.
The following shows an example of redundancy test cases.
Id
Description
input
Sequence
Expected
Result
Actual
Result
Status
Pass / Fail
06-1
Withdraw Honey
torn ATM
1. Insert ATM Card
2. Insert PIN
3. Select -Wthwatf
1 Select MC Type
5. Identify Amount
S. Ciek-OK-
7. Receive Money
'. Receive Carrl
Money is
pluintfl
Baldnceis T^
ialeulated.
ATM Card iJ-^
cturncd. ~
)
M2
hquiryS
Mthdraw Money
broiTi ATM
PIN. Amount
1. Insert ATM Card
2 Insert PIN
». SeleetWCType
5. Click "OK"
> Select -Wth*aw
7. Select AiCType
El Identify Amount
9. Ckck-OK"
1 0. Receive Money
Cuirenlbalancel
s displayed. /U
Monty is /
eturned (
Balance is \f
sacculated. \J
ATM Card is
eturned.
Redu
"kit
umber of
Cases
J
Figure 1 . An Example of Control Flow Graph
From the above figure, there are two test cases, with the
duplicated sequence and expected result, designed to test a
"withdraw" feature in ATM machine. The sequence of the first
test case is: (a) insert ATM card (b) insert PIN (c) select
"withdraw" (d) select account type (e) identify an amount (f)
click "OK" button (g) receive money and (h) receive card. The
sequence of the second test case is similar to the first one.
However, the additional step in the second case is to inquiry a
balance amount before withdrawing the money. Therefore, it is
appear that the first test case is a part of the second test case.
We call the first test case as a redundancy test case.
The study shows that there are many proposed methods to
delete those duplicated test cases such as McMaster's work [24]
[25], Jeffs method [13] and Khan's approach [23]. Also, the
study shows that one of the most interesting research issues is
to minimize those duplicated tests and reduce cost of executing
tests. Although there are many proposed methods to resolve
that issue, that issue is still remaining. Thus, it is a challenge
for researchers to continuously improve the ability to remove
duplicated tests.
Last, test cases are designed to reveal faults during software
testing phase. The empirical studies [8], [10], [19], [20], [21],
[30], [32], [39] describe that reducing test cases may impact to
the ability of detect faults. Many reduction methods decrease a
capability of testing and reveal those faults. Therefore, one of
outstanding research challenges for researchers is to remove
tests while preserving the ability to defect faults.
V. PROPOSED METHOD
For evolving software, test cases are growing dramatically.
The more test cases software test engineers have, the more time
and cost software test engineers consume. The literature review
shows that regression testing activities consume a significant
amount of time and cost. Although, a comprehensive set of
regression selection techniques [8], [9], [10], [19] has been
proposed to minimize time and cost, there is an available room
to minimize size of tests and clean up all unnecessary test
cases. Thus, removing all redundancy test cases is desirable.
There are many approaches to reduce redundancy test cases
and applying an artificial intelligent concept in the test case
reduction process is an innovated approach. The literature
review [5], [27] shows that there are many areas of artificial
intelligent concept, such as artificial neutral network, fuzzy
logic, learning algorithms and CBR concept. Also, it reveals
that CBR has a same research issue as software testing has. The
issue is that cases in the CBR system will be consistency
growing bigger and larger all the time. There are four steps in
CBR that can uncontrollably grow a size of the system:
retrieve, reuse, revise and retain. Therefore, many CBR papers
aim to reduce all redundancy cases, known as "deletion
algorithms". The smaller size of CBR system is better and
desirable. Due to the fact that CBR has the same problem as
software testing and this paper focuses on reduction methods,
therefore, this paper proposes to apply CBR deletion
algorithms to the test case reduction techniques.
This paper introduces three reduction methods that apply
CBR deletion algorithms: TCCF, TCIF and PCF methods.
34
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
Those techniques aim to reduce a number of test cases
generated by path-oriented test case generation technique. This
technique is used for white-box testing only. However, the
generation methods are out of the scope of this paper.
The limitation of the proposed deletion algorithms are: (a)
those methods are perfectly suitable for only white-box testing
techniques and (b) path coverage may not be applicable for a
large system that contains over million lines of code.
A. Example of Test Cases
Given a set of test cases generated, this study discusses the
use of a number of case maintenance techniques, which have
been investigated by CBR researchers in ensuring that only
small amount of cases are stored in the case base, thereby
reducing number of test cases should be used in software
testing. Similar to what happen to software testing, a number of
CBR researchers have focused on finding approaches
especially for reducing cases in the CBR systems' storages.
This paper proposes to use the path coverage criteria in
order to reduce redundancy test cases. This is because path
coverage has a huge benefit of required very thorough testing
activities. The following describes in details of the above path
coverage using in the software testing field. Let S = {si, s2, s3,
s4, s5} to be a set of stage in the control flow graph. The
control flow graph can be derived from the source-code or
program. It is a white-box testing. Thus, each state represents a
block of code. The techniques that aim to generate and derive
test cases from the control flow graph are well-known as path-
oriented test case generation techniques. These techniques are
widely used to generate test cases. There are many research
papers on this area. However, the test case generation
techniques are out of scope in this paper.
Figure 2. An Example of Control Flow Graph
From the above figure, this paper assumes that each state
can reveal a fault. Thus, an ability to reveal faults of five states
is equal to 5. Also, it is assumed that every single transaction
must be tested. This example is used in the rest of paper.
Let TCn = {si, s2, ...,sn} where TC is a test case and sn is
a stage or node in the path-oriented graph that is used to be
tested. From the above figure, a set of test cases can be derived
as follows:
TCj = {s h s 2 J
TC 2 = {s h s 3 J
TC 3 = {si, s 4 J
TC 4 = {si, s 2 , s 3 J
TC 5 = {s h s 3 , s 5 J
TC 6 = {sj, s 4 , s 3 j
TC 7 = {s h s 2 , s 3 , s 5 j
TC 8 = {s h s 4 , s 3 , s 5 j
TC 9 = {s 2 , s 3 J
TC 10 = {s 2 , s 3 , s 5 J
TC n = {s 3 , s 5 J
TC 12 = {s 4 , s 3 J
TC 13 = {s 4 , s 3 , s 5 J
From the figure 2, we assume the following: (a) each state
represents a block of source code (b) each state can reveal only
1 fault; the total ability to reveal faults is 5 and (c) every single
transaction in the control flow graph must be tested.
The following describes the proposed methods that apply
the concept of CBR in details:
B. Path Coverage for Filtering (PCF)
Code coverage analysis is a structural testing technique
(also known as white box testing). Structural testing compares
test program behaviour against the apparent intention of the
source code. This contrasts with functional testing (also
referred to black-box testing), which compares test program
behaviour against a requirements specification. Structural
testing examines how the program works, taking into account
possible pitfalls in the structure and logic. Functional testing
examines what the program accomplishes, without regard to
how it works internally. Structural testing is also called path
testing since you choose test cases that cause paths to be taken
through the structure of the program. The advantage of path
cover is that it takes responsible for all statements as well as
branches across a method. It requires very thorough testing.
This is an effective substitute of other coverage criteria. The
path coverage is used as coverage value in this technique. The
Coverage value is combined into the addition policy for adding
significant case [12]. Within the adding algorithm along with
the coverage weight value stated in the review, the concept of
deletion algorithm and the coverage have been proposed. The
coverage value can specify how many nodes that the test case
can cover. In other words, the coverage value is an indicator to
measure that each test case covers nodes. It means that the
higher coverage value is, the more nodes can be contained and
covered in the test case.
Let Cov(n) = value where Cov is a coverage value, value is
a number of test cases in each coverage group and n is a
coverage relationship.
The procedure of this method can be elaborated briefly as
the following steps.
The first step is to determine a coverage set. From figure 2,
each coverage set can be identified as follows:
Coverage (l) = {rC;}
35
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
Coverage (2) = {TC 2 }
Coverage (3) = [TC 3 ]
Coverage (4) = {TC h TC 4 , TC 9 ]
Coverage (5) = [TC 2 , TC 5 , TC n ]
Coverage (6) = {TC 3 , TC 6 , TC 12 ]
Coverage (7) = {TC h TC 4 , TC 7 , TC 9 , TC 1(h TC n ]
Coverage (8) = {TC 3 , TC 6 , TC 8 , TC n , TC 12 , TC 13 }
Coverage (9) = {TC 9 }
Coverage (10) = {TC 9 , TC 10 , TC n }
Coverage (11) = {TC n }
Coverage (12) = {TC 12 }
Coverage (13) = {TC lh TC 12 , TC 13 }
The second step is also to determine a reachability set. The
reachability set can be figured out from the above coverage set,
based on the given definition in this paper. Therefore, the
reachability set can be identified as follows:
Reachability (TCj) = {1,4, 7}
Reachability (TC 2 ) = {2, 5}
Reachability (TC 3 ) = {3, 6, 8}
Reachability (TC 4 ) = {4,1}
Reachability (TC 5 ) = {5}
Reachability (TC 6 ) = {6, 8}
Reachability (TC 7 ) = {7}
Reachability (TC 8 ) = {8}
Reachability (TC 9 ) = {4, 7, 9, 10}
Reachability (TC 10 ) = {7, 10}
Reachability {TC n ) = {5, 7, 8, 10, 11, 13}
Reachability (TC 12 ) = {6, 8, 12, 13}
Reachability (7C i5 ) = {8, 13}
The next step is to calculate a coverage value. This paper
proposes to calculate a coverage value based on a number of
test cases in each coverage group. Therefore, the coverage
value can be computed as follows:
Cov{\) = 1, Cov{2) = 1, Cov(3) = 1, Cov{A) = 3, Cov(5) = 3,
Cov(6) = 4, Cov(l) = 6, Cov(8) = 6, Cov(9) = 1, Cov(lO) = 3,
Cov(ll) = 1, Cov(12) = 1 and Cov(l3) = 3
Afterward, the step is to determine potential removable test
cases. These test cases can be identified when their number of
members in the reachability set is greater than a number of
members in the coverage set. Therefore, the potential removal
test cases are: TC h TC 2 , TC 3 , TC 9 , TC n and TC 12 .
The last step is to removes all test cases with minimum
coverage value, in the potential removable test cases.
Unfortunately, TC h TC 2 , TC 3 , TC 9 , TC n and TC 12 are removed
due to that they have the minimum coverage value.
C. Test Case Complexity for Filtering (TCCF)
A complexity of test case is the significant criteria in this
proposed method [1], [16]. In this paper, the complexity of test
case measures a number of states included in each test case. We
define the test case complexity as follows:
Definition 1: Let Cplx(TC) = {High, Medium, Low J where
Cplx is a complexity of test case, TC is a test case and the
complexity value can be measured as:
• High when a number of states are greater than an
average number of states in the test suite.
• Medium when a number of states are equal to an
average number of states in test suites.
• Low when a number of states are less than an
average number of states in the test suites.
The procedures of this method can be described briefly in
the following steps. The first two steps are to identify coverage
and reachability set.
Next, the step is to define an auxiliary set. Test cases that
can be included in the auxiliary set have a greater number of
members in the reachability set than a number of members in
the coverage set. From figure 2, therefore, the auxiliary set can
be identified as follows:
Auxiliary set = {TC h TC 2 , TC 3 , TC 9 , TC n , TC 12 }
Afterward, the method computes a complexity value for all
test cases in the above auxiliary set. From figure 2 and test
suites that contain 13 test cases, the average number of states is
(2+2+2+3+3+3+4+4+2+3+2+2+3)/13, which is equal to 3.
Based on the average number of states, the complexity value
for each test case can be computed as follows:
Cplx(TCj) = Low, Cplx(TC 2 ) = Low, Cplx(TC 3 ) = Low,
Cplx(TC 4 ) = Medium, Cplx(TC 5 ) = Medium, Cplx(TC 6 ) =
Medium, Cplx(TC 9 ) = Low, Cplx(TC 10 ) = Medium, Cplx(TC n )
= Low, Cplx(TC 12 ) = Low and Cplx(TC 13 ) = Medium
Finally, the last step removes test cases with minimum of
complexity value from the auxiliary set. Thus, TC b TC 2 , TC 3 ,
TC 9 , TC]j and TC 12 are removed.
D. Test Case Impact for Filtering (TCIF)
The study [21] shows that software is error-ridden in part
because of its growing complexity. Software is growing more
complex every day. The size of software products is no longer
measured in thousands of lines of code, but it measures in
millions. Software developers already spend approximately 80
percent of development costs [18] on identifying and correcting
defects, and yet few products of any type other than software
are shipped with such high levels of errors. Other factors
contributing to quality problems include marketing strategies,
limited liability by software vendors, and decreasing returns on
testing and debugging, according to the study. At the core of
these issues is difficulty in defining and measuring software
quality. Due to the fact that defining and measuring a quality of
software is important and difficult, the impact of inadequate
36
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
testing must not be ignorance. The impact of inadequate testing
could be lead to the problem of poor quality, expensive costs
and huge time-to-market. In conclusion, software testing
engineers require identifying the impact of each test case in
order to acknowledge and understand clearly the impact of
ignoring some test cases.
In this paper, an impact value is an impact of test cases in
term of the ability to detect faults if those test cases are
removed and not be tested. We define the test case impact as
follows:
Definition 2: Let ImpiTC) = {High, Medium, Low} where
Imp is an impact if a test case is removed, TC is a test case and
the impact value can be measured as:
• High when the test case has revealed at least one
fault for many times.
• Medium when the test case has revealed faults for
only one time.
• Low when the test case has never revealed faults.
The procedure of this method is similar to the previous
method. The only different is that this method aims to use an
impact value instead of complexity value. Therefore, the fire
three steps are to: identify coverage set, define reachability set
and determine an auxiliary set. Afterward, the next step is to
compute and assign an impact value. The method computes the
impact value for all test cases in the above auxiliary set. From
figure 2, the impact value for each test case can be computed as
follows:
Imp(TCj) = Low, Imp(TC 2 ) = High, Imp(TC 3 ) = Medium,
Imp(TC 4 ) = Low, Imp(TC 5 ) = High, Imp(TC 6 ) = Medium,
Imp(TC 9 ) = Low, Imp(TC 10 ) = Low, Imp(TC n ) = Low,
ImpiTC 12) - Low and Imp(TC 13 ) = Low
Finally, the last step removes test cases with minimum of
impact value from the auxiliary set. Thus, TC h TC 4 , TC 7 , TC 9 ,
TC] , TC]], TC] 2 and TC ]3 are removed.
VI.
EVALUATION
This section describes an experiments design,
measurement metrics and results.
This paragraph designs an experiment used to evaluate and
determine the best reduction methods. This paper proposes the
following three steps. First, the experiment proposes to
randomly generate 2,000 test data used in the
telecommunication industry. In this experiment, the test data is
represented as test case. Second, the experiment executes
reduction methods with the generated test cases and compares
among the following reduction methods: RD, UD, FD, FUD,
ICF and three proposed methods (e.g. TCCF, TCIF and PCF).
This step randomly simulates defects for each test case in
order to determine an ability to reveal faults. Third, the
experiment aims to run the above methods for 10 times in
order to calculate the average value for each metric. The
metrics used in this experiment are described in details in next
section. Afterward, the experiment compares the values and
evaluates a result by generating a comparison graph in order to
determine the most recommended reduction approach.
The following table lists the description of each test data
that need to be generated randomly.
TABLE II.
A Form of Test Cases
Attribute
Description
Data
Type
Test Id
A unique index to reference
test data. The value is a
sequence number, starting
atl.
Numeric
A Set of Input Data
Full Name
A first and last name who
own the mobile phone.
String
Name
A mobile brand name. The
value is a range of iPhone,
BlackBerry, Nokia, LG,
Sony Ericsson and
Samsung.
String
Model
A mobile model.
String
Price
A price of mobile. The unit
of price is baht.
Numeric
Weight
A weight of mobile. The
unit of weight is gram (g).
Numeric
Height
A height of mobile. The
unit of height is centimeter
(cm).
Numeric
Graphics
A graphics mode option.
The value can be T or F
Boolean
WAP
A WAP mode. The value is
T or F
Boolean
Color
A mobile color. The color
can be Black, Gold, Silver,
Blue, and White
String
Game
A game mode. The value is
T or F
Boolean
Warranty
A mobile warranty. The
unit of warranty is month.
Numeric
The following table describes an approach to generate
random data using the above attributes respectively.
TABLE III.
Approach to Generate Random Test Cases
Attribute
Approach
Test Id
Generate randomly from the following
combination: t + Sequence Number.
For example, tl, t2, t3, ..., tn.
Name
Random from the following values:
iPhone, BlackBerry, Nokia, LG, Sony
and Samsung.
Model
Random from the following values:
iPhone - iPhone 2G and iPhone 3G.
BlackBerry - BlackBerry Bold 9000,
BlackBerry Bold 9700, BlackBerry
37
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
Curve 8300, BlackBerry Curve 8520,
BlackBerry Curve 8900, BlackBerry
Pearl, BlackBerry Pearl Flip 8200 and
BlackBerry Storm.
Nokia - Nokia N97 Mini, Nokia E90, Nokia
E72, Nokia 8800, Nokia N86 and Nokia
E66.
LG - Lotus Elite LX610, Accolade VX5600,
Chocolate Touch VX8575, Arena
GT950 and eXpo GW820.
Sony Ericsson - Xperia X10, Vivaz Pro,
Xperia X10 Mini Pro, Elm and Aspen.
Samsung - Samsung Gravity 2, Behold II,
Comeback, SGH-T139, SGH-T659 and
SGH-T239.
Price
Random from the following values:
High: 30,000 baht - 50,000 baht
Medium: 10,000 baht -29,999 baht
Low: 3,000 baht - 9,999 baht
Weight
Random from the following values: 4-15
Height
Random from the following values: 5-15
Graphics
Random from the following values:
True or False
WAP
Random from the following values:
True or False
Color
Random from the following values:
Black, Gold, Silver, Blue and White
Game
Random from the following values:
True or False
Warranty
Random from the following values:
Long: 12 months - 18 months
Medium: 6 months - 1 1 months
Short: 1 month - 5 months
methods are: RD, UD, FD, FUD, ICF, TCCF, TCIF and PCF.
Additionally, this section shows a graph format. There are two
dimensions in the following graph: (a) horizontal and (b)
vertical axis. The horizontal represents three measurements
whereas the vertical axis represents the percentage value.
The paragraph lists the measurement metrics used in the
experiment. The first measurement is a number of test cases.
The large number of test cases consumes time, effort and cost
more than the smaller size of test cases. Many reduction or
minimization approaches [6], [7], [8], [9], [10], [11], [19],
[24], [25], [36], [37], [39] have been proposed to minimize
size of test cases. This has proven that size is one of important
metrics in software testing area. The second is an ability to
reveal faults. It aims to measure the percentage of faults
detection. One of the goals of test case with a set of data is to
find defects. Thus, this metric is important criteria to measure
and determine which reduction methods can preserve the high
ability to reveal faults. The last measurement is a total of
reduction time: It is the total number of times running the
reduction methods in the experiment. This metric is related to
time used during execution time and maintenance time of test
case reduction methods. Therefore, less time is desirable.
This paragraph discusses an evaluation result of the above
experiment. This section presents the reduction methods
results in term of: (a) a number of test cases (b) ability to
reveal faults and (c) total reduction time. The comparative
Figure 2. A Graph Comparison of Deletion Methods
The above graph presents that both of FD and PCF
minimize a number of test cases by far better than other
reductions methods, approximately over 15%. Meanwhile,
both of them are the worst methods for preserving an ability to
reveal faults. FUD, TCCF and TCIF are best top three
methods to reserve a capability to detect faults. They are
greater than other methods over 22%. Unfortunately, they are
also the worst three methods that require a lot of time during a
reduction process. In the mean time, both of RD and PCF take
the least total reduction time among other methods.
From the above figure, this study determines and ranks the
above comparative methods into five ranking: 5 -Excellent, 4-
Very good, 3-Good, 2-Normal and 1-Poor. This study uses a
maximum and minimum value to find an interval value for
ranking those methods.
For a number of test cases, the maximum and minimum
percentage is 56% and 13%. The different between maximum
and minimum value is 43%. An interval value is equal to a
result of dividing the different values by 5. As a result, the
interval value is 8.6. Thus, it can be determined as follows: 5-
Excellent (since 13% to 21.6%), 4-Very good (between 21.6%
and 30.2%), 3-Good (between 30.2% and 38.8%), 2-Normal
(between 38.8% and 47.4%) and 1-Poor (from 47.4% to 56%).
For an ability to reveal faults, the maximum and minimum
percentage is 94% and 37%. The different value is 57%. The
interval value is 11.4. Therefore, it can be determined as
follows: 5-Excellent (since 82.6% to 94%), 4-Very good
(between 71.2% and 82.6%), 3-Good (between 59.8% and
71.2%), 2-Normal (between 48.4% and 59.8%) and 1-Poor
(from 37% to 48.4%).
For a total reduction time, the maximum and minimum
percentage is 64% and 21%. The different between maximum
and minimum value is 43%. An interval value is equal to a
result of dividing the different values by 5. As a result, the
interval value is 8.6. Thus, it can be determined as follows: 5-
Excellent (since 21% to 29.6%), 4-Very good (between 29.6%
38
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
and 38.2%), 3-Good (between 38.2% and 46.8%), 2-Normal
(between 46.8% and 55.4%) and 1-Poor (from 55.4% to 64%).
Therefore, the experiment result of those eight comparative
methods can be shown below:
TABLE IV.
A Comparison of Test Case Reduction Methods
Algorithm
A
Number of
Test Cases /
Size
Abili
tyto
Reveal
Faults
Total
Reduction
Time
Random Deletion
(RD)
2
3
5
Utility Deletion
(UD)
1
3
4
Footprint
Deletion (FD)
5
1
3
Footprint Utility
Deletion (FUD)
2
5
1
Iterative Case
Base Filtering (ICF)
4
3
3
Test Case
Complexity for
Filtering (TCCF)
1
5
1
Test Case Impact
for Filtering (TCIF)
2
5
2
Path Coverage
for Filtering (PCF)
5
1
5
The above result suggests that FD and PCF is perfectly
suitable for a scenario that does not directly concern about an
ability to reveal faults and total reduction time. Both of FD
and PCF are two of the most excellent methods to minimize a
number of test cases. Meanwhile, FUD, TCCF and TCIF are
the most recommended methods to delete tests while
preserving the ability to detect faults. In addition, both of RD
and PCF are excellent in case that total reduction time is
matter.
The above graph presents that both of FD and PCF
minimize a number of test cases by far better than other
reductions methods, approximately over 15%. Meanwhile,
both of them are the worst methods for preserving an ability to
reveal faults. FUD, TCCF and TCIF are best top three
methods to reserve a capability to detect faults. They are
greater than other methods over 22%. Unfortunately, they are
also the worst three methods that require a lot of time during a
reduction process. In the mean time, both of RD and PCF take
the least total reduction time among other methods.
The evaluation result suggests that FD and PCF is
perfectly suitable for a scenario that does not directly concern
about an ability to reveal faults and total reduction time. Both
of FD and PCF are two of the most excellent methods to
minimize a number of test cases. Meanwhile, FUD, TCCF and
TCIF are the most recommended methods to delete tests while
preserving the ability to detect faults. In addition, both of RD
and PCF are excellent in case that total time is matter.
VII. CONCLUSION
This paper reveals that there are many research challenges
and gaps in the test case reduction area. Those challenges and
gaps can give the research direction in this field. However, the
research issues that motivated this study are: a large number of
test cases and a decrease of ability to reveal faults. This paper
combines an approach to maintain CBR and test case
reduction in order to minimize a number of redundancy tests
while maintaining an ability to detect faults.
In this paper, we proposed deletion algorithms to reduce a
number of test cases that generated by widely-used white-box
testing techniques. As part of our research, we conducted an
experiment with 2,000 test cases used in the
telecommunication industry in Thailand. Consequently, our
evaluation result reveal that the proposed method is one of the
most recommendation techniques to maintain ability to reveal
faults and minimize a number of redundancy test cases.
However, the limitation of the proposed techniques is that path
coverage may be not an effective coverage factor for a huge
system that contains million lines of code. This is because it
requires an exhaustive time and cost of identify coverage from
a huge amount of codes.
In future research, we plan to develop deletion algorithms
with other coverage factors, such as function coverage and
block-statement coverage. Also, we aim to implement deletion
algorithms for multiple test suites. Finally, the evaluation
experiment should be conducted for a large commercial
system
REFERENCES
[1] Barry Smyth and Keane. "Remembering To Forget: A
Competence Preserving Deletion Policy for Case-Based
Reasoning Systems." Proceedings of the 14th International Joint
Conference on Artificial Intelligence. Montreal, Quebec,
Canada: Morgan-Kaufman Inc., 1995. 377-382.
[2] Beizer, Boris. Software Testing Techniques. New York, USA:
Van Nostrand Reinhold Inc., 1990.
[3] BO Qu, Changhai Nie, Baowen Xu and Xiaofang Zhang. "Test
Case Prioritization for Black Box Testing." Proceeding with 31st
Annual International Computer Software and Applications
Conference (COMPSAC 2007). Beijing, China, 2007. 465-474.
[4] Boehm, B.W. "A spiral model of software development and
enhancement." IEEE Software Engineering (IEEE Computer
Society), 1988: 61-72.
[5] Daengdej, Jirapun. Adaptable Case Base Reasoning Techniques
for Dealing with Highly Noise Cases. PhD Thesis, The
University of New England, Australia: The University of New
England, 1998.
[6] Gregg Rothermel and Mary Jean Harrold. "A Safe, Efficient
Regression Test Selection Technique." ACM Transactions on
Softw. Eng. And Methodology, 1997: 173-210.
[7] Gregg Rothermel and Mary Jean Harrold. "Analyzing
Regression Test Selection Techniques." IEEE Transactions on
Software Engineering, 1996: 529-511.
[8] Gregg Rothermel, Mary Jean Harrold, Jeffery Ostrin and
Christie Hong. "An Empirical Study of the Effects of
Minimization on the Fault Detection Capabilities of Test
Suites." Proceedings of IEEE International Test Conference on
Software Maintenance (ITCSM'98). Bethesda, Maryland, USA:
IEEE Computer Society, 1998. 33-43.
[9] Gregg Rothermel, Mary Jean Harrold, Jeffery von Ronne and
Christie Hong. "Empirical Studies of Test-Suite Reduction."
39
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
Journal of Software Testing, Verification, and Reliability 12, no.
4 (December 2002): 219-249.
[10] Gregg Rothermel, Roland H. Untch, Chengyun Chu and Mary
Jean Harrold. "Prioritizing Test Cases For Regression Testing."
IEEE Transactions on Software Engineering, 2001: 929-948.
[11] Jefferson Offutt, Jie Pan and Jeffery M. Voas. "Procedures for
Reducing the Size of Coverage-based Test Sets." Proceedings of
the Twelfth International Conference on Testing Computer
Software. Washington D.C, USA, 1995. 111-123.
[12] Jun Zhu and Quiang Yang. "Remembering To Add
Competence-preserving Case Addition Policies for Case Base
Maintenance." Proceedings of the 16th International Joint
Conference in Artificial Intelligence. Stockholm, Sweden :
Morgan Kaufmann Publishers Inc, 1999. 234-241.
[13] Kaner, Cem. "Exploratory Testing." Quality Assurance Institute
Worldwide Annual Software Testing Conference. Orlando,
Florida, USA: Florida Institute of Technology, 2006.
[14] Lehmann, E. and J. Wegener. "Test case design by means of the
CTE XL." Proceedings of the 8th European International
Conference on Software Testing,. Kopenhagen, Denmark: ACM
Press, 2000. 1-10.
[15] Mary Jean Harrold, Rajiv Gupta and Mary Lou Soffa. "A
Methodology for Controlling the Size of A Test Suite." ACM
Transactions on Software Engineering (ACM) 2, no. 3 (July
1993): 270-285.
[16] Nicha Kosindrdecha and Jirapun Daengdej. A Deletion
Algorithm for Case-Based Maintenance Based on Accuracy and
Competence. MS Thesis, Faculty of Science and Technology,
Assumption University, Bangkok, Thailand: Assumption
University, 2003.
[17] Nicha Kosindrdecha and Siripong Roongruangsuwan.
"Reducing Test Cases Created by Path Oriented Test Case
Generation." Proceedings of the AIAA Conference and
Exhibition. Rohnert Park, California, USA: NASA AIAA, 2007.
[18] NIST. The economic impacts of inadequate infrastructure for
software testing. Technical Report, USA: NIST, 2002.
[19] Rothermel, G., R.H. Untch, C. Chu and M.J. Harrold. "Test case
prioritization: An empirical study." Proceedings of the 15th
IEEE International. Oxford, England, UK: IEEE Computer
Society, 1999. 179-188.
[20] S. Elbaum, A. G. Malishevsky and G. Rothermel. "Prioritizing
Test Cases for Regression Testing." Proceedings of the
International Symposium on Software Testing and Analysis.
2000. 102-112.
[21] S. Elbaum, A. Malishevsky, and G. Rothermel. "Test Case
Prioritization: A Family of Empirical Studies." IEEE Trans, on
Software Engineering 28 (February 2002): 159-182.
[22] S. Elbaum, P. Kallakuri, A. G. Malishevsky, G. Rothermel, and
S. Kanduri. "Understanding the effects of changes on the cost-
effectiveness of regression testing techniques." Journal of
Software Testing, Verification, and Reliability 13, no. 2 (June
2003): 65-83.
[23] Saif-ur-Rebman Khan and Aamer Nadeem. "TestFilter: A
Statement-Coverage Based Test Case Reduction Technique."
Proceedings of 10th IEEE International Multitopic Conference.
Islamabad, Pakistan, 2006.
[24] Scott McMaster and Atif Memon. "Call Stack Coverage for GUI
Test-Suite Reduction." Proceedings of the 17th IEEE
International Symposium on Software Reliability Engineering
(ISSRE 2006). North Carolina, USA, 2006.
[25] — . "Call Stack Coverage for Test Suite Reduction." Proceedings
of the 21st IEEE International Conference on Software
Maintenance (ICSM'05). Budapest, Hungary, 2005. 539-548.
[26] — . "Fault Detection Probability Analysis for Coverage-Based
Test Suite Reduction." Proceedings of IEEE International
Conference on Software Maintenance (ICSM 2007). Paris,
France, 2007. 335-344.
[27] Siripong Roongruangsuwan and Jirapun Daengdej. Techniques
for improving case-based maintenance. MS Thesis, Faculty of
Science and Technology, Assumption University, Bangkok,
Thailand: Assumption University, 2003.
[28] Siripong Roongruangsuwan and Jirapun Daengdej. Test case
reduction. Technical Report 25521, Bangkok, Thailand:
Assumption University, 2009.
[29] Smyth, Barry. Case Based Design. PhD Thesis, Department of
Computer Science, Trinity College, Dublin, Ireland: Trinity
College, 1996.
[30] Sprenkle, S., S. Sampath and A. Souter. "An empirical
comparison of test suite reduction techniques for user-session-
based testing of web applications." Journal of Software Test.
Verificat. Reliabil., 2002: 587-596.
[31] Sreedevi Sampath, Sara Sprenkle, Emily Gibson and Lori
Pollock. "Web Application Testing with Customized Test
Requirements - An Experimental Comparison Study."
Proceedings of the 17th International Symposium on Software
Reliability Engineering (ISSRE'06). Raleigh, NC, USA: IEEE
Computer Society, 2006. 266 - 278.
[32] Todd L. Graves, Mary Jean Harrold, Jung-Min Kim, Adam
Porter and Gregg Rothermel. "An Empirical Study of
Regression Test Selection Techniques." ACM Transactions on
Software Engineering and Methodology (TOSEM), 2001: 184-
208.
[33] W. Eric Wong, J. R. Horgan, Saul London and Hira Agrawal.
"A Study of Effective Regression Testing in Practice."
Proceedings of 8th IEEE International Symposium on Software
Reliability Engineering (ISSRE'97). California, USA: IEEE
Computer Society, 1997. 264.
[34] W. Eric Wong, Joseph R. Horgan, Saul London and Aditya P.
Mathur. "Effect of Test Set Minimization on the Fault Detection
Effectiveness of the All-Uses Criterion." Proceedings of the 17th
International Conference on Software Engineering. Seattle,
USA: ACM, 1995. 41-50.
[35] Wilson, David C. A Case-Based Maintenance: The husbandry of
experiences. PhD Thesis, Computer and Science, Indiana
University, USA: Indiana University, 2001.
[36] Xiaofang Zhang, Baowen Xu, Changhai Nie and Liang Shi. "An
Approach for Optimizing Test Suite Based on Testing
Requirement Reduction." Journal of Software (in Chinese),
2007: 821-831.
[37] Xiaofang Zhang, Baowen Xu, Changhai Nie and Liang Shi.
"Test Suite Optimization Based on Testing Requirements
Reduction." International Journal of Electronics & Computer
Science 7, no. 1 (2005): 9-15.
[38] Xue-ying MA, Bin-kui Sheng, Zhen-feng HE and Cheng-qing
YE. "A Genetic Algorithm for Test-Suite Reduction."
Proceedings of IEEE International Conference on Systems, Man
and Cybernetics. Hangzhou, China: ACM Press, 2005. 133-139.
[39] Yanbing Yu, James A. Jones and Mary Jean Harrold. "An
Empirical Study of the Effects of Test-Suite Reduction on Fault
Localization." Proceedings of International Conference on
Software Engineer (ICSE'08). Leipzig, Germany: ACM, 2008.
201-210.
40
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
Nat Traversal for Video Streaming Applications
Omar A. Ibraheem
National Advanced IPv6 Centre of
Excellence (NAv6)
Universiti Sains Malaysia
1 1800, Penang, MALAYSIA
Omer Amer Abouabdalla
National Advanced IPv6 Centre
Excellence (NAv6)
Universiti Sains Malaysia
1 1800, Penang, MALAYSIA
Sureswaran Ramadass
National Advanced IPv6 Centre
Excellence (NAv6)
Universiti Sains Malaysia
1 1800, Penang, MALAYSIA
Abstract — This paper presents a novel method that exploits the
strength features of two streaming protocols (Real-time
Transport Protocol (RTP) and Hypertext Transfer Protocol
(HTTP)) to overcome the Network Address Translation (NAT)
and firewall traversal problem. The proposed solution is able to
bypass the RTP over all kinds of NATs (including symmetric
NATs) by adding extra fields to the RTP/UDP packet at
transport layer in the sender side. The NAT and firewall will
detect these packets as TCP packets on the channel that
initialized the connection. The receiver side will then remove the
extra fields and recover the packets to their original content. The
proposed work involves adding two modules, one at the client
and the other at the video streaming server. The proposed work
also avoids any modification to the NAT or the RTP protocol
itself.
Keywords- NAT Traversal; Video Streaming; RTP; TCP; UDP;
Windows OS.
I. Introduction
Video streaming is considering one of the famous
technologies which is used today. It provides the ability to
playback video files remotely through computer networks.
The demand for this technology is rapidly increasing due to
wide spread of Internet and increasing of the network
bandwidthfl]
There are two main application layer protocols that are used
for video streaming: RTP and HTTP. Although the RTP
protocol is originally developed for video streaming and the
HTTP protocol is originally developed for browsing, which
has many weakness points when dealing with video streaming,
HTTP protocol is still used and more spread than RTP when
used with video streaming. This is due to the simplicity of
using the HTTP protocol and in order to avoid the problems
that faced the RTP protocol.
While HTTP protocol uses one TCP port at the transport
layer, RTP can use many ports. RTP can use UDPs or TCPs
ports at the transport layer depending on how much the packet
path is suffered from packet loss [2]. In low packets loss
environment, the use of RTP over UDP protocol is preferable,
since in media streaming, the small ratio of packets loss better
than packets delay. Hence, the higher reliability of the TCP is
not desired[3]. UDP/RTP has also the multicasting feature and
has the ability to deal with real time communication due to its
features in bandwidth, jitter, reliability and end node's
processing.
RTP/TCP can cause the video streaming to suffer from
discontinuity because the need to reordering,
acknowledgement, and retransmission and the packets,
whereas RTP/UDP can suffer from dropping the packets by
some filters (firewalls) in the Internet Service Provider (ISPs).
Some ISPs drop UDP packets because they are connectionless
hence unfair against TCP traffic. They also need high
processing power and memory to ensure security [4]. But the
main issue that can occur is when using the RTP with the
Network Address Translation (NAT). NAT drops any
RTP/UDP or RTP/TCP packets that are initialized from the
outside (Internet) when incoming to the end-systems (behind
the NAT).
The NAT is a technology that permits many computers on
the same network to share a public Internet Protocol (IP)
address for accessing the Internet. The main reason behind the
wide spread of using the NAT is the limited number of the
available IPv4 addresses [5].
The use of RTP/UDP or RTP/TCP video streaming is
started with a TCP connection that is established by a request
from the client to the server, after initial negotiation using the
RTSP protocol on the same established TCP channel, the
server starts video streaming through UDP or TCP ports
initialized from the server not through the established
RTSP/TCP channel [2].
The NAT permits to pass the outgoing connections requests
produced from a host behind the NAT into the outside
network (like Internet) [6], however it does not permit to pass
any connection request produced from the outside network
(like Internet) to any host behind the NAT [7]. This is because
the translation table entry is constructed only when a client
(behind the NAT) initializes a request to connect to a host on
the outside network (Internet) [8], [9]. If the initialized request
comes from a host outside the network of the NAT into the
inside network, the NAT cannot identify the destination host
for this request and the connection between the outside host
and the inside one cannot be occur [8], [10]. Regarding to the
RTP/UDP video streaming, the NAT will not allow the UDP
video streaming channels to pass to the client behind the NAT,
since the RTP/UDP channels are initially established from the
server (on the Internet).
Considering the RTP weakness points, the HTTP protocol,
is the preferable choice for video streaming. However, HTTP
protocol also has known weakness points: the user can suffer
41
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
from quality reduction and playback discontinuity due to the
probing behaviour of TCP protocol. This can also cause an
oscillating throughput and slow recovery of the packet rate.
In contrast, the UDP protocol provides a mean to keep the
desired sending rate constant. It also keeps streaming smooth
and eliminates the TCP related processing.
This paper presents a novel method to utilize the benefits of
both TCP and UDP. The proposed method enables the video
streaming to traverse the NAT by converting each RTP/UDP
and RTCP/UDP packet into fake TCP packet just before being
sent (at data link layer) by adding a fabricated TCP header
before each UDP video streaming packet and making the
necessary modifications to the length and checksums fields.
These fabricated TCP packets will pass the NAT, since
they will be transmitted on the channel (IP, TCP port) that
firstly initialized (RTSP/TCP channel) by the client behind the
NAT. In this paper, this channel is called the active channel.
The receiver, on the other side has to restore the original
UDP packet before being processed by the corresponding
transport layer. The restoration is based on a specific signature.
In order to restore the packets, every fabricated TCP packet
has to have a known signature. Depending on that signature,
the receiver will restore the original packet. All of the
previous changes are performed at the data link layer.
The rest of this paper is organized as follows: section II,
looks at some related work. In section III, the proposed
methodology and algorithm are presented. In section IV, the
experiments of the implemented proposed method and its
discussions are described. In section V, the evaluation of the
proposed method and comparisons between the proposed
method and the existing technologies are presented. The paper
is concluded in section VI.
II. Related work
Limited to our knowledge, no many similar works are
presented. However, [4] present a method to overcome the
RTP/UDP issues by putting a proxy server between the client
and the streaming server (at the global network). The proxy
receives a HTTP request (TCP) from the client and translates
it to a RTSP/RTP request to the server (TCP+UDP). The
proxy has two different connections (one for the client and the
other for the streaming serve). The main function of the proxy
is to translate the HTTP streaming protocol into RTSP/RTP
streaming protocol. This can overcome the NAT problem due
to that the HTTP request (TCP) is initialized by the client and
the reply will pass through the same TCP port. However a
third device is needed. In addition it is still using the
constraints of the TCP between the proxy and the client (e.g.
retransmission and reordering ...etc) (in addition to the
increase of traffic to the network). Another issue is that there
are too many operations in order to convert a complete
application protocol into another one. Beside, this method
loses the real time property that is needed for end to end
communication because all the packets must be forwarded at
the proxy server.
Vol. 8, No. 6, September 2010
III. Proposed Methodology
In this work, both the client and the server are assumed to
convert all the RTP/UDP streaming packets into fabricated
TCP packets that can be sent to the other side using the active
channel.
This fabrication process which is implemented for
Windows Operating System (OS) requires a full control of the
incoming/outgoing packets. However, there is the issue of
source code of the TCP/IP (non open source for Windows OS)
is not readily accessible and Windows does not allow the
manipulation of the packets in any TCP/IP protocol suite from
level above the TCP/IP driver layer.
To overcome the inaccessibility issue, a hooking technique
is used in order to control the (frame/packet) at the point that
links between the protocol driver and the NIC card(s), which
is represented by the Network Driver Interface Specification
(NDIS).
Hooking is a technique that can convert the calling of one
operating system function into a new one that in turn calls the
old one. The new function can do extra job before moving the
execution to the old one. This can be done without the need
for the source code of the old one [11].
The proposed modules is implemented and run in windows
user mode. When the module can hook the NDIS, it can
monitor, control, add, and modify the NDIS
incoming/outgoing packets easily.
The NDIS-Hooking driver inserts itself between TCP/IP
and all of the adapters that bind with it as shown in figure (1).
User Mode
Kernel Mode
TCP/IP Stack
NDIS
N^l^JiooJkm^ Driver
Network Adapter Deriver
Figure 1 . NDIS hooking driver with relation to user mode 1
When TCP/IP sends a packet, it reaches the NDIS-Hooking
driver (as a frame) before sending to the adapter. Likewise,
packets that are to be indicated (received) on TCP/IP will go
to the NDIS-Hooking driver first.
1 http://www.ntkernel.com/w&p.php?id=7
42
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
The fabricated TCP header is inserted/deleted in the data link
frame, this means that the original RTP/UDP protocol is used
without modification. Nonetheless the fabricated packets can
still bypass the NAT as authenticated ones.
Frame
Header
IP
Header
Fabricated
TCP Header
UDP
Header
RDP Header
-KPayLoad
CRC
Checksum
Figure 2. Proposed frame structure
As these extra bytes (fabricated TCP header) will be added
when the packet is in the data link layer, this may cause the
packet to exceed the Maximum Transfer Unit (MTU) of the
network. Since, no packet must exceed the Maximum Transfer
Unit (MTU) of the network [12], [13], therefore, the sender's
MTU must be decreased by length of the fabricated TCP
header length (20 bytes).
The whole proposed system is composed of two main
modules. The first module resides on the streaming client
while the second resides on the streaming server. Figure (3)
shows the video streaming network topology.
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
the client's and server's streaming ports, then record this
connection's information into an array. This is happened
normally at the setup phase of the RTSP connection. Later
(when the video streaming data is transfered), the client will
check every TCP packet if it contains a specified signature. If
this signature is raised (in the TCP header), this mean that this
TCP packet is fabricated and it contains the original
RTP/UDP packet. The program will remove the TCP header
and recomputed the UDP and IP checksums. All these steps
are done before sending the packet to the rest of TCP/IP
protocol stack
Outgoing packet: If the packet is outgoing to the streaming
server and the outgoing packet was a RTP/UDP packet, then
insert a new fabricated TCP header before the UDP header.
This fabricated TCP header contains the TCP connection
information taken from the appropriate record from an array
containing all streaming connections' details. This TCP
header also contains a specified signature that has to been
recognized from the streaming server in order to return the
packet back to its original RTP/UDP packet. This operation
also needs to recompute the checksums. All these steps are
done before sending the packet to the adapter. Figure 4 shows
the flowchart of client side module.
Client NAT
Figure 3. Video streaming network topology
Each module consists of the following components:
A component (hooking function in Fig. 1) that provides a
way to access the frame at the data link layer. This component
accesses the frames in data link layer which is in the kernel
mode and moves it into the user mode and vice versa.
A component that finds the required frame based on its
content. This component extracts the specified packets from
the frames which have to be changed (fabricated/restored)
depending on sending direction (income/outcome).
A component that makes the required modifications
(fabricating/restoring) to the predetermined packets. This
component changes the predetermined packets depending the
sending direction (send/receive). In sending, the component
changes the RTP/UDP packet into fabricated TCP packet. In
receiving, the component restores the fabricated TCP packet
into its original RTP/UDP content. This component also re-
computes length and checksums.
A. Client Side Module
As mentioned earlier, the module has to access the kernel
(at data link layer). This is done by accessing the NDIS driver.
The module listens until a packet event has occurred. There
are two possible scenarios:
Incoming packet: If the packet is coming from the
streaming server, then the program will look for the TCP that
contains an RTSP packet. If this RTSP packet contains both
Access the Data
link layer (in
kernel mode)
Figure 4. Flowchart of the client side module
43
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
B. Server Side Module
In server side module, similar steps to the client are also
implemented. The difference is that the system gets the RTSP
connection's details from the outgoing TCP packet instead of
incoming TCP packet in the client. Figure (5) shows the
flowchart of the main steps of the server module.
Return back to it;
original UDP
packet
o
I
Recompute the
"
Move the packet
to the kernel mode
Send packet to
the upper
layer
Figure 5. Flowchart of the server side module
IV. Experiments and discussions
A. Experiments Setup
In this experiment, we use three PCs running windows XP.
Two PCs with one LAN card (client and the server). The other
PC (working as a NAT) contains two LAN cards.
RedMug streaming server commercial software is used on
the server site. The VLC media player (version 1.0.5) is used
on the client side. The VLC media player is set to use the
RTSP protocol by giving a URL of one movie on the
streaming server. The proposed method (client and server
modules) is implemented in VC++.Net Framework and it is
running in windows OS environment in user mode. A
windows device driver (Windows Packet Filter Kit
"winpkfilter 3.0" from NT Kernel Resources,
http://www.ntkernel.com) is used for the hooking purpose.
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
B. Experimental Results and Discussion
In the first experiment (before using the proposed method),
the client tries to access the movie on the streaming server
using the above system configuration. The connection's
establishment and the video streaming negotiations between
the client and the server are established normally. However,
the connection fails at the stage of data streaming
transformation (see Fig. 6).
User- Agent: Ub VLC ,1.1.2 (LIVE555 Streaming Media v20 10.03. 16)
RTSP Session identifier: 5. Sent to client: "RTSP/ 1.0 200 OK
RTSP Session identifier: 5. Received RTSP message from 150.150.100.110:02452. Message: 'DESCRIBE rtsp://ll
User- Agent: Lib VLC/ 1.1.2 (LIVE555 Streaming Media v20 10.03. 16)
RTSPSession identifier: 5. Is the "C:\red MUG \ Contents, sample. mp 4" MP4file in cache? 'true']
RTSPSession identifier: 5. Video track name to stream: track ID/ 2, audio track name to stream: trackID/1]
RTSPSession identifier: 5. Sent to client: 'RTSP/ 1.0 200 OK
RTSPSession identifier: 5. Received RTSP message from 150. 150. 100.1 10:62452. Message: 'SETUP rtsp:// 150. 1
User- Agent: Lib VLC/ 1.1.2 (LIVE555 Streaming Media v20 10.03. 16)
RTSPSession identifier: 5. Sent to client: 'RTSP/ 1.0 200 OK
Session: 257655783;timeout=360
RTSPSession identifier: 5. Received RTSP message from 150. 150. 100.1 10:62452. Message: 'SETUP rtsp: ,, 150. 1
Session: 257655783
User- Agent: Lib VLC/ 1.1.2 (LIVE555 Streaming Media v20 10.03. 16)
RTSPSession identifier: 5. Sent to client: 'RTSP/ 1.0 200 OK
Session: 257655783;timeout=360
RTSPSession identifier: 5. Received RTSP message from 150.150.100.110:62452. Message: 'PLAY rtsp://150. 15
Session: 257655783
User- Agent: Lib VLC/ 1.1.2 (LIVE555 Streaming Media v20 10.03. 16)
RTSPSession identifier: 5. Called the seek method. Start time in sees: 0.000000.]
RTSPSession identifier: 5. Sent to client: 'RTSP/ 1.0 200 OK
Session: 257655783;timeout=360
Client 6 is opened
RTSPSession identifier: 5. Received RTSP message from 150. 150. 100.1 10:62452. Message: 'TEARDOWN rtsp://
Session: 257655783
User- Agent: Lib VLC/ 1.1.2 (LIVE555 Streaming Media v20 10.03. 16)
RTSPSession identifier: 5. Sent to client: 'RTSP/ 1.0 200 OK
RTSPSession identifier: 5. Connection from 150.150.100.110:62452 closed.]
RTSPSession identifier: 6. Connection arrived from 150.150.100.110:62453.]
RTSPSession identifier: 6. Received RTSP message from 150. 150. 100.1 10:62453. Message: 'OPTIONS rtsp:// 15(
User- Agent: Lib VLC/ 1.1.2 (LIVE555 Streaming Media v20 10.03. 16)
RTSPSession identifier: 6. Sent to client: 'RTSP/ 1.0 200 OK
RTSPSession identifier: 6. Received RTSP message from 150.150.100.110:62453. Message: 'DESCRIBE rtsp://ll
User- Agent: Lib VLC/ 1.1.2 (LIVE555 Streaming Media v20 10.03. 16)
RTSPSession identifier: 6. Is the "C:, red MUG ■.Contents, sample. mp4" MP 4 file in cache? "true"]
RTSPSession identifier: 6. Video track name to stream: trackID/2, audio track name to stream: tracklD/lj
RTSPSession identifier: 6. Sent to client: 'RTSP/ 1.0 200 OK
RTSPSession identifier: 6. Received RTSP message from 150. 150. 100.1 10:62453. Message: 'SETUP rtsp: ,, 150. 1
User- Agent: Lib VLC/ 1.1.2 (LIVE555 Streaming Media v20 10.03. 16)
RTSPSession identifier: 6. Sent to client: 'RTSP/ 1.0 500 Internal Server Error
RTSPSession identifier: 6. Connection from 150.150.100.110:62453 closed.]
Figure 6. Connection breakdown when data streaming transforming began
(server side)
The reason for the success of the initialization of the client-
server connection and all the negotiations needed to transfer
the video streaming are that the connection request is a TCP
and the initialization is coming from the client (behind the
NAT) and the video streaming negotiations are done by the
RTSP that uses the active channel. However, the client could
not receive the video streaming data since the NAT dropped
the RTP/UDP video streaming packets. The client then sends
a teardown command to inform the server that the negotiation
is over. The client starts one additional negotiation tries before
it close the connection.
In the second experiment, we used the proposed client and
server modules. After running, the two modules start
monitoring the data link frames. The client monitors the
outgoing streaming request while the server monitors the
incoming streaming request.
When the client request a video streaming from the server,
The connection's establishment and the video streaming
negotiations between the client and the server are established
normally and the client started to display the video streaming
data as shown if figure (7A and 7B).
44
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
uient n is opened
RTSP Session identifier: 9. Connection arrived from 130.130.100.110:62488.]
RTSP Session identifier: 9. Received RTSP message from 150.150.100.110:62488. Message: "OPTIONS rtsp://15i
User- Agent: UbVLC, 1.1.2 (LIVE5S5 Streaming Media V 2010.03.16)
RTSP Session identifier: 9. Sent to client: 'RTSP, 1.0 200 OK
RTSP Session identifier: 9. Received RTSP message from 150.150.100.110:62488. Message: 'DESCRIBE rtsp:, ,, 1"
User- Agent: LibVLC/1.1.2 (LIVE 5 55 Streaming Media v20 10.03.16)
RTSP Session identifier: 9. Is the 'C:\redMUG\Contients\sample.mp4' MP4file in cache? 'true']
RTSP Session identifier: 0. Video track name to stream: track ID/ 2, audio track name to stream: track ID/ 1]
RTSP Session identifier: 0. Sent to client: "RTSP/ 1.0 200 OK
RTS P Session identifier: 0. Received RTSP message from 150.150.100.110:62488. Message: 'SETUP rtsp:// 150.1
User- Agent: UbVLC, 1.1.2 (LIVE 5 55 Streaming Media v2010.03.16)
RTSP Session identifier: 9. Sent to client: 'RTSP/1.0 200OK
Session: 247776867;tjmeout=360
RTSP Session identifier: 0. Received RTSP message from 150.150.100.110:62488. Message: 'SETUP rtsp:// 150.1
Session: 247776867
User- Agent: UbVLC: 1. 1.2 (LIVE555 Streaming Media v2010.03.16)
RTSP Session identifier: 0. Sent to client: "RTSP/ 1.0 200 OK
Session: 247776867;tjmeout=360
RTS P Session identifier: 0. Received RTSP message from 150.150.100.110:62488. Message: 'PLAY rtsp:// 150. 15
Session: 247776867
User- Agent: UbVLC, 1.1.2 (LIVE555 Streaming Media v2010.03.16)
RTSPSession identifier: 0. Called the seek method. Start time in sees: 0.000000.]
RTSP Session identifier: 0. Sent to client: 'RTSP/ 1.0 200 OK
Session: 247776867;timeout=360
Figure 7A. Connection still active when the data streaming are transforming
(server side)
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
Internet gateway. Streaming might fail at times even if the
gateway has a built-in RTSP NAT.
Reference [4] utilizes the two streaming protocols
separately by using a third device (proxy) between the client
and server (every side with whole streaming protocol
advantages and disadvantages), the proposed method utilizes
the benefits of the RTP and HTTP protocols without using any
extra device.
V
'^r
fjf /. m
*-'r^-t
ft,
4
.■■
ME
Figure 7B. Video streaming is displayed in the client (behind the
NAT)
When negotiation is started, the host records the connection
details: IP, TCP port and the streaming UDP ports. The host
will insert the fabricated TCP header (after the UDP header)
in the video streaming packet before sending it.
The reason for the success of transforming the streaming
data is that the sending host converts each streaming UDP
packet into a fabricated TCP packet that bypasses the NAT
because it uses the active channel. The receiving host in turn
restores the fabricated TCP packet into the UDP streaming
data at the data link layer before sending it to the upper layer.
V. Evaluation
A comparison between our proposed method and the
existing technologies is presented in Table 1. The proposed
method has several advantages over the existing technologies,
although the new packet size is 20 bytes larger than the
normal RTP/UDP packet, but less compared with the HTTP.
This has a little impact on the network performance.
The proposed method can traverse the video streaming over
all types of NAT. It can also traverse the firewall that blocks
the UDP ports that RTP may use, commonly with home
Table I. CURRENT AND PROPOSED METHOD COMPARISON
FEATURE
HTTP
RTP/TCP
RTP/UDP
PROPOSED
METHOD
Directional
Bidirectional
Bidirectional
Uniary
Uniary
Playback
hiccups
Yes
Yes
No
No
Quality
Reductions
Yes
Yes
No
No
Oscillating
throughput
Yes
Yes
No
No
Slow recovery
Yes
Yes
No
No
ISP firewall
Traverse
Traverse
Blocked
Traverse
NAT traversal
Yes
No
No
Yes
End-to-End
Delay
Long
Long
Short
Short
Window buffer
and reordering
Yes
Yes
No
No
Streaming
method
Downloading
or progressive
Streaming
Streaming
Streaming
VI. Conclusion
The two main transport layer protocols: TCP and UDP can
be used in streaming but with the whole advantages and
disadvantages of using that protocol. In this paper, a new
method is presented and implemented that can merge some
advantages of both protocols. It enables the client and server
to use UDP advantages in each side for streaming. Both client
and server gains scalability by not having to deal with some
TCP processing feature (e.g. Acknowledgement and window
buffering ...etc). In the other hand, utilize the benefit of the
TCP advantages to traverse the NAT and the firewall. In other
words, the stream is not discarded and traverses the NAT and
the firewall. The experimental results show that the new
method achieves the firewall traversal and Nat traversal even
with the most difficult NAT (symmetric NAT).
References
[1] Chu-Hsing, L., et al., Energy Analysis of Multimedia Video Streaming
on Mobile Devices, in Proceedings of the 3rd International Conference
and Workshops on Advances in Information Security and Assurance.
2009, Springer- Verlag: Seoul, Korea.
[2] Matthew Syme, P.G., Optimizing Network Performance with Content
Switching: Server, Firewall and Cache Load Balancing. 1st ed. 2004:
Prentice Hall. 288.
[3] Philip, W.F., et al., Server- efficient high-definition media dissemination,
in Proceedings of the 18th international workshop on Network and
operating systems support for digital audio and video. 2009, ACM:
Williamsburg, VA, USA.
45
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
[4] H, et al., Transparent protocol translation for streaming, in
Proceedings of the 15th international conference on Multimedia. 2007,
ACM: Augsburg, Germany.
[5] Sourour, M., B. Adel, and A. Tarek. Security Implications of Network
Address Translation on Intrusion Detection and Prevention Systems, in
Network and Service Security, 2009. N2S '09. International Conference
on. 2009.
[6] Sanmin, L., et al. TCPBridge: A software approach to establish direct
communications for NAT hosts, in Computer Systems and Applications,
2008. AICCSA2008. IEEE/ACS International Conference on. 2008.
[7] Yeryomin, Y., F. Evers, and J. Seitz. Solving the firewall and NAT
traversal issues for SIP -based VoIP, in Telecommunications, 2008.
ICT 2008. International Conference on. 2008.
[8] Takabatake, T. A Scheme of Relay Server Selection Methods for NAT
Traversal through Home Gateways, in Telecommunication Networks
and Applications Conference, 2008. ATNAC 2008. Australasian. 2008.
[9] P.Srisurensh, K.E., Traditional IP Network Address Translator
(Traditional NAT). RFC 3022, 2001. IETF.
[10] Khlifi, H., J.C. Gregoire, and J. Phillips, VoIP and NAT /firewalls:
issues, traversal techniques, and a real-world solution.
Communications Magazine, IEEE, 2006. 44(7): p. 93-99.
[11] Ivanov, I., API Hooking Revealed. 2002.
[12] Hasegawa, T. and T. Ogishi. A framework on gigabit rate packet
header collection for low-cost Internet monitoring system, in
Communications, 2002. ICC 2002. IEEE International Conference on.
2002.
[13] Jelger, C.S. and J.M.H. Elmirghani. Performance of a slotted MAC
protocol for WDM metropolitan access ring networks under self-
similar traffic, in Communications, 2002. ICC 2002. IEEE
International Conference on. 2002.
AUTHORS PROFILE
Wi
Omar A. Ibraheem (PhD) is currently a
Post Doctoral Research Fellow in National
Advanced IPv6 Centre of Excellence
(NAv6) at Universiti Sains Malaysia
(USM). Dr. Omar obtained his bachelor,
master, and doctorate in computer science
from Mosul University in 1998, 2000, and
2006 respectively. He has joined NAv6
since January 20 10. Before that, Dr. Omar
was a senior lecturer at the computer
science department, College of Computer
Science and Mathematics of Mosul
University in Iraq. His research area include
the Network protocols, Routing protocols,
Network security and Multimedia
communications.
Omer Amer Abouabdallah (PhD) is a
senior lecturer and post graduate
coordinator of the National Advanced IPv6
Centre of Excellence (NAV6) in Universiti
Sains Malaysia. Dr. Omar obtained his
Bachelor of Computer Science from Al-
Fateh University, Tripoli, Libya in 1993.
He obtained his Master of Computer
Science and doctorate from Universiti Sains
Malaysia in 1999 and 2004 respectively.
Dr. Omar is heavily involved in researches
carried by NAv6 centre, such as IPv6 over
Fiber Object and the Multimedia
Conferencing System. The highlights of Dr.
Omar's achievements include the winner of
Sanggar Sanjung Award 2005 and 2006 by
USM on 2007 and the winner of Innovative
Product Award for NAT AND FIREWALL
TRAVERSAL SOLUTION for MSCv6 as
well as the gold medal winner in
International Invention Innovation
Industrial Design & Technology Exhibition
2006 (ITEX2006).
Sureswaran Ramadass (PhD) is a
Professor and the Director of the National
Advanced IPv6 Centre (NAv6) at
Universiti Sains Malaysia (USM). He is
also the founder of Mlabs Systems Berhad
(MLABS), a public listed company on the
MESDAQ. Prof Dr Sureswaran obtained
his BsEE/CE (Magna Cum Laude) and
Masters in Electrical and Computer
Engineering from the University of Miami
in 1987 and 1990 respectively. He
obtained his doctorate from USM in 2000
while serving as a full time faculty in the
School of Computer Sciences. His
research areas include the Multimedia
Conferencing System, Distributed Systems
and Network Entities, Real Time
Enterprise Network Monitoring, Real
Time Enterprise System Security, Satellite
and Wireless Networks, IPv6 Research,
Development and Consultancy, and
Digital Library Systems.
46
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) InternationalJournal of Computer Science and Information Security,
Vol 8, No. 6, September 2010
The Integration of GPS Navigator Device with
Vehicle Tracking System for Rental Car Firms
Omarah Omar Alharaki
Faculty of Information and
Communication Technology,
International Islamic University
Malaysia,
Kuala Lumpur, Malaysia.
Fahad Saleh Alaieri
Faculty of Information and
Communication Technology,
International Islamic University
Malaysia,
Kuala Lumpur, Malaysia.
Akram M. Zeki
Faculty of Information and
Communication Technology,
International Islamic University
Malaysia,
Kuala Lumpur, Malaysia. .
Abstract — the aim of this research is to integrate the GPS
tracking system (tracking device and web-based application)
with GPS navigator for rental cars, allowing the company to
use various applications to monitor and manage the cars.
This is enable the firms and customers to communicate with
each other via the GPS navigator. The system should be
developed by applying new features in GPS tracking
application devices in vehicles. This paper also proposes new
features that can be applied to the GPS Navigator. It also
shows the benefits that the customers and staff will get from
this system.
Keywords', GPS tracking system, GPS devices, GPRS, Garmin.
I. Introduction
The Global Positioning System (GPS) is a satellite-
based navigation system made up of a network of 24
satellites to specify a position on the surface of the earth
[11] [4]. It also provides highly accurate location with the
use of special GPS receivers and their augmentations
[18]. In 1973 GPS was intended for USA military
applications [7], but in the 1980s, the government made
the system available for civilian use. GPS works in any
weather conditions, anywhere in the world, 24 hours a
day. There are no subscription fees or setup charges to
use GPS [11].
From NASA (2009) : The uses of GPS have extended
to include both the commercial and scientific worlds.
Commercially, GPS is used as a navigation and
positioning tool in airplanes, boats, cars, and for almost
all outdoor recreational activities such as hiking, fishing,
and kayaking [7]. GPS is also playing an increasing role
in the tracking of motor vehicles [2].
General Packet Radio Service (GPRS) is an
enhancement of GSM networks to support packet
switched data services such as email and web browser in
addition to existing GSM data services such as Short
Message Service (SMS). GPRS operates on the existing
GSM network infrastructure that it utilizes available time
slots during each frame transmission. Thus, it does not
overload the existing GSM network traffic and can
efficiently provide data services. The GPRS can transfer
data at the maximum rate of 115.2 kbps. Due to a very
large coverage area of GSM networks around the world,
GPRS becomes the largest data service network available
and always-on; thus, it is most suitable for a real-time
tracking management system. [1].
GPS tracking system developed that transmit
vehicle's data in real time via cellular or satellite
networks to a remote computer or data centre [15] [10].
Vehicle tracking system signifies the monitoring and
management of vehicle, trucks, etc by using GPS system
that can get in real time the current location, situation ,
history, performance and emissions and control them. [1]
[16].
The web based tracking system allows users to
securely log in and track their cars in real- time over the
Internet. The user sees moving dots on a map in a web
browser. It also display all transmitted information to the
users along with displaying location of vehicle on a map
[15].
The tracking system allows users to locate any car at
any time of day. Also, it can replay a past trace of the cars
history. It can remotely control the car such as run alarms
and locking devices. It enable the users to keep track of
the vehicles without the intervention of the driver where,
as navigation system helps the driver to reach the
destination [12].
Many shipment companies in the world use GPS
tracking systems in their trucks. It is very important for
the fleets, especially when they have big number of
trucks and staffs, to manage this huge number of vehicles
and people. In addition, The shipments and conveying the
tracking information to customers are perceived to be
important customer service components and they are
often considered industry norms rather than a potential
competitive advantage for shipment service providers
[14].
47
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) InternationalJournal of Computer Science and Information Security,
Vol 8, No. 6, September 2010
However, the biggest challenge in the cars rental
companies are car thieves and delay car return. This will
cause, lost a lot of money and cars. However, the rental
cars firms tried to find the solution to save their cars from
these troubles from their clients and staff.
The vehicle tracking system presented in this paper as
a system that is designed to track and manage rental cars
that are used by the customer, using a GPS tracking
technology with GPS navigator.
The system comprises of vehicle GPS tracking
devices (Tracking device and GPS navigator), GPRS and
a web-based application. Through this system, the
company staff will have the facility to monitor the
movement and relevant information of each vehicle.
Moreover, to keep in touch with their customers by GPS
navigator technology [13].
The paper highlights how to apply new application in
GPS navigator that allows the firm staff and customers to
communicate between each other via GPS navigator in
the rental cars. Moreover, satisfying customer needs and
building high confidence between the firm and
customers.
This paper illustrates the integration of multiple
technologies to achieve a common goal. It shows how
such technologies can be synergistically combined to
address real rental cars firm problems [13].
II. GPS TRACKING SYSTEM
GPS device is a small unit that receive signals from
satellites and send other signals to antennas (GPRS). This
device is a major part of the system and it will be
installed into the vehicle which is responsible for
capturing the following information for the vehicle such
as the Current location of vehicle, Speed of vehicle, Door
open/close status, Ignition on/off status, etc. [15].
This device is also responsible for transmitting this
information to the Tracking Server located anywhere in
the world. Also, it has to install the unit in a hidden and
safe place inside the vehicle [15].
The information about the vehicle saved in this unit
will be sent to antennas by GPRS, and there are many
applications (Figure 1 shows one of the application)
connected to the internet that can calculate it and put it on
the map to integrate with it.
(Figure 1 : web based tracking application)
The information includes location, speed, fuel level,
engine situation, start driving point, and car situation.
Also via this application, the company staff can control
the cars, for example, on/off lights, on/off air
conditioning, on/off car engine, on/off security system
and others, by sending some codes to the tracker in the
car.
A. GPS Tracking System Framework
The GPS tracking system consists of client-server
architecture where the web browser is the client and the
server functions are shared between a web server, a
communication server, a database server and a map
server [13]. The process of web application tracking
consists of four parts: a location-aware web system,
location determination, location-dependent content query
and personalized presentation. The location-aware web
system is the underlying infrastructure. It allows the
exchange of the location information between the web
browser and the web server, and the automatic update of
the web pages in the web browser when the location
changes [17].
GPRS is the main method of communication between
the tracking device and the web server. GPRS, being a
2.5G mobile technology, is available all over the world. It
is also ideally suitable for data transfer over an always
on-line connection between a central location and mobile
devices. The cost is per kilobyte of data transferred, in
comparison to SMS where the cost is per message. [13].
The location information collected through the GPS in
real time is placed in a central database that is owned by
the firm's staff via GPRS antenna. Each user of the
system may access this information via the Internet [15].
Figure 2 shows the system framework.
48
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) InternationalJournal of Computer Science and Information Security,
Vol 8, No. 6, September 2010
(Figure 2 : GPS tracking system framework)
B. Features of GPS Tracking System:
Some rental car firms offer GPS devices for rent when a
customer comes to rent a car by adding charges for it per
day. This is beneficial for customers to locate their rental
cars, know the location direction and their current
location. Moreover, it also helps the rental car firms to be
informed how the customers are using the cars.
This system is also used to prevent theft and retrieve
stolen/lost vehicles. The signal sent out by the installed
device help the rental car firms to track the vehicle. These
tracking systems can be used as an alternative for
traditional car alarms or in combination of it. Installing
tracking systems can thus bring down the insurance costs
for these vehicles by reducing the risk factor [10].
Moreover, this system gives benefits to the rental car
firms and to their customers such as: [9].
• Real-time location of vehicles,
• Historical vehicle reports,
• Security code/pin,
• Trip computer,
• Engine idle and start/stop reporting,
• Real-time tracking alerts and reports after hour vehicle
monitoring,
• Create custom Geo-fences and landmarks,
• Historical movement of vehicle,
• Mileage reporting,
• Optional starter disable/enable,
• Notification of doors opening and closing
C. Case Study:
Some fleet firms that applied this tracking system which
are the following:
Delhi Transport Corporation is the one of the largest City
Road Transport Undertaking in India [3]. It has a fleet of
around 15,000 vehicles carrying on the business of
passenger transport in 800 routes from 33 depots all over
the state of Delhi with a product mix comprising of City
and Inter-city services. After they implemented the
tracking system in vehicles and other functions under
Automatic Fleet Management System, the benefits they
obtained are [3]:
• Better bus scheduling.
• Quick replacement in case of breakdown/accident en-
route.
• Effective control over the drivers & checking bus-
stops skipping.
• Check on over-speeding.
• Basic communication between driver and control
room in emergencies.
• Automation of Fleet Operations minimizing human
intervention.
• Improved Fleet utilization leading to better services
and thereby enhancing commuter satisfaction.
Another car rental Firm is in the United Arab Emirates
(UAE) that fitted their cars with high-tech C track
GPS/GPRS satellite tracking units to prevent thieves from
driving away with the cars [8].
Thieves in Dubai and Sharjah are increasingly posing
as clients who rent a car and then ship the car out of the
country, mostly to Russia, North Africa and Eastern
Europe. This prompted car rental firms to integrate
vehicle with GPS tracking system to protect their
business and save themselves time and money. Al
Mumtaz Rent-a-Car in Dubai gives their vehicles some
level of safety and it can monitor their vehicle all the
time. [8].
The system is cost-effective, as it saves time and
money in the recovering of stolen vehicles. According to
Diamond Lease, they can also monitor the movement of
all their vehicles and can thus establish whether a vehicle
is being misused. The GPS tracking device can keep track
of the car's engine condition by recording harsh braking,
speeding and even the removal of any of the car's parts,
thus saving us more money [8].
In Saudi Arabia from rental cars company report they
have more than 300 cars and their experience more than
25 years in this field. Moreover, they have more than 10
branches in East of Saudi Arabia. However, during this
time the rental cars faced many troubles from either
customers or employees such as car lost, personal car use
from the staff and delays car return. From these problems
the rental cars lose money, lose cars, low service quality,
and short cars use period, [one of authors experience]
On the other hand, after they installed the tracking
system into the rental cars they save almost all of their
money, time and cars. Moreover, the firm got
improvement of their services quality and maintained the
cars quality for a longer period.
III. Proposed method and Recommendation
This paper shows that this system is easy to apply,
very important to manage, save and control the cars. In
49
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) InternationalJournal of Computer Science and Information Security,
Vol 8, No. 6, September 2010
this section, the paper will propose some helpful and
useful features in the GPS system.
There are some new features recommended to be
implement in the system that will save Firm time,
money, and manage their cars safely which are as
following (Figure 3):
Integrate GPS navigator device with
tracking system
i GPS
Create new application system for GPS
navigator
Connect car electricity with GPS
navigator device
Connect GPS navigator device with web based
application
Figure 3 : The process of applying GPS navigator system
1) Integrate GPS navigator device with GPS tracking
system
This device usually work as a guide for travelers to show
them where they are , where they want to go , maps ,
roads, shops and other useful information .
The integration framework is between GPS navigator
device and GPS tracking system as shown in figure 4.
(Figure 4: framework for the integration of GPS devices)
2) Create new application system for GPS navigator
that can link customers with rental car company via GPS
tracking device. This system will contain new menus that
provide new services for customers in addition to its main
function as guidance. The proposed new menu can help
the customers to contact with staff for any help such as
lost directions, lost key or any other assistants as shown
in figure 5.
LOST
KEY
LOST
DIRECTION
I NEED
HELP
(Figure 5: Proposed new menu in GPS nav. Application)
GPS navigator device must be installed in the cars
because it will link with the GPS tracking device.
3) Connect car electricity with GPS navigator device
that controls the cars, so that nobody can drive the car
without logging in using the specific username and
password. When the customer wants to drive the car after
he/she has signed the contract, he/she has to enter the
name that is in the contract and the contract number as
shown in figure 6 and figure 7.
J Enter Your Name please
< >
► ^
A
H
B C
■ 1 '
D E F f G
i
J
K
L
M
N
P Q R S T U
V
w
x y z „ a
Back Mode 12 3 Done
(Figure 6: field for entering customer name)
jEnter your contract number 1
1 1 2 1 3
F
n
#
*
%
T
5 I 6
&
{
)
*
7
8 I 9
+
1
/
■_i
f
?
@
(Figure 7: field for entering contract number)
4) connect GPS navigator device with web based
application through GPS tracking device that can send
messages from the company to the customer, such as,
reminders of expiry date of contract, last promotions, and
other useful information (as shown in figure 8 & 9) .
50
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) InternationalJournal of Computer Science and Information Security,
Vol 8, No. 6, September 2010
H
Dear Mr./Ms. "Customer Name"
Your contract has almost expired, you are
required to return the car within 5 hours or if
you wish to extend the contract kindly contact
us immediately.
Phone
Volume Tools
(Figure 8: one of the proposed reminder messages from
the rental car firm to their customers)
Dear Mr./Ms. "Customer Name"
There is construction in
"Street Name" and "Street Name"
For your safety, kindly avoid driving there at
this time.
®O0
Phone Volume
Tools
(Figure 9: one of the proposed notice messages from the
rental car firm to their customers)
IV. Conclusion
In conclusion, this paper presents the development of
a Rental cars tracking system using GPS & GPRS
technologies. It is a typical example of how the
advantages may be forced for the efficient and effective
managing of rental cars firm. However, after
implementation of this proposed system may give
benefits for the rental cars firm customers such as built in
road maps, Route capability Touch screen access,
monitor fuel consumption, vehicle maintenance alerts,
route guidance, Speed limit display [6], last promotions
from the firm, warning reminders, renewal car rental
contract, and keep contact with the rental cars firm.
Moreover, this system show that the customers will
connected with the rental cars firm in whole day via GPS
navigator device with full protection of the rental cars
firms from any violations from the customers.
Acknowledgment
We would like to acknowledge our mentor, Dr. Akram
zeki for his patience, guidance, and knowledge in order to
reach our goal. Also, we would like to convey our
gratitude to our parents who support us to understand the
importance of knowledge and show us the best way to
achieve it.
References
[1] Chadil, N.; Russameesawang, A.; Keeratiwintakorn, P. (2008),
"Real-time tracking management system using GPS, GPRS and
Google earth", proceeding of ECTICON.2008, pp.393 - 396.
[2] Craig A. Scott (1994), "Improved GPS Positioning for Motor
Vehicles Through Map, Matching", University of
Technology, Sydney, Presented at ION-94, Salt Palace
Convention Center, Salt Lake City.
[3] DELHI Transportation Corporation case study (2007),
http : //www. cmcltd. com/casestudies/transportation/GPSSy stem
_Case_Study_DTC.pdf
[4] General Information Document of GPS-based Fleet Management
and Tracking Systems (2008), Exaterra Inc., Ottawa, Canada:
Canadian company based in Ottawa.
[5] Hariharan, Krumm, Horvitz (2005), "Web-Enhanced GPS",
School of Information and Computer Sciences, University of
California, Irvine, USA.
[6] http : //gpstrackit. com/gps-tracking-products/garmin-integration
[7] http://scign.jpl.nasa.gov/learn/gpsl .htm
[8] http://www.digicore.com/ , (UAE Press Release August,2007)
[9] http://www.imardainc.com/smarttrack-vehicle-tracking-system
[10] http://www.roseindia.net/technology/vehicle-
tracking/VehicleTrackingSy stems, shtml
[11] http : //www8 .garmin. com/aboutGPS
[12] http://www.gisdevelopment.net/technology/gps/techgp0044.htm
[13] M. Medagama, D. Gamage, L. Wijesinghe, N. Leelaratna, I.
Karunaratne and D. Dias (2008), GIS/GPS/GPRS and Web based
Framework for Vehicle Fleet Tracking, ENGINEER , No. 05, pp.
28-33.
[14] Mikko Ka'rkka'inen, Timo Ala-Risku, Kary Fra'mling (2004),
"Efficient tracking for short-term multi-company networks",
International Journal of Physical Distribution & Logistics
Management Vol. 34 No. 7, pp. 545-564
[15] Muruganandham , P.R.Mukesh (2010), " Real Time Web based
Vehicle Tracking using GPS", World Academy of Science,
Engineering and Technology 61 2010.
[16] Robert W. Bogue (2004), "New on-vehicle performance and
emission monitoring system", Sensor Review, Volume 24, No. 4,
pp. 358-360
[17] Rui Zhou (2008), "Enable web-based tracking and guiding by
integrating location-awareness with the world wide web",
Campus-Wide Information Systems Vol. 25 No. 5, pp. 311-328
[18] S.S.S. Prakash, MadhavN. Kulkarni (2003), " Fleet
Management: A GPS-GIS integrated approach", Department of
Civil Engineering, IIT Bombay, Mumbai, GISdevelopment.net.
51
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, 2010
Process Framework in Global extreme Programming
Ridi Ferdiana, Lukito Edi Nugroho, Paulus Insap
Santoso
Department of Electrical Engineering and Information
Technology
Gadjah Mada University (UGM)
Yogyakarta, Indonesia
Ahmad Ashari
Department of Computer Science and Electronics
Gadjah Mada University (UGM)
Yogyakarta, Indonesia .
Abstract — Software development life cycle works as a process
framework that underlying of the software engineering
framework. In multi-site software development, the impact of not
having a process framework is often quite disastrous. The multi-
site team can work anonymous and have no guidance to work
and collaborate with others. Therefore, several researches have
begun to introduce the important of the process framework
through a legitimate software development life cycle. The most
common process framework that introduced in multi-site
software development is called as a Global Software Development
process framework (GSD). However, many GSD
implementations are reported as an enterprise process
framework that highly iterative and model driven oriented. This
paper will show an alternative way to modify the existing GSD
SDLC into the simplified process framework by integrating the
process framework with an agile method like extreme
Programming. By simplifying the process framework, it will
provide an opportunity for small and medium enterprise to adopt
the proposed SDLC in multi-site development.
Keywords-process fram ework; software
lifecycle; agile; extreme Programming
development
I.
Introduction
Global Software development (GSD) is defined as a
software development process that uses teams from multiple
geographic locations. The physical distant between team is the
key issue in GSD, The team may be from the same
organization, collaboration that involves different organization,
outsourcing to the other organization. The dispersed can
happen in same country or other sides of the world. Developing
software in the distant introduces complex and interesting
issues.
Further research tells us that distributed is done in order to
fulfill well-establish motivators include [3].
• Limited trained workforce in technologies that are
required to build today's complex systems.
• Differences in development costs favoring dispersing
team geographically.
• A "round-clock" work system facilitated by time zone
differentials allowing for shorter times to market.
• Advanced infrastructures (internet bandwidth, and
collaboration software).
• A desire to be close to a local market.
Although the motivations behind GSD often differ from the
other, the main problem is still constant. It is naturally more
difficult to coordinate projects where teams separated by
physical distance. While in collocated software development
there are generally a best practices standard, artifact, and ad-
hoc coordination to do "real-time" software development. It
does not really exist in distributed software development.
Best practices, artifacts, or ad-hoc communication are done
to supply sufficient "shared vision" between the team, and also
client. Shared vision is creating a common understanding about
what the team is trying to do, what the finished product look
like, what the basis of the product, and when they must deliver
the product if it is to have its intended effect [8]. Either the
system is developed in collocated or distributed; the shared
vision its necessity to limit the development risk.
Shared vision in distributed development faces some
challenges and issues. The first matter is commonality. It is
uncommon to have multiple divisions, organizations, cultures,
and languages on a project. Often the team members have not
known each other, may have different level experiences, and
may have motivation conflict. All these factor plots to make it
increasingly difficult to coordinate across teams, manage
evolution, and monitor progress. Past studies have shown that
tasks take about 2.5 times longer that the collocated one [6].
Sangwan et al. [11] creates a set of practices that can be
used on projects that are geographically distributed. The
research identified a crucial factor to the success of GSD
projects and construct best practices for these factors to enable
a successful outcome. The outcome of the research is a process
framework that is leveraged the best practices in the critical
success factors and is based on the agility model. However, the
research still argued about agile implementation as a tradeoff of
its discipline process [2].
52
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
II. GSD Process Framework
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, 2010
approaches will be
Process framework in GSD is divided in four major phases,
which are requirements engineering, project planning,
architecture design, and product development. Those phases
are known as a software development life cycle (SDLC). GSD
SDLC provides wide range phases to initiate the project
(planning and requirement) and to construct the product
(architecture design and product development). Along with the
process framework, GSD also provides organization process,
and monitoring control as parts in its process framework. This
section will discuss the concept in separated sub sections.
Figure 1 shows the GSD process framework.
Mdr*.e1 Intflflt'PwlyCl Vi90*l—
Process Framework
" Goals
RaouirfimflTils Engirwflnng
Ptafici Planning
ArchitaCLiff Design
- ar>d Module -
3£tCih£itont
F+oducL Ob wtopnsm
Figure 1. GSD process framework [11]
Each process can be threatened as a process component that
has several input and output. Although it shows the sequential
one, several implementation of the process framework is
described as highly iterative and incremental. The next section
will discuss the concept in separated sub sections.
A GSD requirements engineering
The major concern in requirement engineering is doing
analysis of the problem, and creating a system based on the
available solution. It can be composed as functional
requirements and non-functional requirements. Requirements
answer questions about what should be implemented, how the
system should behave, what system attributes and constraints.
Requirement engineering process is started by finding
business goals and market intent for the product. The product
should have a reasonable definition to build, the benefit for the
system, and have clear target markets. Without reasonable
definition, it can be difficult to define concretely the users,
overlooked the product, and create an assumption of the
product.
Sangwan, et al. [11] proposes the requirement engineering
process through three main activities; which are elicitation,
modeling, and review. Each activity provides output and
process one or several input.
The elicitation activity is to extract the domain knowledge
from the experts, identify features that are in scope and out of
scope, and document the features in the highly structured and
logical model. In GSD, that activity can be done through
several approaches starting from face-to face meeting (it's
meaning the team should plan to travel), using CSCW tools
like instant messaging, or phone conferences, or indirectly by
using email communication. Those
discussed further in case studies section.
The primary output of the elicitation activity is a feature-
level requirements model. Feature level requirements model
focuses in creating a high-level structure of the product feature
without detailing and completing the model itself. The creation
step to provide the high-level structure by doing function
decomposition and categorize the feature through work based
structure model. Through this way the GSD team will get start
with easy and common understanding through a simple model
like Unified Modeling Language.
The model itself provides indirect control between project
planning and the design model. However, the requirement
model can sometimes become complex, difficult to understand,
and have been conflicting with requirements. Therefore, the
review process is important to validate the project plan,
adjusting the design, and following the software process. Tools
are playing some rules to provide several syntactic reviews
(e.g. model completeness, violation in rules, etc.). The semantic
syntactic should be manually reviewed. In GSD, these step
called requirement review meeting. These meeting have three
goals, which are; ensuring the requirements were correctly
understood and translated into the model, verifying the
requirement model itself is readable and understandable, and
disseminating the requirements itself.
B. GSD Architecture Designs
The drivers for software architecture in GSD project are
shared vision. When the distributed teams have shared vision
about the architectures, the development and design will shape
by itself. The concerns are many people intuitively know how
to create shared vision, but make the intuition explicit about the
remote teams is challenging. Shared vision come when the
team has same domain knowledge, and is able to design the
architecture in well-defined and loosely coupled components.
Define work units is the first step in GSD architecture
design to decompose the whole system into functional units
that can be allocated distributed in development. The drivers
for how to decompose the functional units come from several
considerations such as previous experiences, architecturally
significant requirements, referenced patterns, organization
information, and resource location.
The second step is to identify module responsibility. The
functional unit results are combined together with the
requirement models. The result of this step is module
specification, which is described the detail function of each
module. As a structural result, each module will be categorized
into two main categories, which are static and dynamic
modules. Static modules are defined as a module that works
and dedicated for common purposes such as security, print,
utility, compression, and any modules that work loosely
coupled with the business process. Dynamic modules are
defined as a collection of modules that work tight coupled with
the business process, e.g. payment processing, inventory
checking modules, etc. Both module categories are displayed in
a relation through an UML interaction diagrams.
53
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
The third steps in architecture development are creating the
architecture documents. Architecture document works as a
reference guide in development, not as a manual to build the
system. Clements et al. [4] provides a guidance to build an
architecture document in a sufficient way. According to the
research, an architecture document consists of two main views,
which are; execution views and implementation views.
C. GSD project planning
Project planning in GSD is focused in allocating the amount
work, budget and time in distributed manners. Project planning
in GSD usually uses time-boxed development to approach
synchronous development among the various geographically
distributed development sites. Requirements and architecture
results are the main input for this process. Cost estimation and
project plan is the main output in project planning.
Feature release planning is the first phases in project
planning. It is designed from market intent and functional
specification. The feature release planning is an iterative
activity that can be revised in time constraints depend on the
market intent and requirements changes. Feature release
planning is dividing the features of the system through time
iteration. Iteration may vary from two weeks or months. The
goal of this activity is to identify the minimum set of features
that can be sold in the marketplace.
Development planning is proposed by mixing feature
release plan, design model, and critical path analysis. The result
of the development planning is a project plan. Project plan
consists of schedule planning and integration-test planning.
Schedule planning provides a fix date for features. It is
driven by the critical path analysis and design model that
proposed in architecture sessions. Thus, each development
team must release an operational version to integrate and tested
on the date specified. If the team can make the date, the release
would be skipped, and the team would release on the next fixed
date on the next sprint.
Integration and test planning provides a detail when the
production can hit the date. It is driven by feature release plan
and schedule planning. Iterations will be followed by test and
integration planning. The result from this step is a detail works
of schedule called as a project plan.
Project plan can be a baseline for the team to make budget
proposals. It will become to go or not go to decision. Although
it becomes a trade off in the budget versus feature. The project
plan can be a live artifact for future reference development.
Therefore, GSD proposes three phases in project planning,
which are planning in inception phase (requirement phase),
planning in elaboration phase (architecture design phase), and
planning in construction phase (product development).
D. GSD product development
GSD product development is preliminary started by
structuring the team. GSD proposes a hierarchical team
structure. It contains a central team, and site team. GSD
provides a basic composition of the central team and team site.
Basic composition provides the basic structure of the team.
There are two communication models between central team
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, 2010
and site team, which are vertical communication and horizontal
communication. Horizontal communication is a
communication model that exists between team members in
one site. Vertical communication happens between two
different teams (e.g. central team with the site team) through a
proxy communication model. Supplier manager is
representative of central team, and R&D manager is
representative of site team. Table 1 provides the casual team
role in GSD.
TABLE I.
GSD Team Strucures
Role
Responsibility
Product manager
create release planning and decide what feature
will be released in a milestone
Program manager
plans the entire lifecycle of a product, managing
the requirements, and partitioning component
development in team site
Supplier manager
Plans the entire components and manages a remote
component development team.
Subject matter expert
Provides expertise on the business that the project
is to support. Requirements engineer captures
customer requests, develop the analysis model and
glossary, and translates analysis model to the
requirements document
Requirements engineer
Captures customer requests, develop the analysis
model and glossary, and converts analysis model
to the requirements document.
Architect
Expresses, designates, and communicates
architecture solution, implements an architectural
prototype, develops specification, and acceptance
tests for product components.
QA Experts
Provisions the automated build/test, test
management, and defect tracking systems used by
the architecture project team.
R&D Manager
Plans the entire development lifecycle and manage
the module development plans.
Designer
Focuses to build user interface and designs
modules using appropriate design patterns.
Developer
Gears the module and increment the quality of the
codes.
The communication models exist to solve coordination
between team and inside the team. Solving a problem in GSD
need to know who solve the problem, adjust the schedule, and
shift task to the other's team. It works using traditional or the
internet based communication tools. In line with engineering
and planning activities, preparing an adequate project
infrastructure for distributed development is an essential
project success factor.
Infrastructure supports in GSD should support accessibility,
collaboration, concurrency, processes, awareness, and
integration. Accessibility is a characteristic that the tools must
be available at all development sites. Collaboration is
characteristic that the tools must support both real-time
collaboration and asynchronous collaboration. Concurrency
refers to the extent to which different parties can collaborate
explicitly or implicitly on a single artifact. The process
Characteristics focuses in a specific support tool for a software
engineering process (e.g. Rational Rose for RUP). Awareness
could, for example, be supported by making as much
information as possible available in a single location and
linking different artifacts for navigability. The last
54
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
characteristic which is integration is a nice to have that tools
support the integration between different tools through real-
time integration or file based integration. Conventional tools
may not address all the characteristics; therefore, selecting right
tools is the important factor in GSD. Table II provides previous
researches in the GSD researches.
TABLE II.
GSD Tools Related Researches
Authors
Research Topic
Fitzpatrick et al., 2006
Discussing the tools and its implementation for
GSD management process through notification
and chat.
Pillati et al., 2006
Discussing about software configuration
management tools in GSD.
Bass, 2006
Explains monitoring technique for GSD process
by using shared mental models approach
Taylor et al., 2006
Discussing the implementation opportunity to
adopt Agile in GSD through open collaborative
tools.
Thissen et al., 2007
Discussing several alternatives for GSD
communication tools such as email, web, and
webcasts.
GSD Product development is the last step of GSD process.
After this step, product should be delivered to the customer.
Product in GSD process is defined as software and its artifacts
(e.g. document, user manual, installation guide).
III. GSD Process Framework Current issues
The phase in GSD is over when the product delivered and
passed the user acceptance test. Although GSD provides a
comprehensive process, it does still require software
engineering methodology like RUP to provide technical how-
to. GSD also needs specific tools to make the process become
more effective, some research aware of provide tools like
notification chat, web email, open collaborative tools, and
software configuration models. Joining GSD with effective
software development methodology and sufficient tools
absolutely give advantages in the overall software project.
Effective software development is defined as a construction
method that focused to deliver the right software products in
the right time with the right tools. Developing nowadays
software faces challenges that are not those much different
from building medieval war ships. There is the "unknown"
factor, since many technical problems cannot be understood in
their entirety at beginning of the project.
Stober and Hansmann [12] identified those challenges such
as complexity of the infrastructure on which IT solutions are
built. The challenges arising from international and distributed
teams, not yet identify dependencies, exploding costs, and
rapidly approaching deadlines, requirements that come up late
in the game of decision, unexpected issues to mitigate, and the
required knowledge is not sufficient and to make it worse the
need for cost-effectiveness is going as a market competition.
How do we make software development become effective in
GSD, while the team is separated geographically?
Requirements are the key of the effective software
development [1]. Through the stable requirements, it is easy to
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, 2010
define distinct milestone, development phases, and acceptances
closing phases, although it separates on distant. Unfortunately,
requirement changes are usually happened even in the most
precisely elaborated software development planning.
Therefore, it is important to prepare the team to focus on the
ability to adapt quickly to face any kind of forthcoming
challenges. Involving the development team as stakeholders,
rather than suppliers is a great way to establish an open project
structure which responsibility and changes control is shared
[12].
Since the key of effective software development is
functional software, Agile promises a better way to developing
software with a lightweight and adaptive process. Agile
provides manifestos and twelve principles that can be seen
online at http://www.agilemanifesto.org/. Those principles
focus in delivering working software through individual and
interactions, customer collaboration and responding changes.
Agile process implemented through several methods like
extreme Programming (XP), Scrum, Dynamic Software
Development Method (DSDM), Agile modeling, Crystal, Lean,
etc. Every method has own unique approaches to build in an
agile way.
Relating the agile and GSD is somewhat contradiction
Agile needs intensive communication both within the customer
and the team, but GSD make a distribution of the team that also
contributes to the team dysfunction. Communication between
peers becomes important to the team's collective
understanding. Remote team members miss these, and
consequently, their understanding suffers. Miller [9] shows that
when the direct communication doesn't exist in their practices.
It will weaken the adopted method. Taylor, et al [13] shows the
agile adoption in GSD is just like reinventing the wheel, since
many of the Agile GSD experience reports are not giving any
additional value in the existing GSD guidance. They also
recommend the researcher to create a value or a framework to
integrate Agile in GSD context.
Although it has a contradiction in the term of conditions,
several researchers also provide field reports about successful
agile adoption in GSD. Miller [9] confidently said that his team
at Microsoft delivered sufficient software by integrating Scrum,
XP, and GSD. Hazzan and Dubinsky [5] states the diversity
that exists in GSD is naturally supported by agile software
development. Paasivaara et al [10] captures the Scrum practices
that successfully adopted in three GSD projects. The following
section will discuss how extreme Programming method is
integrated with the current GSD framework to provide a
process framework that the research called as the Global
extreme Programming process framework.
IV. Research methodology
In order to integrate between the GSD process and XP
methodology, the research did case studies that based on the
real project. The organization that we are selected is the small
independent software vendor that contains five development
members with two clients who dedicated to support the project.
The project length is six months and works to develop project
management that running on the web.
55
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
The team itself has two years average of experience in
multi-site development. The client and the team are separated
in the different countries with different time zone. The team
composition as like follows.
(IJCSIS) International Journal of Computer Science and Information Security,
Vol 8, No. 6, 2010
TABLE III.
Team Structures
Roles
Job Descriptions
Customer
Drive the product direction to the team, answer the
developer questions, write stories, declare stories complete,
write acceptance test, accept the release
Coach
A thorough understanding of the XP process, professional
development experience, leadership experience, get the
developers and testers to adhere to the values and practices,
assume a leadership role for the development and testing
teams, pair with other developers and testers as needed,
assist customers in the story writing process, assist
customers in the acceptance test writing process
Developer
Estimate stories, brainstorm tasks, develop unit tests,
develop the tasks, refactor the code, and communicate with
customers when questions arise. There are two developers
in this case study
Client Subject
matter expert
Help the customers define intelligent stories, act as a proxy
for the customer.
Tester
Ensure that a story is testable, assist the customers with
writing acceptance tests, run the acceptance tests
The Tracker
Collect development metrics, produce reports indicating
the team's progress, communicate the team's historical
velocity, and communicate the status of the team's
progress
The research will collect the data by exploring the case
study related artifacts in the proposed workflow. The research
procedures are come as follows.
• The team is incubating and does workshops to
implement the proposed GXP framework process.
• Based on the theory, the team designs the artifacts.
• The evaluator evaluates the execution and captures
several data that can be written as finding of the case
study research.
Based on the project execution the GXP process framework
finding is discussed the next sections.
V. GXP Process Framework
Understanding the GXP process framework understands the
hybrid approaches that integrated between conventional GSD
process and XP method. As we know, conventional GSD
process divides the SDLC into four phases, which are
requirements engineering, project planning, architecture design,
and production. XP method adopts exploration, planning,
iteration, production, and maintenance as a phase in a
development cycle. The GXP SDLC proposes the uses the XP
method as a baseline and integrates the GSD workflow on it.
Figure 2 shows the GXP process framework.
<iXF SULC
^ i
1
Prc-djd Viacn
(HOW)
1. L3J LtjlitllJ -x.
rtududHc-jpr
PrniiKt_
tjtpfcHE^icn
H'ljnnnj
llcialuiB K.im g
rfftafieilLitinr
l f 1Yj
1
1
Balaam Fla'ri-g
h™^ S
a
PluJj.L 1
a
■hkltL-HKC
I
[>.:d.].n.:i h.:.:dl. M*
cnuw>
A.»ipL....H: 1 C :J
Figure 2. GXP Process Framework
GXP process
such as.
framework provides several input and output
Product vision. Product vision works as an input that
shows a reasonable reason definition why the company
building the product, what system will be used for, and
what the target markets are.
Business goal. Business goal works as an input that
described a business background for the product. It can
be related with productivity enhancement, increasing
revenue, gaining more customers, or cutting costs.
Product scope. Product scope works as an input that
show what kind of feature that will be developed in a
timeframe. This input usually depends on available
resources, existing budget, and time limitation.
User stories. User stories work as an output from the
exploration phases and an input for planning phase.
User stories describe the feature level requirements
model. User story details are described in user story
estimation. The detail's version of user stories is called
as the detail requirements model.
Spike solution. Spike solution is a high level
architecture of the product. It works as an input for
iteration phase and an output from exploration phase.
Spike solution also describes the simple design for the
product.
Release planning. Release planning is a milestone
plan for the product. It described the feature that will
be delivered in iteration. Release planning is an output
from the planning phases and works as an input for
production phase.
Iteration planning. Iteration planning is a set of
activities to build features in iteration. The activity is
designed from the task card that structured based on
planning game activity. It works as an output from
planning phase and an input for iteration phase.
Customer feedback. Customer feedback is a customer
feedback based on iteration. It described validation and
verification for a customer request. Changes will
happen when the team built misleading feature. It
works as an input for iteration phase.
56
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
• Small releases are an output from iteration. Small
releases described potential working feature that
delivers in iteration. Small releases also work as an
input for production phase.
• Acceptance test is an input for production phase. It
describes a black box test for product features. The
black box test usually runs by the customer.
• Product is an output from production phase and an
input from maintenance phase. It works as a final
release of the feature.
A Exploration phase
Exploration phase is where discovery of vision and mission
occurs. This phase is equal with requirements engineering
process in the GSD. At this stage, the stakeholder with coming
up with the vision statement. Vision statement is a high-level
statement that describes the goal of the system. For example,
the purpose of this new system is to read books via the Internet.
Vision statement is not more than 20-30 words. Then, this
vision statement is described visually through unique XP
artifacts called system metaphor. System metaphor is how the
team conceptualizes the system and is typically written in
language that is relevant to the business domain.
The detail of the system metaphor is described through a
user story. It is the tool that captures user requirements. User
stories are similar to use cases and are written by the customer
in non-technical language. Most user stories are around 1-4
sentences long. For example, library member can borrow 10
books for a maximum. User stories are written in index card or
sticky note that posted on a whiteboard.
Since the GXP is working in multi-site development the
user stories and vision statement should be documented in an
artifact. GXP proposes the user stories' artifacts such as
follows.
• Users, persona or actor that will use the system.
• User stories, the lists of the user scenarios that
supported by the system.
• Task cards, it provides a detail tasks that should be
performed for each story.
• Software estimation. The software estimation can be
done through estimation techniques such as user stories
points, or early estimation by the developer.
The user stories' artifacts are composed by the project
manager and the clients. The artifacts will be uploaded and
maintained through the document repositories that run on the
internet.
B. Planning Phase
Planning phase starts with the activity called planning
game. Planning game is a short workshop that plays by
customer and development team. Customer purposes are
determining the value, and development team determines the
cost. There are two stages of planning game, the first one is
releasing planning game, and the second is iteration planning
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, 2010
game. In the release planning game, the goal is to define the set
of features in the next release. The customer writes the stories,
development team estimates the story point, and customer
plans the overall release. In iteration planning game, the
customer chooses the stories for the iteration, and the
development team estimate how long, and accepts the
corresponding task. Estimation is based on experiences or
intuition of the development team. The goal in iteration
planning game is to define how long the iteration and what
kind of feature that delivered in iteration.
The planning game execution actually runs parallel with the
exploration phase. Therefore, the information in user story
artifact is refined in the planning phase. This phase also
discusses about creates an artifact called Iteration and Release
Planning. This artifact is created with a composition as follows.
• Categorized user story. In this section team should
create user stories that already categorized based on its
urgency. GXP introduces four levels of priority, which
are high (essentially needed), medium (adding business
value), low (nice to have) and none (verbose). Team
can easily throw none level and negotiate for low level.
• Project plan summary. It contains how long is the
project, how long the iteration length, how much the
user stories, and how many iterations in a project.
• Iteration detail. It describes the user story that executed
in the iteration. Actually, the iteration detail discusses
the work assignment for developer.
The artifact is also stored and maintained by the project
manager. It will be updated by the developer when the user
story is completed
C. Iteration Phase
Iteration phase is the real work of development happens.
The development team selects the story in current iteration.
Since user stories are generally too high-level to execute as
programming tasks, so the next step is to break out the actual
development tasks needed to support the story. Development
team works together to investigate the tasks for each user story.
The tasks are written either on the back of the story, on
separate task cards, or on some other tracking mechanism.
Developers begin work on a new task by writing the test first
and adding it to their test framework. Programming continues
in pairs with each partner taking turns to "drive" the keyboard
from time to time. At the end of the iteration, customers
perform the acceptance tests they have written. Any user stories
that fail acceptance will be fixed during the next iteration; there
is why the iteration phase provides feedback to planning phase.
The result of the iteration is software that works in staging
environment. Staging environment is the customer's software
environment that mirrors, or replicates, their live environment.
Staging is used to perform final prerelease tests and
performance checks. Customer will evaluate the staging system
and then approve the system shift to the production
environment.
In the iteration phase, the artifacts that live in the internet
are stored in the source code control systems. The tools like
57
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
Team Foundation Server, Team City, or hosted source code
online in third party are the several options to store the most
important artifacts in the software development. In GXP the
developer should have several agreements based on the codes
such as follows.
• The coding standard and naming conventions of the
codes. This little agreement gives benefits for the
other's team to understand the code better.
• The comment agreement. This approach clearly makes
the codes speak what developer did with the codes.
Having automatic commenting software like
GhostDoc, will make it happen.
• The developer notes. This approach makes an
agreement how the developer communicate each other
through the special comment token. A word like
TODO, BUGBUG or DEBUG will make the developer
communication is better in term exchanging
information.
D. Production and Maintenance Phase
Production is the user acceptance test that happens with the
developed codes or modules. The developer set up the system
in the development system and tests it by the client. The
approved features is copied into the production system or
staging system. The approval model works as follows.
• Developers build the codes based on the user stories,
and then uploaded into the development system.
• The client and developer do online meeting and discuss
through shared screen and review sheets.
• The client fills the review sheet approves it or rejects it
for further modification.
After the production, the system goes into the maintenance
mode. In this mode, development team can create an upgrade,
patch, or changes code with confidence since the change's
monitoring can be easily defined through test case. The hard
part in this phase is creating data migration and system
migration. Since those parts are technology specific, XP does
not provide the sufficient techniques to overcome it. The
artifact that exists in this phase is.
• Review sheet for user acceptances tests
• Defects document that fulfilled also by the client
VI. DISCUSSION AND FUTURE WORKS
In our studies, we do a novel approaches to integrate the
extreme Programming with the Global Software Development
process. Although both are contradict in the first vision, both
can be integrated through the power of artifacts and the tools.
The gap in direct communication between client and customer
can be limited by using a formal process framework and the
tools like source code version document management, or
others. The research also notes several finding that related with
the artifacts that designed by the team to tackle a
communication gaps between of them.
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, 2010
In the future, evaluation can be further improvements of the
research. For example, it will be a good idea to compare the
proposed process framework with the existing one. The impact
of the process framework in terms of productivity, product
quality, and resources will make the benefit of the framework
can be easily quantified. We are currently in the process of
defining some empirical studies with real multi-site distributed
projects.
References
[I] Belford, P. C, Bond, A. F., Henderson, D. G., and Sellers, L. S. 1976.
Specifications a key to effective software development. In Proceedings
of the 2nd international Conference on Software Engineering (San
Francisco, California, United States, October 13 - 15, 1976).
International Conference on Software Engineering. IEEE Computer
Society Press, Los Alamitos, CA, 71-79.
[2] Boehm, B. and Turner, R. 2004. Balancing Agility and Discipline:
Evaluating and Integrating Agile and Plan-Driven Methods. In
Proceedings of the 26th international Conference on Software
Engineering (May 23 - 28, 2004). International Conference on Software
Engineering. IEEE Computer Society, Washington, DC, 718-719.
[3] Carmel, E. 1999 Global Software Teams: Collaborating Across Borders
and Time Zones. Prentice Hall PTR.
[4] Clements, P., Bachmann, F., Bass, L., Garlan, D., Ivers, J., Little, R.,
Nord, R., and Stafford, J. 2002. Documenting Software Architectures:
Views and Beyond, Addison- Wesley.
[5] Hazzan, O. and Dubinsky, Y. 2008. Agile Software Engineering.
Springer.
[6] Herbsleb, J. D. and Mockus, A. 2003. An Empirical Study of Speed and
Communication in Globally Distributed Software Development. IEEE
Trans. Softw. Eng. 29, 6 (Jun. 2003), 481-494.
[7] Herbsleb, J. D., Paulish, D. J., and Bass, M. 2005. Global software
development at Siemens: experience from nine projects. In Proceedings
of the 27th international Conference on Software Engineering (St. Louis,
MO, USA, May 15 - 21, 2005). ICSE '05. ACM, New York, NY, 524-
533.
[8] McCarthy, J. and McCarthy, M. 2006 Dynamics of Software
Development. Second. Microsoft Press.
[9] Miller, A. 2008. Distributed Agile Development at Microsoft Patterns
and Practices. Microsoft.
[10] Paasivaara, M., Durasiewicz, S., and Lassenius, C 2009. Using Scrum
in Distributed Agile Development: A Multiple Case Study. In
Proceedings of the 2009 Fourth IEEE international Conference on
Global Software Engineering (July 13 - 16, 2009). ICGSE. IEEE
Computer Society, Washington, DC, 195-204.
[II] Sangwan, R., Bass, M., Mullick, N., Paulish, D. J., and Kazmeier, J.
2007. Global Software Development Handbook (Auerbach Series on
Applied Software Engineering Series). Auerbach Publications.
[12] Stober, W, and Hansmann, U. 2010. Agile Software Development: Best
Practices for Large Software Development Projects. Springer.
[13] Taylor, P. S., Greer, D., Sage, P., Coleman, G., McDaid, K., and
Keenan, F. 2006. Do agile GSD experience reports help the
practitioner?. In Proceedings of the 2006 international Workshop on
Global Software Development For the Practitioner (Shanghai, China,
May 23 - 23, 2006). GSD '06. ACM, New York, NY, 87-93.
AUTHORS PROFILE
Ridi Ferdiana. Mr. Ridi Ferdiana was born in 1983. He is a doctoral student
at Gadjah Mada University, Yogyakarta since 2008. He earned his master
degree from the same university in 2006. In his professional area, he holds
several professional certifications such as MCP, MCTS, MCPD, MCITP,
MVP and MCT. In his daily research activities he really enjoys to learn about
software engineering, business platform collaboration, and programming
optimization.
58
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, 2010
Lukito Edi Nugroho. Born in 1966, Dr. Lukito Edi Nugroho is an Associate
Professor in the Department of Electrical Engineering and Information
Technology, Gadjah Mada University. He obtained his M.Sc. and PhD
degrees from James Cook University in 1995 and Monash University in 2002,
respectively. His areas of interest include software engineering, distributed
and mobile computing, and application of ICT in education.
Paulus Insap Santosa. Insap was born in Klaten, 8 January 1961. He
obtained his undergraduate degree from Universitas Gadjah Mada in 1984,
master degree from University of Colorado at Boulder in 1991, and doctorate
degree from National University of Singapore in 2006. His research interest
including Human Computer Interaction and Technology in Education.
Ahmad Ashari Place and date of birth: Surabaya, May 2 n 1963. Get
Bachelor's degree 1988 in Electronics and Instrumentation, Physics
department Gadjah Mada University, Yogyakarta. Master degree 1992 in
Computer Science, University of Indonesia, Jakarta Doctor Degrees 2001 in
Informatics, Vienna University of Technology. Major Field of study is
distributed system, Internet, Web Services, and Semantic Web.
59
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol.8,No.6,2010
A Hybrid PSO-SVM Approach for Haplotype
Tagging SNP Selection Problem
Min-Hui Lin
Department of Computer Science and Information
Engineering, Dahan Institute of Technology,
Sincheng, Hualien County , Taiwan, Republic of China
Chun-Liang Leu
Department of Information Technology, Ching Kuo
Institute of Management and Health,
Keelung , Taiwan, Republic of China
Abstract — Due to the large number of single nucleotide
polymorphisms (SNPs), it is essential to use only a subset of all
SNPs called haplotype tagging SNPs (htSNPs) for finding the
relationship between complex diseases and SNPs in biomedical
research. In this paper, a PSO-SVM model that hybridizes the
particle swarm optimization (PSO) and support vector machine
(SVM) with feature selection and parameter optimization is
proposed to appropriately select the htSNPs. Several public
datasets of different sizes are considered to compare the proposed
approach with other previously published methods. The
computational results validate the effectiveness and performance
of the proposed approach and the high prediction accuracy with
the fewer htSNPs can be obtained.
Keywords : Single Nucleotide Polymorphisms (SNPs),
Haplotype Tagging SNPs (htSNPs), Particle Swarm Optimization
(PSO), Support Vector Machine (SVM).
I. INTRODUCTION
The large number of single nucleotide polymorphisms
(SNPs) in the human genome provides the essential tools for
finding the association between sequence variation and
complex diseases. A description of the SNPs in each
chromosome is called a haplotype. The string element of each
haplotype is or 1, where denotes the major allele and 1
denotes the minor allele. The genotype is the combined
information of two haplotypes on the homologous
chromosomes and is prohibitively expensive to directly
determine the haplotypes of an individual. Usually, the string
element of a genotype is 0, 1, or 2, where represents the
major allele in homozygous site, 1 represents the minor allele
in homozygous site, and 2 is in the heterozygous site. The
genotyping cost is affected by the number of SNPs typed. In
order to reduce this cost, a small number of haplotype tagging
SNPs (htSNPs) which predicts the rest of SNPs are needed.
The haplotype tagging SNP selection problem has become
a very active research topic and is promising in disease
association studies. Several computational algorithms have
been proposed in the past few years, which can be divided into
two categories: block-based and block-free methods. The
block-based methods [1-2] firstly partition human genome into
haplotype blocks. The haplotype diversity is limited and then
subsets of tagging SNPs are searched within each haplotype
block. A main drawback of block-based methods is that the
definition of blocks is not a standard form and there is no
consensus about how these blocks should be partitioned. The
algorithmic framework for selecting a minimum informative
set of SNPs avoiding any reference to haplotype blocks is
called block- free methods [3]. In the literature [4-5], feature
selection technique was adopted to solve for the tagging SNPs
selection problem and achieved some promising results.
Feature selection algorithms may be widely categorized
into two groups: the filter approach and the wrapper approach.
The filter approach selects highly ranked features based on a
statistical score as a preprocessing step. They are relatively
computationally cheap since they do not involve the induction
algorithm. Wrapper approach, on the contrary, directly uses the
induction algorithm to evaluate the feature subsets. It generally
outperforms filter method in terms of classification accuracy,
but computationally more intensive. Support Vector Machine
(SVM) [6] is a useful technique for data classification. A
practical difficulty of using SVM is the selection of parameters
such as the penalty parameter C of the error term and the kernel
parameter y in RBF kernel function. The appropriate choice of
parameters is to get the better generalization performance.
In this paper, a hybrid PSO-SVM model that incorporates
the Particle Swarm Optimization (PSO) and Support Vector
Machine (SVM) with feature selection and parameter
optimization is proposed to appropriately select the htSNPs.
Several public benchmark datasets are considered to compare
the proposed approach with other published methods.
Experimental results validate the effectiveness of the proposed
approach and the high prediction accuracy with the fewer
htSNPs can be obtained. The remainder of the paper is
organized as follows: Section 2 introduces the problem
formulation. Section 3 describes the PSO and SVM classifier.
In Section 4, the particle representation, fitness measurement,
and the proposed hybrid system procedure are presented. Three
public benchmark problems are used to validate the proposed
60
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
approach and the comparison results are described in Section 5
Finally, conclusions are made in Section 6.
II. PROBLEM FORMULATION
As shown in Figure 1, assume that dataset U consists of n
haplotypesjfy.}^.^ , each with p different SNPs{S j } 1 < j < p , U is
n*p matrix. Each row in U indicates the haplotype fy and each
column in U represents the SNPSj . The element d t . denotes
the 7-th SNP of z-th haplotype, d tj e{0,l} . Our goal is to
determine a minimum size g set of selected SNPs (htSNPs)
V = {v k }, k e {1, 2,..., p} , g = |v| , in which each random
variable v k corresponding to the /c-th SNP of haplotypes in U,
to predict the remaining unselected ones with a minimum
prediction error. The size of V is smaller than a user-defined
value R (g <R), and the selected SNPs are called haplotype
tagging SNPs (htSNPs) while the remaining unselected ones
are named as tagged SNPs. Thus, the selection set V of htSNPs
is based on how well to predict the remaining set of the
unselected SNPs and the number g of selected SNPs is usually
minimized according to the prediction error by calculating the
leave-one-out cross-validation (LOOCV) experiments [7].
(IJCSIS) International Journal of Computer Science and Information Security,
Vol.8,No.6,2010
where d =1,2,...,D , z=l, 2,...,S , and D is the dimension of
the problem space, S is the size of population, k is the iterative
times; vf d is the z-th particle velocity, x k id is the current particle
solution, ptf d is the z-th particle best (p besf ) solution achieved
so far; gb. d is the global best ( g best ) solution obtained so far by
any particle in the population; rj and r 2 are random values in
the range [0,1], both of q andc 2 are learning factors, usually
q = c 2 = 2 , w is a inertia factor. A large inertia weight
facilitates global exploration, while a small one tends to local
exploration. In order to achieve more refined solution, a
general rule of thumb suggests that the initial inertia value had
better be set to the maximum w max = 0.9, and gradually down
to the minimum w . = 0.4 .
h
5,
d u
J 2
d i,l d i,2
d n-l,l d n-l,2
d„
i
d.
2,j
d-u
**.,
J p-1
d ^
d ,,"
d 2,p-l
d 2 , P
V.
d l;P
Cl.p-l
d n-l,p
d -.,p-l
d n , P _
"n,l "n,2
Figure 1 The haplotype tagging SNP Selection Problem.
III. RELATED WORKS
A Particle Swarm Optimization
The PSO is a novel optimization method originally
developed by Kennedy and Eberhart [8]. It models the
processes of the sociological behavior associated with bird
flocking and is one of the evolutionary computation techniques.
In the PSO, each solution is a 'bird' in the flock and is referred
to as a 'particle'. A particle is analogous to a chromosome in
GA. Each particle traverses the search space looking for the
global optimum. The basic PSO algorithm is as follow:
-n-ipK ~4)+Ci - r i '(qK -*£)
(i)
According to the searching behavior of PSO, the gbest
value will be an important clue in leading particles to the global
optimal solution. It is unavoidable for the solution to fall into
the local minimum while particles try to find better solutions.
In order to allow the solution exploration in the area to produce
more potential solutions, a mutation-like disturbance operation
is inserted between Eq. (1) and Eq. (2). The disturbance
operation random selects k dimensions (1 < k < problem
dimensions) of m particles (1 < m < particle numbers) to put
Gaussian noise into their moving vectors (velocities). The
disturbance operation will affect particles moving toward to
unexpected direction in selected dimensions but not previous
experience. It will lead particle jump out from local search and
further can explore more un-searched area.
According to the velocity and position updated formula
mentioned above, the basic process of the PSO algorithm is
given as follows:
1.) Initialize the swarm by randomly generating initial
particles.
2.) Evaluate the fitness of each particle in the population.
Compare the particle's fitness value to identify the both
3.)
of A
5.)
6.)
70
and g hest values.
x ; ,
=vi, +i +**
Update the velocity of all particles using Equation (1).
Add disturbance operator to moving vector (velocity).
Update the position of all particles using Equation (2).
Repeat the Step 2 to Step 6 until a termination criterion
is satisfied (e.g., the number of iteration reaches the pre-defined
maximum number or a sufficiently good fitness value is
obtained).
The authors [8] proposed a discrete binary version to allow
the PSO algorithm to operate in discrete problem spaces. In the
binary PSO (BPSO), the particle's personal best and global
best is updated as in continuous value. The major different
between discrete PSO with continuous version is that velocities
of the particles are rather defined in terms of probabilities that a
bit whether change to one. By this definition, a velocity must
be restricted within the range [V mh ,V nax ] . If v, k d +1 £ (V min ,V max )
thenv^/ 1 = max(min(y max , v ( k d +1 ),V min ) . The new particle position
(2) is calculated using the following rule:
61
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
If rand( ) < S(v* +1 ) , then x* +1 = 1 ; else x*
, where S(v* +1 ) = -
1+e"
The function S(v ld ) is a sigmoid limiting transformation and
rand() is a random number selected from a uniform distribution
in [0, 1]. Note that the BPSO is susceptible to sigmod function
saturation which occurs when velocity values are either too
large or too small. For a velocity of zero, it is a probability of
50% for the bit to flip.
B. Support Vector Machine Classifier
SVM starts from a linear classifier and searches the optimal
hyper-plane with maximal margin. The main motivating
criterion is to separate the various classes in the training set
with a surface that maximizes the margin between them. It is an
approximate implementation of the structural risk minimization
induction principle that aims to minimize a bound on the
generalization error of a model.
Given a training set of instance-label pairs
(x ( , y. ), i = 1, 2, ..., m where x z . e R n and y. e {+1, -1} . The
generalized linear SVM finds an optimal separating hyper-
plane f(x) = (w-x) + b by solving the following optimization
problem:
(IJCSIS) International Journal of Computer Science and Information Security,
Vol.8,No.6,2010
(3) Linear SVM can be generalized to non-linear SVM via a
mapping function O , which is also called the kernel function,
and the training data can be linearly separated by applying the
(4) linear SVM formulation. The inner product (0(x ( ) • 0(x j )) is
calculated by the kernel function k^x^Xj) for given training
data. By introducing the kernel function, the non-linear SVM
(optimal hyper-plane) has the following forms:
Minimize — w T w + C^% {
Subject to: y i (<w-x i >+b) + £. -1>0, £ >0
(5)
where C is a penalty parameter on the training error, and ^ is
the non-negative slack variables. This optimization model can
be solved using the Lagrangian method, which maximizes the
same dual variables Lagrangian L D (a) (6) as in the separable
case.
m I m
Maximize L D (a) = £ a t - - £ a i a j y i y j < x. • x . >
m
Subject to: < a. < C, z = 1, 2, ..., m and ^ a.y. =
(6)
To solve the optimal hyper-plane, a dual Lagrangian
L D (a) must be maximized with respect to non-negative a {
m
under the constraint ^a.y. = and < a ( < C . The penalty
;=1
parameter C is a constant to be chosen by the user. A larger
value of C corresponds to assigning a higher penalty to the
errors. After the optimal solution a* is obtained, the optimal
hyper-plane parameters w* and b* can be determined. The
optimal decision hyper-plane f(x,a,b*) can be written as:
f(x,a,b*) = ^ d y.a* <x i -x>+b* =(w -x) + b* (7)
f (x, a , b* ) = J] y, a* < O(x) • ® (x, ) > +5*
m
(8)
Though new kernel functions are being proposed by
researchers, there are four basic kernels as follows.
• Linear: k(x i ,x j ) = xjxj (9)
• Polynomial: k(x i ,x.) = (yxj x . + r) d , y > (10)
• RBF: k(x,,x j ) = exp(- r ||x,-x j || 2 ),/>0 (11)
Sigmoid: k(x ; . , x. ) = tanh(/x i T x j + r)
(12)
where y , r and d are kernel parameters. Radial basis function
(RBF) is a common kernel function as Eq. (11). In order to
improve classification accuracy, the kernel parameter y in the
kernel function should be properly set.
IV. METHODS
As the htSNPs selection problem mentioned above in
Section 2, the notations and definitions are used to present our
proposed method. In the dataset U of nxp matrix, each row
(haplotypes) can be viewed as a learning instance belonging to
a class and each column (SNPs) are attributes or features based
on which sequences can be classified into class. Given the
values of g htSNPs of an unknown individual x and the known
full training samples from U, a SNP prediction process can be
treated as the problem of selecting tagging SNPs as a feature
selection problem to predict the non-selected tagging SNPs in x.
Thus, the tagging SNPs selection can be transformed to solve
for a binary classification of vectors with g coordinates by
using the support vector machine classifier. Here, an effective
PS O- SVM model that hybridizes the particle swarm
optimization and support vector machine with feature selection
and parameter optimization is proposed to appropriately select
the htSNPs. The particle representation, fitness definition,
disturbance strategy for PSO operation and system procedure
for the proposed hybrid model are described as follows.
A Particle Representation
The RBF kernel function is used in the SVM classifier to
implement our proposed method. The RBF kernel function
requires that only two parameters, C and y should be set. Using
the RBF kernel for SVM, the parameters C , y and SNPs
viewed as input features which must be optimized
62
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No.6, 2010
simultaneously for our proposed PSO-SVM hybrid system. The
particle representation consists of three parts including: C
and y are the continuous variables, and the SNPs mask are the
discrete variables. For the feature selection, if n f features are
required to decide which features are chosen, then n f +2
decision variables in each particle must be adopted.
Table 1 shows the particle representation of our design.
The representation of particle z with dimension of n f + 2 ,
where n f is the number of SNPs (features) that varies from
different datasets. x u ~ x,. e {0,1} denotes the SNPs mask,
x t n +1 indicates the parameter value C , x in +2 represents the
parameter value y . If x ik =l,k=l,2,...,n f represents the k-th
SNP on the z-th particle to be selected, and vice versa.
TABLE I.
The particle
' representation.
Discrete -variables
Continuous-variables
SNPs mask
c r
X i,l X i,2 '" X i,n f
X i,n f +1 X i,n f +2
A random key encoding method [9] is applied in the PSO
algorithm. Generally, random key encoding is used for an
order-based encoding scheme where the value of random key is
the genotype and the decoding value is the phenotype. Note
that the particle in each {x ik \< k < n is assigned a random
number on (0, 1), and to decode in ascending order with regard
to its value. In the PSO learning process, the particle to be
counted larger tends to evolve closer to 1 and those to be
counted smaller tends to evolve closer to 0. Therefore, a repair
mechanism such as particle amendment in [5] to guarantee the
number of htSNPs after update process in PSO is not required.
B. Fitness Measurement
In order to compare the performance of our proposed
approach with other published methods SVM/STSA in [4] and
BPSO in [5], the leave-one-out cross validation is used to
evaluate the quality of fitness measurement. The prediction
accuracy is measured as the percentage of correctly predicted
SNP values on non-selected SNPs. In the LOOCV experiments,
each haplotype sequence is removed one by one from dataset U,
the htSNPs are selected using only the remaining haplotypes to
predict these tagged SNPs values for the removed one. This
procedure is repeated such that each haplotype in U is run once
in turn as the validation data.
C. The Proposed Hybrid System Procedure
Figure 2 shows the system architecture of our proposed
hybrid model. Based on the particle representation and fitness
Data Preprocessing
Dataset
Split dataset by LOOCV
_£
Training set
htSNPs (Feature) Selection
l
Testing set
Selected htSNPs (features) subset F
PSO parameter : htSNPs mask
Testing set tagged SNPs
Training set htSNPs
SVM parameter Optimization
Training SVM classifier
PSO parameters : C and r
Learned SVM classifier
PSO operation
Fitness calculation
No
PSO
operation
Optimized C , r , and feature subset F
Figure 2 The flowchart of the proposed PSO-SVM model.
measurement mentioned above, details of the proposed hybrid
PSO-SVM procedure are described as follows:
Procedure PSO-SVM hybrid model
1.) Data preparation
Given a dataset U is considered using the leave-one-out
cross-validation process to split the data into training and
testing sets. The training and testing sets are represented as
U JR andU TE , respectively.
2. ) PSO initialization and parameters setting
Set the PSO parameters including the number of iterations,
number of particles, velocity limitation, particle dimension,
disturbance rate. Generate initial particles comprised of the
features mask, C and y .
3.) Selected htSNPs (features) subset
Select input features for training set according to the feature
mask which is represented in the particle from 2), then the
features subset can be determined.
63
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol.8,No.6,2010
TABLE II. Results to compare PSO-SVM with SVM/STSA [4] and BPSO [5] on four real haplotype datasets.
Datasets
Prediction accuracy
%
(num of SNPs)
80
85
90
91
92
93
94
95
96
97
98
99
5q31
(103)
SVM/STSA
1
1
3
3
4
5
6
8
10
22
42
51
BPSO
1
1
2
3
4
5
6
7
9
14
29
42
PSO-SVM
1
1
2
2
3
4
5
6
7
10
23
36
TRPM8
(101)
SVM/STSA
1
1
2
5
5
6
7
8
10
15
15
24
BPSO
1
1
2
5
5
6
7
8
9
13
14
22
PSO-SVM
1
1
2
4
4
5
6
7
8
11
13
21
LPL
(88)
SVM/STSA
2
3
4
10
13
20
25
30
35
39
42
47
BPSO
2
3
6
9
12
16
18
21
25
28
31
37
PSO-SVM
2
3
4
7
10
12
13
17
20
22
26
31
4.) SVM model training and testing
Based on the parameters C and y which are represented in
the particle, to train the SVM classifier on the training dataset,
then the prediction accuracy for SVM on the testing dataset by
LOOCV can be evaluated.
5.) Fitness calculation
For each particle, evaluate its fitness value by the prediction
accuracy obtained from previous step. The optimal fitness
value will be stored to provide feedback on the evolution
process of PSO to find the increasing fitness of particle in the
next generation.
6.) Termination check
When the maximal evolutionary epoch is reached, the
program ends; otherwise, go to the next step.
7.) PSO operation
In the evolution process, discrete valued and continuous
valued dimension of PSO with the disturbance operator may be
applied to search for better solutions.
V. EXPERIMENTAL RESULTS AND COMPARISONS
To validate the performance of the developed hybrid
approach, three public experimental SNP datasets [4] including
5q31, TRPM8 and LPL are used to compare the proposed
approach with other previously published methods. When there
are missing data exist in haplotype datasets, the GERBIL [4-5]
program is used to resolve them. The chromosome 5q31
dataset was from the 616 kilobase region of human
chromosome 5q31 and the SNPs were 103. The TRPM8 which
consists of 101 SNPs was obtained from HapMap. The human
lipoprotein lipase (LPL) gene was derived from the haplotypes
of 71 individuals typed over 88 SNPs.
Our implementation platform was carried out on the Matlab
7.3, a mathematical development environment by extending the
Libsvm which is originally designed by Chang and Lin [10].
The empirical evaluation was performed on Intel Pentium IV
CPU running at 3.4GHz and 2 GB RAM. Through initial
experiment, the parameter values of the PSO were set as
follows. The swarm size is set to 200 particles. The searching
ranges of continuous type dimension parameters are:
Ce[10" 2 ,10 4 ]and y e [10" 4 ,10 4 ] . The discrete type particle
for features mask, we set [V min ,V max ] = [-6,6] , which yields a
range of [0.9975,0.0025] using the sigmoid limiting
transformation by Eq. (4). Both the cognition learning factor q
and the social learning factor c 2 are set to 2. The disturbance
rate is 0.05, and the number of generation is 600. The inertia
weight factor w m in = 0.4 and w m ax = 0.9 . The linearly
decreasing inertia weight is set as Eq. (13), where i now is the
current iteration and z m ax is the pre-defined maximum iteration.
w=w„ iI -^(w liI -w lil )
(13)
To compare the proposed PSO-SVM approach with the
SVM/STSA in [4] and BPSO in [5] on the three haplotype
datasets by LOOCV experiments, the computational results of
prediction accuracy according to the numbers of selected
htSNPs are summarized in Table 2. As mentioned in [4], it is
astonished that only one SNP for the 80% prediction accuracy
in 5q31 and TRPM8 datasets can be achieved. In practice, if
one guesses each SNP as 0, the prediction accuracy of 72.5%
for 5q31 dataset and 79.3% for TRPM8 dataset would be
obtained. Therefore, the appropriate selection of one htSNPs to
correctly predict 80% on the rest of non-selected SNPs is
reasonable. It is obvious that the proposed PSO-SVM hybrid
model achieves higher prediction accuracy with fewer selected
htSNPs in the three haplotype datasets. In general, the
prediction accuracy is increased refers to the incremental
selected htSNPs number. From Figure 3 to Figure 5 show that
the numbers of selected htSNPs on haplotype datasets are
proportional to the prediction accuracy and the PSO-SVM
algorithm has very good performance for haplotype tagging
SNPs selection problem in the three testing cases.
64
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
I i 5 6 7
<S ]0 II L? B L4 15 L6 ]7 IS W 20
Number of lilSNfc
Figure 3 The comparison result of prediction accuracy associated with selected
htSNPs on 5q31 datasets.
Figure 4 The comparison result of prediction accuracy associated with
selected htSNPs on TRPM8 datasets.
1
i *
(IJCSIS) International Journal of Computer Science and Information Security,
Vol.8,No.6,2010
VI. CONCLUSION
In this paper, a hybrid PSO-SVM model that combines the
particle swarm optimization (PSO) and support vector machine
(SVM) with feature selection and parameter optimization is
proposed to effectively solve for the haplotype tagging SNP
selection problem. Several public datasets of different sizes are
considered to compare the PSO-SVM with SVM/STSA and
BPSO previously published methods. The experimental results
show that the effectiveness of the proposed approach and the
high prediction accuracy with the fewer number of haplotype
tagging SNP can be obtained by the hybrid PSO-SVM system.
REFERENCES
K. Zhang, M. Deng, T. Chen, M. Waterman and F. Sun, "A dynamic
programming algorithm for haplotype block partitioning," Proc. Natl.
Acad. Sci., vol. 99, pp. 7335-7339, 2002.
K. Zhang, F. Sun, S. Waterman and T. Chen, "Haplotype block partition
with limited resources and applications to human chromosome 21
haplotype data," Am. J. Hum. Genet., vol. 73, pp. 63-73, 2003.
H. Avi-Itzhak, X. Su, and F. de la Vega, "Selection of minimum subsets
of single nucleotide polymorphisms to capture haplotype block
diversity," In Proceedings of Pacific Symposium on Biocomputing, vol.
8, pp. 466-477, 2003.
He Jingwu and A. Zelikovsky, "Informative SNP Selection Methods
Based on SNP Prediction," IEEE Transactions on NanoBioscience, Vol.
6, pp. 60-67, 2007.
Cheng-Hong Yang, Chang-Hsuan Ho and Li-Yeh Chuang, "Improved
tag SNP selection using binary particle swarm optimization," IEEE
Congress on Evolutionary Computation (CEC 2008), pp. 854-860, 2008.
V.N. Vapnik, "The nature of statistical learning theory," New York:
Springer- Verlag, 1995.
E. Halperin, G. Kimmel and R. Shamir, "Tag SNP selection in genotype
data for maximizing SNP prediction accuracy," Bioinformatics, Vol. 21,
pp. 195-203, 2005.
J. Kennedy and R. C. Eberhart, "A discrete binary version of the particle
swarm algorithm," in Proceedings of the World Multiconference on
Systemics, Cybernetics and Informatics, Piscataway, NJ, 1997, pp.
4104-4109.
T
-^-SVMCTSA
— EfSO
1 2 3 4 b G 7 8 9 10 11 12 13 14 li l& 17 1$ 19 20
Nuiiit^j of lilSNFs
Figure 5 The comparison result of prediction accuracy associated with
selected htSNPs on LPL datasets.
[2]
[3]
[4]
[5]
[6]
[7]
[8]
[9]
[10]
J.C. Bean, "Genetics and random keys for sequencing and optimization,"
ORSA J. Comput., Vol. 6, pp. 154-160, 1994.
C.C. Chang, and C.J. Lin, LIB SVM: a library for support vector
machines, Software available at:
http://www.csie.ntu.edu.tw/~cjlin/libsvm, 2001.
65
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
PAPR Reduction Technique for LTE SC-FDMA
Systems Using Root-Raised Cosine Filter
Md. Masud Rana, Jinsang Kim and Won-Kyung Cho
Deptartment of Electronics and Radio Engineering, Kyung Hee University
1 Seocheon, Kihung, Yongin, Gyeonggi, 449-701, Republic of Korea
Email: mamaraece28@yahoo.com
Abstract — Recently, mobile radio communications have de-
veloped rapidly due to the endless demand for broadband
multimedia access and wireless connection anywhere, and any
time. With the emergence of diverse fourth generation (4G)
enabling technologies, signal processing has become ever in-
creasingly important for small power, small chip resources,
and efficient physical implementations of potential multimedia
wireless communication systems. In this paper, we analytically
derive the time and frequency domain single carrier-frequency
division multiplexing (SC-FDMA) signals. Simulation results
show that the SC-FDMA sub-carrier mapping scheme has a
significantly lower peak-to average power ratio (PAPR) compared
to orthogonal frequency division multiplexing (OFDMA). In
addition, the interleave FDMA (IFDMA) sub-carrier mapping
scheme with root raised cosine filter reduced PAPR significantly
than localized FDMA (LFDMA) and distributed (DFDMA) sub-
carrier mapping scheme. As a results, improves the mean power
output from a battery driven terminal equipment and power
amplifier efficiency.
Index Terms— CCDF, IFDMA, OFDMA, PAPR, root-raised
cosine, SC-FDMA.
I. Introduction
The further increasing demand on high data rates in wireless
communication systems has arisen in order to support broad-
band services. The third generation partnership project (3 GPP)
members started feasibility study on the enhancement of the
universal terrestrial radio access (UTRA) in December 2004,
to improve the mobile phone standard to cope with future
requirements. This project was called long term evolution
(LTE) [1].
LTE uses single carrier frequency division multiple access
(SC-FDMA) for uplink transmission and orthogonal frequency
division multiple access (OFDMA) for downlink transmission
[6]. SC-FDMA is a promising technique for high data rate
transmission that utilizes single carrier modulation and fre-
quency domain equalization. Single carrier transmitter struc-
ture leads to keep the peak-to average power ratio (PAPR)
as low as possible that is reduced the energy consumption.
SC-FDMA has similar throughput performance and essentially
the same overall complexity as OFDMA [3], [10], [12]. A
highly efficient way to cope with the frequency selectivity
of wideband channel is OFDMA. OFDMA is an effective
technique for combating multipath fading and for high bit
rate transmission over mobile wireless channels. In OFDMA
system, the entire channel is divided into many narrow sub-
channels, which are transmitted in parallel, thereby increasing
the symbol duration and reducing the intersymbol-interference
(ISI) [4], [8]. Despite many benefits of OFDMA for high speed
data rate services, it suffer from high envelope fluctuation
in the time domain, leading to large PAPR. Because the
high PAPR is detrimental to user mobile equipment (UE)
terminals, SC-FDMA has drawn great attention as an attractive
alternative to OFDMA for uplink data transmission. It can be
regarded as DFT-spread OFDMA (DFTS-OFDM), where time
domain data signals are transformed to frequency domain by
a DFT before going through OFDMA modulation. The main
benefit of DFTS-OFDM compared to OFDM scheme, is re-
duced variations in the instantaneous transmit power, implying
the possibility for increased power- amplifier efficiency, low-
complexity high-quality equalization in the frequency domain,
and flexible bandwidth assignment [12].
In order to solve the high PAPR problem seen in the
uplink of OFDMA, research is now addressing techniques such
as a SC-FDMA. The most of the previous work related to
3 GPP LTE uplink has been mainly focused on implementation
problems in the physical layer [2], [5], [9], [13]. In [10], [12]
proposed raised-cosine pulse shaping method that compare
PAPR characteristics using the complementary cumulative
distribution function (CCDF) for different subcarrier mapping.
PAPR reduction is of the most importance performance
parameter in case of high amplitude signals subject to non
linear power amplification. This situation more and more occur
due to the ever-growing demand in high spectral efficiency
advanced mobile telecommunications systems implying multi
dimensional waveforms considerations for which the PAPR
is high. Pulse shaping is required for a single carrier sys-
tem to bandlimit the transmit signal. This paper addresses
a theoretical analysis of the PAPR reduction of LTE SC-
FDMA systems when root-raised cosine (RRC) filter is used.
RRC is used as the transmit and receive filter in a digital
communication system to perform matched filtering. The
combined response of two such filters is that of the raised-
cosine filter. In this paper, we analytically derive the time and
frequency domain SC-FDMA signals. Simulation results show
66
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
that the SC-FDMA has a significantly lower PAPR compared
to OFDMA system. In addition, we comparing the three forms
of SC-FDMA sub-carrier mapping scheme and find that the
interleave FDMA (IFDMA) sub-carrier mapping with root
raised cosine based pulse shapping method reduced PAPR
significantly than localized FDMA (LFDMA) and DFDMA
sub-carrier mapping scheme. As a results, improves the mean
power output from a battery driven terminal equipment and
power amplifier efficiency.
The rest of the paper is organized as follows. We describes
the 3GPP LTE and LTE SC-FDMA system model in sec-
tion II and III, respectively. In section IV, we describes the
different SC-FDMA sub-carrier mapping scheme. In section
V, we describes the PAPR reduction technique for LTE SC-
FDMA systems. In section VI, we simulated and compare
the proposed method with OFDMA for different sub-carrier
mapping scheme. Finally, conclusions are made in section VII.
II. 3GPPLTE
The main purposes of the 3GPP LTE are substantially
improved end-user throughputs, low latency, reduced user
equipment (UE) complexity, high data rate, and significantly
improved user experience with full mobility. First 3 GPP LTE
and LTE-advanced (LTE-A) specification is being finalized
within 3GPP release 9 and release 10, respectively [1].
One radio frame = 20 Slots =10 Sub-frames = 10 ms
1 Slot = 7 OFDM symbols = 0.5 ms
2 Slots = 1 Sub-frame = 10 ms
TT^
UL
16
17
19
20
; ■
< l : 2
3
4
5
6 ; 7
A A
► Cyclic prefix
M
7 OFDM symbols
••••►
.§ N
S3
CH
j__ Resource block:
Short CP:
7 symbols x 12 sub-carriers
Long CP:
6 symbols x 12 sub-carriers
— 1 resource element
Pilot □
Fig. 2. LTE generic frame structure.
blocks fit in a carrier of 1.4 MHz and 100 resource blocks
fit in a carrier of 20 MHz. Slots consist of either 6 or 7
OFDM symbols, depending on whether the normal or extended
cyclic prefix (CP) is employed. The CP is added in front of
each block. The details transmission scheme parameters of
the 3GPP LTE system are shown in Table I [7]. LTE uses SC-
FDMA scheme for the uplink transmissions and OFDMA in
downlink transmission.
TABLE I
LTE SYSTEM PARAMETERS
2008
2009
2010
2011
-*- Release 7 study phase (HSPA+)
► Release 8 work phase (LTE)
► Release 9 test specs
t t t
-► Release 10 (LTE
advanced)
Release 6
1 1 SPA uplink
Core First
drafted specs
Fig. 1. LTE release timeline.
Specifically, the physical layer has become quite stable re-
cently for a first implementation. LTE supports multipule input
multiput output (MIMO) with one, two, four, and eight antenna
elements at base station (BS) and mobile terminal. Both closed
and open loop MIMO operation is possible. The target of LTE-
A is to reach and surpass the international telecommunication
union (ITU) requirements. One of the important LTE-A ben-
efits is the ability to leverage advanced topology networks;
optimized heterogeneous networks with a mix of macros with
low power nodes such as picocells, femtocells, ensures user
fairness, worldwide roaming, and new relay nodes [1].
In 3GPP LTE, the basic unit of a transmission scheme is a
radio frame which is ten msec long. They are divided into ten
sub-frames, each sub-frame one msec long. Each sub-frame is
further divided into two slots, each of half msec duration. Fig.
2, shows the basic LTE generic frame structure [11]. The sub-
carrier spacing in the frequency domain is 15 kHz. Twelve
of these sub-carriers together (per slot) is called a resource
block therefore one resource block is 180 kHz. Six resource
Trans, bandwidth (MHz)
1.25
2.5
5
10
15
20
FFT size
128
256
512
1024
1536
2048
Occupied sub-carrier
76
151
301
601
901
1200
Sampling frequency (MHz)
1.92
3.84
7.68
15.36
23.04
30.72
No. of available PRBs
6
12
25
50
75
100
User plane latency (ms)
< 5
PRB bandwidth (kHz)
180
Frame duration (ms)
0.5
Sub-carrier bandwidth (kHz)
15
Coverage (km)
5-30
Mobility (km/hr)
15-350
Peak data rates (Mbits/s)
DL: 100, and UL: 50
Antenna configuration
DL: 4x2, 2x2, 1x2, lxl, and UL: 1x2, lxl
Spectrum efficiency
DL: 3-4 x HSDPA, and UL: 2-3 x HSUPA Rel.6
Control plane latency (ms)
100 (idle to active), and 50 (dormant to active)
Radio resource
DL: 3-4 fold higher than Rel.6
III. LTE SC-FDAMA SYSTEM MODEL
The basic principle of a LTE SC-FDMA transmission sys-
tem is shown in Fig. 3.
At the transmitter side, a baseband modulator transmits the
binary input to a multilevel sequences of complex number
mi (q) in one of several possible modulation formats including,
quandary phase shift keying (QPSK), and 16 level-QAM.
These modulated symbols are perform a N-point discrete
Fourier transform (DFT) to produce a frequency domain
representation [3]:
si(ra)
1
N-l
^rai(g)e
(1)
q=0
where mi is the discrete symbols, q is the sample index, j is
the imaginary unit, and mi(q) is the data symbol. The output
67
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
Input data
Output data
Modulation
X
Demodulation
mi(q)
Size-N DFT
I R(k)
Size-N IDFT
si(n)
Subcarrier
mapping
4 s 3( m )
Equalization
S2(q)
X
Subcarrier
demapping
Size-M IDFT
Channel
Noise
s(m)
r(m)
Size-M
DFT
Cyclic
prefix(CP)
insertion
i
■^
Remove CP
Fig. 3. LTE SC-FDMA transceiver system model [6].
of the DFT is then applied to consecutive inputs of a size-M
inverse DFT (M >N) and where the unused inputs of the IDFT
are set to zero. If they are equal (M=N), they simply cancel
out and it becomes a conventional single user single carrier
system with frequency domain equalization. However, if N is
smaller than M and the remaining inputs to the IDFT are set
to zero, the output of the IDFT will be a signal with 'single-
carrier' properties, i.e. a signal with low power variations, and
with a bandwidth that depends on N. The SC-FDMA is single-
carrier, not single frequency. The data signal of each user
consists of a lot of frequency. DFT of SC-FDMA is used to
filter the frequency items and maps them into IDFT to reform
single user waveform. This may justify the reduced peak- to-
average power ratio (PAPR) experienced in the IDFT output.
The details description of the sub-carrier mapping mode are in
section IV. PAPR is a comparison of the peak power detected
over a period of sample occurs over the same time period. The
PAPR of the transmit signal is defined as [14]:
max 0<m<T |s(ra)| 2
PAPR
l_ f TN
(2)
w . \s(m)\ 2 dm
where T is the symbol period of the transmitted signal s(m).
PAPR is best described by its statistical parameter, com-
plementary cumulative distribution function (CCDF). CCDF
measures the probability of signal PAPR exceeding certain
threshold [12], [14]. To further reduce the power variations
of the DFTS-OFDM signal, explicit spectrum shaping can be
applied. Spectrum shaping is applied by multiplying the fre-
quency samples with some spectrum- shaping function, e.g. a
root-raised-cosine function (raised-cosine-shaped power spec-
trum). The IDFT module output is followed by a CP insertion
that completes the digital stage of the signal flow. A CP is
used to eliminate ISI and preserve the orthogonality of the
tones. Assume that the channel length of CP is larger than the
channel delay spread [8].
The transmitted symbols propagating through the radio
channel can be modeled as a circular convolution between the
channel impulse response (CIR) and transmitted data blocks.
At the receiver, the opposite set of the operation is performed.
The CP samples are discarded and the remaining N samples
are processed by the DFT to retrieve the complex constellation
symbols transmitted over the orthogonal sub-channels. The
received signals are de-mapped and equalizer is used to
compensate for the radio channel frequency selectivity. After
IDFT operation, the corresponding output is demodulated and
soft or hard values of the corresponding bits are passed to the
decoder.
IV. SC-FDMA SUB-CARRIER MAPPING SCHEME
There are two principal sub-carrier mapping modes-
localized mode, and distribution mode. An example of SC-
FDMA transmit symbols in the frequency domain for two
user, three sub-carrier per user and six sub-carriers in total
is illustrated in Fig. 4 [6].
si(l)
L
Li 1 i T i ttt
In tttttt
Zeros
Complex weight
si(2)
(x)
(y)
(z)
si(3)
.
ttrttrtrt
User 1
User 3
User 2
Time domain
Frequency domain ■
Fig. 4. Multiple access scheme of SC-FDMA: (x) IFDMA mode, (y)
DFDMA mode, and (z) LFDMA.
In distributed sub-carrier mode, the outputs are allocated
equally spaced sub-carrier, with zeros occupying the unused
sub-carrier in between. While in localized sub-carrier mode,
the outputs are confined to a continuous spectrum of sub-
carrier [10], [12]. Except the above two modes, interleaved
sub-carrier mapping mode of SC-FDMA (IFDMA) is another
special sub-carrier mapping mode. The difference between
DFDMA and IFDMA is that the outputs of IFDMA are
allocated over the entire bandwidth, whereas the DFDMA' s
outputs are allocated every several sub-carriers. If there are
more than one user in the system, different sub-carrier map-
ping modes give different sub-carrier allocation [10], [12].
In order to accommodate multiple access to the system and
to preserve the constant envelope property of the signal, the
68
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
elements of transmitted signal s(m) are mapped, by using the
LFDMA or IFDMA sub-carrier mapping rule.
Here is output symbol calculation in IFDMA in time
domain. The frequency samples after IFDMA sub-carrier
mapping is [12], [16]:
i(q/C) = l 0<n<N-l
else
s i(q)
where q = Cn, and C = M/N and it is the bandwidth
expansion factor of the symbol sequence. If N = M/C
and all terminals transmit N symbols per block, the system
can handle C simultaneous transmissions without co-channel
interference. After M-point IDFT operation (M > N), time
domain signal can be described as follows. Let m = Nc + q,
where < c < C — 1 and < q < N — 1. The time domain
IFDMA transmitted signal can be express as [12], [16]:
M-
s(m)
S2(q)e~
-y
N-l
^si(n)
1
~NC
n=0
1 1 N ~ 1
y si[n)e N
C N
n=0
N-l
1 r 1 ^— V , . J27rqn
C V N
(3)
n=0
The square backed [.] of the above equation represent the N-
point IDFT operation. The above equation can be rewritten
as
1
s ( m ) = g m i(q)i
(4)
is the N-point IDFT
where rra(q) = ± E n =o si(n)e'
operation. Therefore in case of IFDMA, every output symbol is
simple a repeat of input symbol with a scaling factor of 1/C in
time domain. When the sub-carrier frequency allocation starts
from fth sub-carrier i.e. q = Cn + / then the frequency
samples after IFDMA sub-carrier mapping is
i(q/C-f) = l 0<n<N-l
else
s 2(q)
and the time domain transmitted signal can be express as
1 r 1
s(m)
N-l
C [ N ^ 5 i( n ) e ^
n=0
1 22,
— e *
C
~mi(q).
(5)
Thus, when the sub-carrier frequency allocation starts from
fth instead of zero then there is an additional phase rotation
of e J2irmf/M^
Here is output symbol calculation in LFDMA sub-carrier
mapping in time domain. After DFT operation the frequency
domain sub-carrier mapping signal can be written as
si(l) = 1 0<Z <N -1
N <l< M -1
s 2(q)
After IDFT operation, time domain signal can be described
as follows. Let m = Cq + c, where < c < N — 1 and
< c < C — 1. Then time domain transmitted signal can be
represent as [12], [16]:
M-l
5 ( m ) = m Yl s ^
j27rml
e M
1 1
CN
N-l
£ s i(o
j2n(Cq + c)l
e CN
(6)
if c = 0, then
1 1 N ~ 1
5 ( m ) = 7^77 ^2 s i( l ) e
C l N
mi(q)
(7)
Since mi(q) = ^ p=0 mi(p)e n , and if c ^ 0, then
N-l N-l
j27Tlp j27T(Cq + c)l
CN
5 ^ = ~nc E£ mi W e ^ P ] e
1=0 p=0
N-l N-l
= 4n E E m 1 (p)e^ 2 -^-^/ JV+c / cw )'
1 _ e J2K(q-p) e J2irc/C)
NC
1
Z=0 p=0
iV-1
NC 2^ mi W ! _ e (j2ir(q-p)/N+c/CN)
1 _ e J27Tc/C
p=0
N-l
^E^w r
c
(1-e-
7VC ^ "" 1V/ ' / 1 - e U^{q-p)/N+c/CN)
p=0
N-l
J2nc/C\±_ V^ _
J N ^ 1
n=0
™>l(p)
D (j27r(q-p)/N+c/CN)
(8)
So, the time domain LFDMA sub-carrier mapping signal has
exact copies of input time signals with a scaling factor of 1/C
in the N-multiple sample positions and in between values are
sum of all the time input symbols in the input block with
different complex- weighting (CW) [12].
Here is output symbol calculation in DFDMA in time
domain. The frequency samples after DFDMA sub-carrier
mapping is [12], [16]:
, x f sAq/C) = 1 0<n<N -1
52 ^ ) = {o else ~
where < c < C - 1, q = Cn, and < C < C. Let
m = Nq + c where 0<c<C-landO<g<iV- land
< q < N — 1 . The time domain DFDMA transmitted signal
can be express as [12], [16]:
M-l
5 0) = m Z^ S2 ( q ) e M
1=0
1 1 N ~ 1
1 1 V^ /7\ 327r(C q + c)l ~
CN^ Sl(l)e CN ° l
1=0
(9)
69
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
if c = and according to the previous procedure, we obtain
TABLE II
THE SYSTEM PARAMETERS FOR SIMULATIONS
(10)
Since mi(q) = ^2 p=0 mi(p)e n , if c ^ and according
to the previous procedure, we obtain
7V-1
sim) = ^l-e^'%Y.- l
mi(p)
n=0
e (j2Tr(Cq-p)/N+Cc/CN)
^
So, the time domain symbols of DFDMA sub-carrier mapping
have the same structure as those of LFDMA sub-carrier
mapping.
V. PAPR REDUCTION TECHNIQUE FOR LTE SC-FDMA
SYSTEMS
In case of high amplitude signals subject to non linear
power amplification, PAPR reduction is one of the most
importance performance parameter. This situation more and
more occur due to the ever-growing demand in high spectral
efficiency telecommunications systems implying multi dimen-
sional waveforms considerations for which the PAPR is high.
In a single-carrier communication system, pulse shaping is
required to bandlimit the signal and ensures it meets the
spectrum mask. In this paper, a root raised cosine (RRC) filter
is used to pulse shape the SC-FDMA signals. RRC is used
as the transmit and receive filter in a digital communication
system to perform matched filtering. The combined response
of two such filters is that of the raised-cosine filter. The raised-
cosine filter is used for pulse-shaping in digital modulation
due to its ability to minimize intersymbol interference (ISI).
Its name stems from the fact that the non-zero portion of the
frequency spectrum of its simplest form (3 = 1 (called the
roll-off factor) is a cosine function, 'raised' up to sit above
the / (horizontal) axis. The RRC filter is characterized by
two values; /3, and T s (the reciprocal of the symbol-rate). The
impulse response of such a filter can be given as:
1-/3 + 4/3/tt,£ =
P/V2[(l + 2/tt) sin(7r//54) + (1 - 2 /it) cos(tt//34)],
t = ±T s //?4
sin[7rt/T s S (l-/3)]+4/3t/T s cos[7rt/T 5 (l+/3)] ,
7rt/T s [l-(4/3t/T s )2]) i eibe
It should be noted that unlike the raised-cosine filter, the
impulse response is not zero at the intervals of ±T S . However,
the combined transmit and receive filters form a raised-cosine
filter which does have zero at the intervals of ±T S . Only in
the case of (3 = 0, does the root raised-cosine have zeros at
±T a .
VI. PERFORMANCE ANALYSIS
The performance of the aforementioned PAPR reduction
technique is explored by performing extensive computer sim-
ulations. All simulation parameters of the LTE SC-FDMA
systems are summarized in Table II [6].
The CCDF)of PAPR, which is the probability that PAPR
is higher than a certain PAPR value PAPRO, is calculated by
Monte Carlo simulation. We compare the PAPR value that is
h(t)
System parameters
Assumptions
System bandwidth
hMHz
Number of sub-carriers
512
Data block size
16
Roll of factor
0.0999999999
Overs ampling factor
4
Number of iteration
10 4
Sub-carrier mapping schemes
DFDMA, IFDMA, LFDMA
Modulation data type
Q-PSK and 16-QAM
Spreading factor for IFDMA
32
Spreading factor for DFDMA
31
exceeded with probability less than 0.1 percentile PAPR. The
PAPR calculation using various sub-carrier mapping schemes
for SC-FDMA and OFDMA system is shown in Fig. 5. The
modulation scheme used for the calculation of PAPR is QPSK.
It can be seen that SC-FDMA sub-carrier mapping schemes
o" 1
X \
\ v \ \
\ \ Y \\ \ :
\ \ \\ A \
\ \ \\ \ \
\ \ y \ \
^
- - Proposed DFDMA
— Existing DFDMA
Proposed LFDMA
Existing LFDMA
Proposed IFDMA
Existing IFDMA
OFDMA
\ \ \\ \ -
M V \\ \
i \ \ M \
\ \ \ l\ \
u
i Kl \ :
3 4 5 6 7
PAPRo dB
Fig. 5.
QPSK.
Comparison of CCDF of PAPR for SC-FDMA with OFDMA using
gives lower PAPR values as compared to OFDMA scheme.
In addition, the root raised cosine pulse shaping method has
lower PAPR than the case of existing pulse shapping method
by more than 3dB. Due to the complex- weighting of LFDMA
and DFDMA equation would increase the PAPR. Due to the
phase rotation it is unlikely that the LFDMA samples will
all add up constructively to produce a high output peak after
pulse shaping. But it is shown that LFDMA has a lower PAPR
than DFDMA when pulse shaping is applied. Another PAPR
simulation using various sub-carrier mapping schemes for LTE
SC-FDMA systems is shown in Fig. 6. The modulation scheme
used for the calculation of PAPR is 16-QAM. It show that
IFDMA has lowest value of PAPR at 7.8dB which is 6.7dB in
case of QPSK as modulation technique. Finally, we conclude
70
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
o" 1
:
\
o- 2
o- 3
- - Proposed DFDMA
— Existing DFDMA
Proposed LFDMA
Existing LFDMA
Proposed IFDMA
Existing IFDMA
— OFDMA
\ \ M \
\ \ \ \\\ \
l \ \ -\\ \
i \ \ iu \
\ \ \ \\ \ "
i \ \ \ \ ;
i V \ 1 \
1 \ v 1
|_ 1 iV 1
1 ' 1
Fig. 6. Comparison of CCDF of PAPR for SC-FDMA with OFDMA using
16-QAM.
that the higher values of PAPR by using 16-QAM which is
undesirable because they cause non linear distortions at the
transmitter.
VII. Conclusions
The efficiency of a power amplifier is determined by the
PAPR of the modulated signal. In this paper, we analysis
different sub-carrier mapping scheme for LTE SC-FDMA
systems. We derive the time and frequency domain signals
of different sub-carrier mapping scheme, and numerically
compare PAPR characteristics using CCDF of PAPR. We come
to the conclusion that the IFDMA sub-carrier mapping with
RRC pulse shaping method has lowest PAPR values compare
to the other sub-carrier mapping methods. As a results, im-
proves the mean power output from a battery driven terminal
equipment and power amplifier efficiency. Therefore, SC-
FDMA is attractive for uplink transmissions since it reduces
the high PAPR seen with OFDMA.
[3] B. Karakaya, H.Arslan, and H. A. Cirpan, "Channel estimation for LTE
uplink in high doppler spread," Proc. WCNC, pp. 1126-1130, April 2008.
[4] J. Berkmann, C. Carbonelli, F.Dietrich, C. Drewes, and W. Xu, "On
3G LTE terminal implementation standard, algorithms, complexities and
challenges," Proc. Int. Con. on Wireless Communications and Mobile
Computing, pp. 970-975, August 2008.
[5] A. Ancora, C. Bona, and D.T.M. Slock, "Down-sampled impulse response
least-squares channel estimation for LTE OFDMA," Proc. Int. Con. on
Acoustics, Speech and Signal Processing, Vol. 3, pp. 293-296, April 2007.
[6] M. M. Rana, M. S. Islam, and A. Z. Kouzani, "Peak to average power ratio
analysis for LTE aystems," Proc. Int. Con. on Communication Software
and Networks, pp. 516-520, February 2010.
[7] A. Ancora, C. B. Meili, and D. T. Slock, "Down-sampled impulse
response least- squares channel estimation for LTE OFDMA," Proc. Int.
Con. on Acoustics, Speech, and Signal Processing, April 2007.
[8] L. A. M. R. D. Temino, C. N. I Manchon, C. Rom, T. B. Sorensen, and
P. Mogensen, "Iterative channel estimation with robust wiener filtering in
LTE downlink," Proc. Int. Con. on Vehicular Technology Conference, pp.
1-5, September 2008.
[9] J. Zyren, "Overview of the 3GPP long term evolution physical layer," Dr.
Wes McCoy, Technical Editor, 2007.
[10] H. G. Myung, J. Lim, and D. J. Goodman, "Single carrier FDMA for
uplink wireless transmission," IEEE Vehicular Technology Magazine, vol.
1, no. 3, pp. 30-38, September 2006.
[11] M. Noune and A. Nix, "Frequency-domain precoding for single carrier
frequency- division multiple access," IEEE Commun. Magazine, vol. 48,
no. 5, pp. 86-92, May 2010.
[12] H.G. Myung, J. Lim, and D. J. Goodman, "Peak to average power ratio
for single carrier FDMA signals," Proc. PIMRC, 2006.
[13] S. Maruyama, S. Ogawa, and K.Chiba, "Mobile terminals toward LTE
and requirements on device technologies," Proc. Int. Con. on VLSI
Circuits, pp. 2-5, June 2007.
[14] S. H. Han, and J. H. Lee, "An overview of peak to average power ratio
reduction techniques for multicarrier transmission," IEEE Transction on
Wireless Communications, April 2005.
[15] H. G. Myung, "Introduction to single carrier FDMA," Proc. Int. Con.
on European Signal Processing (EUSIPCO), Poznan, Poland, September
2007.
[16] A. Sohl, and A. Klein, "Comparison of localized, interleaved and block-
interleaved FDMA in terms of pilot multiplexing and channel estimation,"
Proc. Int. Con. on PIMRC, 2007.
Acknowledgment
This research was supported by the Basic Science Research
Program through the National Research Foundation of Korea
(NRF) funded by the Ministry of Education, Science and
Technology (20100017118).
References
[1] Q. Li, G. li, W. Lee, M. II. Lee, D. Clerckx, and Z. li, "MIMO techniques
in WiMAX and LTE: a feature overview," IEEE Commun. Magazine, May
2010.
[2] E. Dahlman, S. Parkvall, J. Skold, and P. Beming, "3G evolution HSPA
and LTE for mobile broadband," Academic Press is an Imprint of Elsevier,
2007.
71
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
Survey of Routing Protocols and Channel
Assignment protocols in Wireless Mesh Networks
Vivek M R athod, Suhas J M anangi, S atish E , S a u m y a H e g d e
National Institute of Technology Karnataka - Surathkal
Abstract: This paper is a survey on wireless mesh networks.
Here we mention the basics of wireless mesh network, their
purpose, channel assignment techniques and routing
protocols. This survey is prepared towards helping those
working on the relationship between channel assignment
and routing protocols.
Keywords: Wireless Mesh Networks, Routing protocols,
Channel Assignment, Multi Hop, Multi Radio.
I.
INTRODUCTION
[1] : c
A wireless mesh network (WMN) is a communication
network made up of radio nodes organized in a mesh
topology. The nodes which constitute the WMN are in
adhoc mode so as to realize mesh topology.
Wireless mesh architecture is an effort towards providing
high-bandwidth network over a specific coverage area.
Wireless mesh architecture's infrastructure is, in effect, a
router network minus the cabling between nodes. It's built
of peer radio devices that don't have to be cabled to a
wired port like traditional WLAN access points (AP) do.
The traditional WLANs consist only of single hop end-to
end connection (i.e., between the client and access point).
In contrast, Mesh architecture sustains signal strength by
breaking long distances into a series of shorter hops.
Intermediate nodes not only boost the signal, but
cooperatively make forwarding decisions based on their
knowledge of the network, i.e. perform routing. Such
architecture may with careful design provide high
bandwidth, spectral efficiency, and economic advantage
over the coverage area.
This paper is organized in the following sections.
1. Types of wireless mesh networks (network
architectures).
2. Essential characteristics of WMN.
3. Components of WMN and their alternatives.
4. Routing purposes, problems and protocols.
5. Areas for research.
II. NETWORK ARCHITECTURE
The types of network structures being used for WMNs can
be classified into three types in a very broad sense.
1. Client wireless mesh networks
2. Infrastructure wireless mesh networks
3. Hybrid wireless mesh networks [8] .
A Client WMN:
Client mesh networks or simply ad-hoc networks are formed
by client devices with no supporting fixed infrastructure.
Each node plays same role and participates in packet
forwarding.
B. Infrastructure WMN:
In contrast to client WMN, infrastructure WMN consists of
routers and client devices. The routers are interconnected
via wireless links to form a multi-hop backhaul
infrastructure. One or more routers are connected to the
wired network and are called gateways of the WMN.
Generally mesh router has two or more radio interfaces. One
of which is an access interface for the clients to access the
network. The second radio interface is a relay interface for
forwarding and routing data packets. This is basically used
for inter-router communication. Client devices associate
themselves with the nearest mesh router to access the
network. They don't participate in routing or relaying of
packets. Therefore even if two clients are within the wireless
range of each other, they cannot directly communicate. It
has to happen through their respective routers.
C. Hybrid WMN [8] :
This architecture is the combination of Infrastructure and
client meshing; clients can access the network through
mesh routers as well as by directly meshing with other
mesh clients. A hybrid WMN is an extension to the
Infrastructure WMN. In a hybrid WMN the clients not
only connect to the wireless backhaul, but also serve as a
gateway to for the clients which are located too far from
the wireless mesh router. Therefore a hybrid WMN is
more robust and more scalable than the previous two. A
well-built hybrid WMN would enable fast, cheap and easy
deployment of networks, leading to interesting
applications such as emergency networks.
III. ESSENTIAL CHARACTERISTICS OF WMN
WMN has mainly following essential characteristics:
1. Multihop and Multi-pathing: multiple paths
between two point in a WMN leads to the increase
in bandwidth. This increase in the bandwidth is
because the RTT (round trip time) for shorter paths
(hops) is less than that of a single end to end path.
Multiple packets can travel simultaneously
72
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
between two ends. Multi-pathing strips the data to
be sent to a destination and sends these chunks via
multiple paths, which increases the throughput
significantly. It also provides robustness to paths,
because there is always an alternative unless the
destination itself is not connected to the network.
2. Self-healing, self-forming and self-organising:
Since most of the nodes of the WMN are mobile,
the WMN is always aware of its surroundings. It
dynamically changes the routing paths based on the
current state of the network. If a participating node
quits, the network is reconfigured so as to keep the
remaining nodes connected. Similarly the dynamic
changes in the network must also take place based
on the network traffic at different routes.
3. Compatibility and interoperability: The WMNs
built on the IEEE 802.11 standards must be capable
of supporting conventional Wi-Fi clients.
4. Cost Factor: WMNs can be very cost effective
because we can build and configure a WMN with
minimal existing resources. A WMN could provide
an effective and good internet bandwidth to the
group of users who share a single internet link.
IV. COMPONENTS OF WIRELESS MESH
NETWORKS
A WMN consists of two types of wireless nodes. Mesh
Routers and Mesh Clients. The Mesh Routers have
improved computational, communication and power
resources as compared to Mesh Clients. Mesh Routers are
generally static and form the multi-hop backhaul network
with optional access to other auxiliary networks. In addition,
Mesh Routers are also typically equipped with multiple
wireless network interfaces (IEEE 802.11 [3] ) and are
therefore able to establish high capacity connections. Mesh
Clients are mobile devices, which take advantage of the
existing communication infrastructure provided by the Mesh
Routers.
A 1-Radio VS Multi-Radio Approaches:
In 1-radio approach the participating nodes have only one
radio each. Consider a network where both the clients and
the mesh routers have only one radio, and then mesh
routers would not be able to listen to the backhaul and the
client simultaneously. Collisions would be very frequent.
This will result in very low throughput. Thus one radio
WMN is inferior to multi-radio infrastructure mesh
networks in Multihop situations. In the case of 1 radio ad
hoc mesh networks, available bandwidth is reduced by
50% with each hop: bandwidth available at the 3rd hop is
1/8 of the available capacity. However, while one-radio ad
hoc mesh networks are unsuitable for Multihop situations,
they are useful in one-hop situations for quickly
establishing p2p communications. Conversely, 2-radio
infrastructure meshes are ideal for Multihop situations
with no restriction on the number of hops. Thus One radio
mobile client mesh network combined with two or more
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
radio routers' backhaul support provides the best hybrid
WMN; ubiquitous connectivity but with multiple levels of
redundancy built in.
V. MULTI RADIO MESH ROUTER
There are mainly 4 types of Multi Radio Mesh Routers:
1. Single unit mesh router
2. split wireless router
3. Multi-Channel Multi-Transceiver single radio
4. low cost mesh router construction
A Single Unit Mesh Router:
Single unit mesh router is a single package with multiple
radios in it. All these radios' operate in non overlapping
channels. Some of these could be used to relay packets
between routers, while the others to provide connectivity to
the clients or client adhoc network. Even though the radios
operate in non-overlapping channels, the practical results
have shown that there is a significant amount of interference
between them due to the near-field effect, resulting in
reduced throughput
B. Split Wireless Router:
Split mesh router is a network (wired) of two or more
single radio routers. This design has gained motivation
from the limitations of the single unit multi-radio routers.
We refer to the single radio routers which are part of split
router as nodes hereafter.
The commercially available single-radio routers often
provide multiple interface technologies like the Ethernet,
fiber or ATM. Two or more such units are connected via a
backhaul using one of the available interface options like
the Ethernet. Since the separation between these nodes is
determined by the cable length forming the backhaul, the
interference can be significantly reduced by increasing the
distance between them. This is an effective solution for
the interference due to near-field effect in the single unit
mesh router.
Since our mesh router unit is a combination of 3-single
radio routers, we need a software abstraction by which the
assembly appears like a single unit to the network. Each
single radio router must here be aware of the neighbors of
the other two.
C. Multi-Channel Multi Transceiver Single Radio [4] :
In this kind of routers, a radio includes multiple parallel RF
front-end chips and baseband processing modules to support
several simultaneous channels. On top of the physical layer,
only one MAC layer module is needed to coordinate the
functions of multiple channels. So far no multi-channel
multi-transceiver MAC protocol has been proposed for
WMNs.
D. Low Cost Mesh Router Construction
A low cost router can be set up using two USB or PCI radio
cards on a low-end computer. But this would also require a
MAC layer which supports multiple NICs simultaneously.
73
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
VI. CHANNEL ASSIGNMENT IN MULTI RADIO
ROUTERS [2] [4]
In case of a two radio router network, there is no much
flexibility, because one of the radios is used to
communicate with other mesh routers on same channel
and the other radio is used to communicate with clients.
But if we have more than two radios on each mesh router,
we could use one for communicating with clients and the
other radios could be intelligently assigned different
channels so that they form channel diversified routes
among mesh routers. Whether or not two routers are
neighbors is decided by the channels assigned to them.
The channel assignment can be made based on link quality
and the topology.
There are various algorithms proposed for the channel
assignment problem. They can be classified into two
categories;
1. Interference-aware channel assignment(IACA)
2. Traffic-aware channel assignment (TACA)
Some of the algorithms are:
1. Identical channel assignment
2. Hybrid channel assignment
3. Centralized channel assignment
4. Maxflow based channel assignment routing
(MCAR)
5. Topology and interference-aware channel
assignment (TIC)
A Identical Channel Assignment:
In this method first radio is assigned channel 1, second is
assigned next non overlapping channel and so on. Though
this preserves connectivity, this method in no way makes
any effort in reducing interference.
B. Hybrid Channel Assignment:
In this strategy some radios are statically assigned channels
while other radios are assigned channels dynamically.
C. Centralized Channel Assignment:
In this method the links are visited in some order and a
common channel is assigned to interfaces on both the ends.
If all the interfaces of the end node are already assigned a
channel and they don't share any common channel, then it is
necessary to replace one on the channel assignments. This
ends up in a recursive channel assignment procedure. The
visit can be in the decreasing order of the number of links
falling in the interference range and the least used channel in
that range is selected (interference aware). It could also be
based on the estimated link flow rates(traffic aw are). The
algorithm might then visit all the links in decreasing order of
expected link flow rate and select the channel which
minimizes the sum of expected flow rates of all the links in
the interference region that are assigned the same channel.
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
D. Maxflow Based Channel Assignment Routing:
MCAR is an improvement over centralized channel
assignment algorithm. The interdependence among channel
assignments across the whole network is taken into account
by first identifying the groups of links that need to be
assigned the same channel in order for the number of
different channels on every router not to exceed the number
of radios. Then, the actual channel assignment stage exploits
the result of the first stage to assign channel in such a way
that no replacement of previously assignments are
necessary.
E. Topology and Interference aware Channel Assignment:
This algorithm undergoes two phases. One is Topology
discovery and the other is channel selection.
Topology discovery: Prior to the channel assignment the
topology is discovered. Topology discovery for every
router is the identification of band-specific set of
neighboring routers and the measurement of quality of
link to each of these neighbors. Each router tunes itself to
various channels on which band topology is to be
discovered. This activity is co-ordinated by the channel
management server. The link quality is measured by ETT
(estimated transmit time).
Channel selection: Dijkstra's shortest path algorithm is
used in TIC to discover frequency-diversified routes
between the gateway and routers. The interference
between mesh links is generated using conflict- graph
model. For generating the above model interfering mesh
links have to be identified in the first place. Thus the data
generated in the first phase (topology discover) can be
used to construct conflict graph. Thus the interfering links
are assigned non-overlapping channels.
Cross-layer work: In most of the situation the throughput
of configured WMN depends on both the channel
assignment and routing algorithm chosen. So there is a lot
of research in developing the cross-layer protocols which
deals with the channel assignment and routing jointly.
VII. Routing Protocols [9]
Routing protocols lie at the heart of designing a WMN
network. They, in simple terms, specify the relay routes for
packets in the network. Most of the protocols neglect the
traffic between the mesh nodes and only consider the
traffic between the nodes and the internet.
Network Asymmetry: This is the situation in which
forward direction of a network is significantly different
from the reverse direction in terms of bandwidth, loss rate,
and latency. Forward path routing protocols are effective
in routing of the packets from the mesh nodes to the
gateway of the WMN, backward routing protocols are
effective in routing the packets from the internet to the
mesh nodes.
Some of the most popular protocols being used are AODV
and OLSR.
74
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
A. Ad-hoc On Demand Distance Vector Routing (AODV):
The Ad hoc On Demand Distance Vector (AODV) routing
algorithm is a routing protocol designed for ad hoc mobile
networks. AODV is capable of both unicast and multicast
routing. It is an on demand algorithm, meaning that it
builds routes between nodes only as desired by source
nodes. It maintains these routes as long as they are needed
by the sources. Additionally, AODV forms trees which
connect multicast group members. The trees are composed
of the group members and the nodes needed to connect the
members. AODV uses sequence numbers to ensure the
freshness of routes. It is loop-free, self-starting, and scales
to large numbers of mobile nodes.
AODV builds routes using a route request / route reply
query cycle. When a source node desires a route to a
destination for which it does not already have a route, it
broadcasts a route request (RREQ) packet across the
network. Nodes receiving this packet update their
information for the source node and set up backwards
pointers to the source node in the route tables. In addition
to the source node's IP address, current sequence number,
and broadcast ID, the RREQ also contains the most recent
sequence number for the destination of which the source
node is aware. A node receiving the RREQ may send a
route reply (RREP) if it is either the destination or if it has
a route to the destination with corresponding sequence
number greater than or equal to that contained in the
RREQ. If this is the case, then it unicast a RREP back to
the source. Otherwise, it rebroadcasts the RREQ. Nodes
keep track of the RREQ's source IP address and broadcast
ID. If they receive a RREQ which they have already
processed, they discard the RREQ and do not forward it.
If the RREP propagates back to the source, then nodes set
up forward pointers to the destination. Once the source
node receives the RREP, it may begin to forward data
packets to the destination. If the source later receives a
RREP containing a greater sequence number or contains
the same sequence number with a smaller hop count, it
may update its routing information for that destination and
begin using the better route.
As long as the route remains active, it will continue to be
maintained. A route is considered active as long as there
are data packets periodically travelling from the source to
the destination along that path. Once the source stops
sending data packets, the links will time out and
eventually be deleted from the intermediate node routing
tables. If a link break occurs while the route is active, the
node upstream of the break propagates a route error
(RERR) message to the source node to inform it of the
now unreachable destination(s). After receiving the
RERR, if the source node still desires the route, it can
reinitiate route discovery.
Advantage and disadvantages:
The main advantage of this protocol is that routes are
established on demand and destination sequence numbers
are used to find the latest route to the destination. The
connection setup delay is less. One of the disadvantages of
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
this protocol is that intermediate nodes can lead to
inconsistent routes if the source sequence number is very
old and the intermediate nodes have a higher but not the
latest destination sequence number, thereby having stale
entries. Also multiple RouteReply packets in response to a
single RouteRequest packet can lead to heavy control
overhead. Another disadvantage of AODV is that the
periodic beaconing leads to unnecessary bandwidth
consumption.
B. Optimized Link State Routing Protocol(OLSR):
It is a proactive protocol. The Optimized Link State
Routing Protocol (OLSR)
[1] [5]
is a proactive routing
protocol. Every node sends periodically broadcast
"Hello"-messages with information to specific nodes in
the network to exchange neighborhood information. The
information includes the nodes IP, sequence number and a
list of the distance information of the nodes neighbors.
After receiving this information a node builds itself a
routing table. Now the node can calculate with the shortest
path algorithm the route to every node he wants to
communicate. When a node receives an information
packet with the same sequence number twice he is going
to discard it. In these routing tables he stores the
information of the route to each node in the network. The
information is only updated:
1. A change in the neighborhood is detected.
2. A route to any destination is expired.
3. A better (shorter) route is detected for a
destination.
The difference from OLSR to LSR (Links State Protocol)
is that OLSR relies on multi-point relays (MPR). MPR is a
node which is selected by its direct neighbor (one
hop).The first idea of multipoint relays is to minimize the
flooding of broadcast messages in the network. An
information packet should not be sent twice in the same
region of the network. MPR helps to optimize and reduce
that problem. Each node informs its direct neighbors (one
hop) about its MPR set in the "Hello"-messages. After
receiving such a "Hello"-message, each node records the
nodes MPR Selector that selects it as one of their MPRs.
The second idea is that the size of the hello messages is
reduced. It includes only the neighbors that select node N2
as one of their MPR nodes. In this way partial topology
information is propagated. Node N2 can be reached only
from its MPR selectors.
Advantages:
1. Minimal latency.
2. Ideal in high density and large networks.
3. OLSR achieves more efficiency than classic LS
algorithms when networks are dense.
4. OLSR avoids the extra work of "finding" the
destination by retaining a routing entry for each
destination all the time, thus providing low
single-packet transmission latency.
5. OLSR can easily be extended to QoS monitoring
by including bandwidth and channel quality
75
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
information in link state entries. Thus, the quality
of the path (e.g., bandwidth, delay) is known
prior to call setup.
Disadvantages:
1. When the network is sparse, every neighbor of a
node becomes a multipoint relay. The OLSR then
reduces to a pure LS protocol.
2. High control overhead (reduced by MPR usage) .
3. Higher computation.
4. Storage.
5. Implementation complexity.
C. Backward Routing Protocol:
The above mentioned protocols are designed to route in
the direction from the mesh nodes to the internet.
However most services generate asymmetric traffic and
the amount of the downstream from the servers in the
internet to the mesh nodes far exceeds the upstream.
Therefore some routing protocols are proposed which take
care of this backward traffic. Backward path routing in
more involved than routing on forward path because the
data addressed at any host in the internet only needs to be
forwarded to the gateway, while the backward routing
protocol needs to address each node individually.
There are three main families of backward routing
protocols; reactive hop-by-hop routing, proactive hop-by-
hop routing, and proactive source routing.
D. AODV-CGA:
This extended AODV protocol allows the use of multiple
gateways to the internet. It shares most of the mechanisms
with the well-known AODV protocol. The addition to the
existing AODV which is made is that, all gateways are
connected to a dedicated router that acts as a proxy to the
internet. This router has two tasks:
1. On the forward path, it sends route on behalf of
hosts in the internet;
2. On the backward path, it initiates route requests for
nodes in the wireless mesh network.
E. Proactive Field-based Routing (PFR):
Wireless mesh nodes periodically exchange beacons.
These beacons contain a list of all known destinations with
their respective field value. When a new destination
appears, it announces its presence with beacons to its
neighbors in order to establish a field. With this
mechanism, a field on the network is constructed for every
destination. This field assigns a value to every node in the
network; the destination bears the maximum value.
Packets are then routed along the steepest gradient
towards the destination.
Advantages and disadvantages:
This protocol ensures loop freedom. This protocol enables
nodes to consider multiple routes to the destination. This
protocol also has a drawback. Since it proactively
maintains all routes, it incurs communication overhead
even if the traffic is too low.
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
F. Gateway Source Routing (GSR):
In this protocol forward path information from the packets
that arrive at the gateway is reused. In the routing header
of every packet, the intermediate hops from the mesh node
to the gateway are recorded. These paths are then stored in
the gateways. To route packets to a mesh node, the mesh
gateway inverts the recorded forward path and copies it to
the packet header. The gateway then sends the packet to
the first node of the backward path. Each node updates the
path in the header by removing its entry and forward the
packet to the given next hop until the packets reaches the
destination.
GSR requires that a packet towards a host in the internet is
first sent by a mesh node in order to establish the
backward path. This should not be a problem when we
assume that majority of communication is initiated by the
mesh nodes.
If a mesh node has to act as a server, a dedicated
addressing mechanism will have to be used.
G. Hierarchical Routing Protocols:
Some of the protocols which differ from the above are
hierarchical routing and geography based routing.
In hierarchical routing, a certain self-organization scheme
is employed to group network nodes into clusters. Each
cluster has one or more cluster heads. Nodes in a cluster
can be one or more hops away from the cluster head.
Since connectivity between clusters is needed, some nodes
can communicate with more than one cluster and work as
a gateway. When the node density is high, hierarchical
routing protocols tend to achieve much better performance
because of less overhead, shorter average routing path,
and quicker set-up procedure of routing path. However,
the complexity of maintaining the hierarchy may
compromise the performance of the routing protocol.
In WMNs, a mesh client must avoid being a cluster head
because it can become a bottleneck due to its limited
capacity.
H. Geographical Base Routing:
Compared to topology-based routing schemes, geographic
routing schemes forward packets by only using the position
information of nodes in the vicinity and the destination
node. Thus, topology change has less impact on the
geographic routing than the other routing protocols. Early
geographic routing algorithms are a type of single-path
greedy routing schemes in which the packet forwarding
decision is made based on the location information of the
current forwarding node, its neighbors, and the destination
node. However, all greedy routing algorithms have a
common problem, i.e., delivery is not guaranteed even if a
path exists between source and destination. In order to
guarantee delivery, planar-graph-based geographic routing
algorithms have been proposed recently. However, these
algorithms usually have much higher communication
overhead than the single-path greedy routing algorithms.
76
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
VIII. CONCLUSIONS
Here we mentioned the basics of wireless mesh network,
their purpose, various techniques involved and the area of
the research in wireless mesh networks. Further research can
be done studying the relationship between channel
assignment techniques and routing protocol. These two
areas will influence one another and an efficient
combination can be possibly found.
IX. REFERENCES
[1]. Optimization of routing algorithm in wireless mesh networks.
Gupta B.K, Acharya B.M, Mishra M.K. NaBIC 2009
[2]. Distributed channel assignment for multi-radio wireless mesh
networks. Makram S>A, Gunes M. ISCC 2008
[3]. Performance Analysis of IEEE802.il Wireless Mesh Networks. Ye
Yan, Hua Cai, Seung-Woo Seo. ICC 2008
[4]. Routing metrics for multi-radio wireless mesh networks. Guerin J,
Portmann M, Pirzada A. ATNAC 2007
[5]. Routing Packets into Wireless Mesh Networks. Baumann R,
Heimlicher S, Lenders V, May M. WiMOB 2007
[6]. Neighbor selection technique for multi hop wireless mesh
networks. Coll B, Gozalvez J.LCN 2009
[7]. A New multi channel MAC protocol combined with on demand
routing for wireless mesh networks. Guojun Shui, Shuqun Shen.
ICCSSE 2008
[8]. Hybrid routing with periodic updates in wireless mesh networks.
Damle A, Rajan D, Faccin S.M. WCNC 2006
[9]. Multipath routing algorithm based on traffic prediction in wireless
mesh networks. Li Zhi yuan, Wang Ru chuan, Bi Jun lei. ICNC
2009
77 http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
An Approach For
Designing Distributed Real Time
Database
Dr. Dhuha Basheer Abdullah
Computer Sciences Dept./Computers Sciences and
Mathematics College /Mosul University
Mosul- Iraq
Ammar Thaher Yaseen
Computer Sciences Dept./Computers Sciences and
Mathematics College /Mosul University
Mosul- Iraq
Abstract- A distributed Real Time database system is a
transaction processing system that is designed to handle
workloads where transactions have service deadlines. The
emphasis here is on satisfying the timing constraint of transactions
(meet these deadlines, that is to process transactions before their
deadlines expire) and investigating the distributed databases. This
paper produces a proposed system named ADRTDBS.
In this work a prototype of client/server module and
server/server module for distributed real time database has been
designed. Server gets the data from direct user or a group of
clients connected with it, analyze the request; and broad updating
to all servers using 2PC (Two Phase Commit) and executing the
demand by using 2PL (Two Phase Locking). The proposed model
does not concern with data only, but provide a synchronize
replication, so the updating on any server is not saved unless
broadening the updating on all servers by using 2PC, and 2PL
protocols. The database on this proposed system is homogenous
and depend on full replication to satisfy real time requirements.
The transactions have been scheduled on the server by using a
proposed algorithm named EDTDF (Earliest Data or Transaction
Deadline First). This algorithm works to execute transactions that
have smallest deadline at the beginning, either this deadline
specific to the data or to the transaction itself. Implementing this
algorithm helps to execute greater rate of transactions before their
deadlines.
In this work two measures of performance for this system
(proposed model) were been conducted; first, computing the Miss
Ratio (rate of no. of executing transactions that miss their
deadline); second, computing the CPU utilization (CPU utilization
rate), by executing a set of transactions in many sessions.
Keywords: real time, databases, distributed, replication, Scheduling
I. INTRODUCTION
According to the definition provided by Coulouris,
Dollimore & Kindberg [4], a distributed system consists of a set
of autonomous processing elements that are connected via a
communication network and interact via message passing.
A database is a structured set of data maintained by a
database management system (DBMS) that interfaces with a set
of applications or clients that access and modify the data. In a
distributed database system, the data is distributed among
autonomous DBMS instances (nodes or sites) that communicate
via a network. The nodes, potentially along with a central
coordinator, are collectively referred to as a distributed
database management system (DDBMS) [1,7,8,15].
In a distributed database, replication of data objects(The
term object is used for the unit of replication; this could just as
well be a table in a relational database as an object) is often
used to improve fault tolerance and availability in the system
by maintaining several copies of data objects and placing those
copies close to the clients that want to use them [19].
In a real-time system (RTS), the value of a performed task
depends not only on its functional correctness, but also on the
time at which it is produced. For example, when an
autonomous vehicle detects an obstacle in its intended path, it is
crucial that it changes its path before a collision occurs. Real-
time systems are often embedded, meaning that they are a part
of (and interact heavily with) a physical environment.
Typically, embedded systems use specific-purpose rather than
general-purpose computers, such as in the embedded system
controlling fuel injection in a car engine [6,20].
It is paramount that real-time systems have predictable,
bounded and sufficiently low requirements on resources such as
memory, network bandwidth and processor execution time,
since failures due to unpredictable behavior and/or over
consumption of available resources may cause unacceptable
damage to humans or equipment. Real-time systems also need
to be highly and predictably available, meaning that when a
request is made to the system, it can be guaranteed that the
system is available to service that request within a predictable
and bounded time.
A distributed real-time system (DRTS) combines
characteristics of distributed and real-time systems. This means
that in such a system, issues related to distribution (such as
execution of distributed algorithms and network
communication) must be addressed with real-time requirements
in mind.
Real-time database systems (RTDBS) are often used to
manage data in real-time systems, since traditional databases
cannot meet the timeliness and predictability requirements of a
RTS. As many embedded applications with real-time
requirements are inherently distributed, RTDBS are often
distributed over a set of autonomous nodes, creating a need for
distributed real-time database systems (DRTDBS) [10,14,16].
78
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
• Replication in DRTDB S
Data replication can be used to increase availability,
predictability, and reliability of transaction processing in
DRTDB S. Common replication approaches for DRTDB S use
either a primary copy to deterministically apply updates to
replicated data, or use distributed concurrency control and
distributed commit protocols.
The distributed algorithms required to implement, e.g.,
distributed locking (to ensure serializability) and distributed
commit (to ensure mutual consistency and durability) are hard
to make predictable and sufficiently efficient due to their
reliance on correct message delivery. Furthermore, depending
on the replication approach a transaction may be forced to
either wait or roll back and restart due to concurrent execution
of transactions on remote nodes. Such behavior is problematic
in real-time systems, since potential blocking times and
rollbacks must be considered when determining worst-case
execution times of transactions. For this reason, optimistic
replication approaches, where transactions are allowed to
execute as if no concurrent transactions exist, are more suitable
than pessimistic replication approaches in real-time databases.
Optimistic replication increase the availability, predictability
and efficiency of transaction execution at the cost of transaction
conflicts that must be resolved [2,9].
II. RATED WORK
In 1994 the researcher Nandt Subakr and others discuss the
ways in which the Commit Protocol that could be adapted to
the environmental sensitivity of the cases required for real-time.
This protocol depends on the strategies and the installation
optimistic on local compensation [13].
In 1994 also provided a researcher Victor Fiy and other
researchers produce the basic rules to support the necessary
qualities to the environment of the account distributed to RT,
which is the modalities of distribution of time. It provides
general concepts to clarify the application of the expansion of
CORBA [18].
In 1998 the researcher Krayas Shihabi and others discuss
the experience to implement the 2- Server have the same DBMS
are linked through the Internet. Focusing on the intelligence
linking the researchers explained how the firm Optimal query
plan may choose the most expensive mistake. This takes
precedence over the lack of knowledge of the operational
environment [12].
In 2003 the researcher Yuan Wei and others discuss
produce a study on the extraction using real-time updating of
data and strategies on demand in DRTDB and the definition of
certain laws to choose the best policy of modernization. Based
on these laws, the researchers suggested an algorithm to derive
the updated data, the derivation policy of modernization
practical data sets automatically [17].
In 2005 the researchers Broheedi Marcos and steen Andler
illustrate how to bring forward the requirements in the
DARTDBS. It is possible to use a model requirements of the
modalities of information with RT[3].
In 2006 also provided a researcher Benoi Ravindran and
others Where they distributed scheduling algorithm Call CUA.
The parameters indicated it would satisfy for Thread time when
there is failure. Algorithm is the Best-Effort and the Thread of
the highest importance when they arrive at any time be the
possibility of implementing a very high [11].
In 2008 the researcher Alexander Zharkov discuss how to
use the material offers Materialized Views in DRTDBMSs. The
researcher offers an algorithm for building dynamic and
evaluation of the material cost. President difference this
algorithm from its predecessors is taken into consideration the
characteristics of time Temporal Properties of relations
president and data processing [21].
III. CONTRIBUTIONS
ADRTDBS is a real time distributed database management
system prototype that is designed to support distributed
transactions processing with timing constraints.
The ADRTDBS offers many contributions listed below:
• Database in main memory: disk access should be minimize
in a RTS, since reading from disk is both unpredictable and
orders of magnitude slower than memory access.
ADRTDBS is built to keep the entire database resident in
main memory.
• Full Replication: times for network messages are
unpredictable, and accessing data on remote nodes in much
slower than local access. So ADRTDBS employs a full
replication scheme which ensures that local replicas exist
for all objects accessed by a transaction removing the need
for remote object access.
• ADRTDBS Design and Implementation:
■ A structure to add support of executing real time
transactions in distributed environment.
■ Providing scheduling algorithm named EDTDF
Earliest Data or Transaction Deadline First for
Transaction execution.
■ Produce an approach for concurrency control by using
2PL (Two Phase Locking) protocol to managing
concurrent execution of transactions.
■ Execute data shipping and transaction shipping.
■ Provide synchronize replication and synchronization
updating by using 2PC (Two Phase Commit).
■ Provide backup and recovery approach to process
failure.
79
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
IV. THE PROPOSED SYSTEM
Given the important developments in computer and
software industry databases and the increasing use in different
areas of life (such as the management of banks, libraries,
companies, factories ... etc.) and because of its great importance
in a systematic compilation of data and processing, updating
and retrieval with pinpoint accuracy, speed and the urgent need
to provide such techniques in our country to keep pace with this
development software tremendous invaded the whole world,
this system was built to be a first step in the application of
modern techniques and contemporary distributed database
environment in real time. It was named Approach for designing
Distributed Real Time Database System (ADRTDBS). The
system ADRTDBS deals with Homogeneous distributed
databases (i.e. all computers linked to the network is made of
the same company (Pentium IIII) and contains the same version
of the operating system (Windows XP) as well as containing
the same version of the database management system ( DBMS
Oracle 9i), the same version of the program interfaces
(Developer 6i). Has the capacity to implement Soft Real Time
Transactions.
V. SYSTEM ARCHITECTURE
The system architecture consists of the following structure
shown in the figure (1):
-a
Server 1 I II Server 2 |
:3f
Figure (1) Architecture of ADRTDBS System
It contains two computers working as server having same
database, and two computers working as clients connected with
each server. Connection between computers is via HUB. The
database resident in each server connected with the network
and having same data and structure (replicas). The clients
contain interfaces that making connection with servers and
retrieve updating database. Any transaction will be executed on
the database will have one of two cases: either local transaction,
the implementation is the only current computer server. Or
Global transaction, the implementation to all servers linked in
the network.
VI. SYSTEM MODEL
• Database Model
A real time distributed system consists of two autonomous
computers system (sites) connected via a communication
network. Each site maintains a copy of database. In order for
transactions to be applied consistently to all replicas and give a
result within deadline time, a prototype units runs at each site.
Also this prototype architecture gives the distributed nature and
the increased communication burden of such a database system.
The smallest unit of data accessible to the user is called data
object. In this distributed database system with replicated data
objects a logical data object is represented by a set of one or
more replicated physical data object. The database is fully
replicated at all sites. The database consists of two types of data
objects: temporal and non-temporal. Temporal data object are
used to record the state of the object in the external
environment which its value changes frequently with time.
- Shipping Approaches
Two approaches for processing transactions in a
ADRTDBS system: query shipping and data shipping.
• Data Shipping
In the data shipping approach, a transaction initiated by a
client will be processed at the client. While the transaction is
processing, the client sends data requests, which are required by
the transaction, to the database server. The server responds to
the requests by sending the required data objects to the client.
The processing of the transaction will be completed at the
client.
• Query Shipping
In the query shipping approach, the client sends queries to
the database server for the transaction, instead of data requests.
Once the server receives a query, it processes the query and
sends the results back to the client. In the query shipping
approach, the communication cost and the buffer space required
at the client side are smaller than that in the data shipping
approach. Also, the query shipping approach provides a
relatively easy migration path from an existing single-site
system to the client-server environment since the database
engine can have a process structure similar to that of a single-
site database system. On the other hand, the data shipping
approach can off-load functionality from the server to the
clients. This may improve the scalability of the system and
balance the workload in the system. Figure (2) illustrates
flowcharts for query shipping from the point of view for server
and client.
80
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
| S«uJ. mue l&Ajay-tXf TTl*= Sg^qr |
C^Z^ 1
Figure (2) : a- Query Shipping in Server, b- Query Shipping in Client
• Transaction Model
A transaction is a sequence of operations that takes database
from a consistent state to another consistent state. Two types of
transactions are used in this proposed system: query
transactions and update transactions. Query transactions consist
only of read operations that access data object and return their
values to the user. Thus, query transactions do not modify the
database state. Update transactions consist of both read and
write operations.
A transaction Ti in this proposed system characterized by
the following attributes: Ti = ( r i9 we i5 rd } , p } )
r z : released time for the transaction, which represent the
arrival time,
we,: the estimated worst case execution time.
rd { : the relative deadline, it indicates the requirement to
complete the transaction before the instant deadline.
Pi: the priority, of transaction, which depend on the
transaction relative deadline.
VII. ADRTDBS SYSTEM UNITS
An ADRTDBS system capable of executing transactions
with timely constraint in distributed environment. The system
consists of ten of working units, and each server contains copy
of program for these units. Figure (3) illustrate the prototype of
the ADRTDBS system.
• Transaction Admission Control (TAC) Unit:
This unit receives transaction request from database
servers and clients. This unit provides a database interface to
the application. This interface consists of a data manipulated
language, in which the user (application) can query and
manipulate data elements.
• Index Management (IM) Unit-:
It is used to maintain an index for all tuples in the
database. It is capable of transforming a database key into the
memory address of the tuple correspondent to the database
key.
• Memory Management (MM) Unit
This unit is responsible for memory allocation of tuples
and database indexes.
• Transaction Management (TM) Unit -:
This unit responsible of managing transactions coming
from admission control unit TAC and transmit it to scheduler
unit TS to schedule them according to the proposed algorithm.
This unit provides required data to each transaction and
controls and assembles results for each request. This unit also
controls and manages other units and calculate deadline for
each transaction.
A deadline function computes the execution time for each
transaction and predicts the deadline according to the
following equation:
TD = RL (T) + Pr_Ex(T) * SF
81
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
Figure (3) Units of Proposed ADRTDBS System
Where
RL (T) : release time for transaction T
Pr_Ex(T) : Predict execution time for T and this time can
be computed by
Pr_Ex(T) = (T_operation + T_update) * N_op + T_cc
Where
T_operation : time to process an operation
T_update : time to update data
N_op : number of operation
T_cc : communication cost
SF : slack factor 15 >= SF >= 1
• Transaction Scheduler (TS) Unit
This is responsible for scheduling transactions. This unit
maintains the list of transactions in a ready queue and releases
the next transaction when the previous is completed and give it
to the CPU. The ready queue is organized according to the
transaction priority. Each transaction is characterized by a
deadline which defines its urgency with respect to the other
transactions of real time application. The higher priority is
given to transaction with minimum deadline according to the
scheduling algorithm EDTDF (Earliest Data or Transaction
Deadline First). If the system cannot complete transaction
before its deadline, the transaction is aborted.
The algorithm is work like this:
Receive transaction from TM unit.
Determine if the transaction contains temporal data.
Define the deadline for the temporal data (DD).
Compute the deadline for transaction (TD).
Compute the Final Transaction Deadline (FTD) by
if the transaction contain temporal data then
FTD = Min(TD,DD)
else FTD = TD
Put the transaction in the ready queue with
transaction with earliest deadline at the head of the
queue.
1.
2.
3.
4.
5.
6.
» Transaction Deadline Test (TDT) Unit
82
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
This unit is responsible of decision that a transaction is
aborted whenever it is found to have missed its deadline. So,
for each transaction periodically checks whether or not the
transaction will be able to meet its deadline taking into
consideration the fact that the transaction has to update the data
object in its write-set at each database. If the system's current
time plus the time to update all data objects in a transaction's
write-set is greater than the transaction's deadline, it means that
this transaction will not be able to commit before its deadline is
reached. In order not to be waste any system resources, the
transaction will be aborted and removed from the system.
Figure (4) illustrate the work of this unit.
Fjw&rYftTmis'iCTMi from Queua-wili
IJiiIiiijim De-vlhii*
L+iYipm
ite- Current lutein Tin*
H-uinuin Transaction Tim* = Current tine- +
Ei«ut mi Tine- + Up'hte- Tun*
Hv. inu in Transaction Time*
i
n
r
^*»»- Transaction Dwdlii*
.rtbort Tmisiccion
1
Jn
lis Dwdliie- Couute-i .1H:D.= 1II:D+1
ExecutaTnns-iction
i
r
i
r
Fi»l
Figure (4) Transaction Deadline Test unit
• Concurrency Control (CC) Unit
This unit is responsible of synchronous execution for
more than one transaction that require execution on same
database at same time. In this work a 2PL (Two Phase
Locking) protocol was used to control concurrency. 2PL
protocol follows the following three rules:
- When the protocol receive a lock request, it tests whether the
requested lock conflicts with another lock that is already set.
If so, it queues the lock request. If not, it responds to the lock
request by setting the lock.
- The 2PL protocol locks a data item only once, it cannot
release the lock until the DM (Data Manager) has completed
processing of the lock's corresponding operation.
- The 2PL protocol releases a lock for a transaction, it may not
subsequently allow any lock for the same transaction.
• Network Management (NM) Unit
This unit is responsible of managing the transfer of data
between servers Depending on the TCP/IP Protocol which
consider the best protocol that provide high speed for sending
and receiving data over the network.
Database server communicates among each other via 1-to-n
communication which is consider as group communication.
Reliable broadcast parameter of this communication ensures
that a message sent by a correct database server, or delivered
by a correct database server, is eventually delivered to all
correct database servers.
• Replication Control (RC) Unit
This unit controls all updating on the local database. This
unit broad changes on database to remain copies of database on
servers connected by network in synchronizing manner by
using 2PC (Two Phase Committing).
Replication requires to have a specific site - the main copy
- associated with each data item. The clients must send their
requests to one particular server. This server is the main copy.
Because there is only one server executing the transactions,
there are no conflicts across the servers. Any update to the data
item must be first sent to the main copy where it is processed.
The main copy then propagates the update (or its results) to all
other sites. This approach is used in ADRTDBS system to
minimize conflicts among transactions executed over replicated
data.
The steps for this technique are the following:
1. The transaction starts at the primary copy site.
2. Read operations are executed locally.
3. The result of write operations are broadcast to the other
sites, (backups), (i.e. update every where).
4. The main copy site starts the Two Phase Commitment
Protocol (2PC).
5. The transaction is committed on all sites.
•Recovery and Backup (RB) Unit
This unit makes backups of database and recovering it when
required. This unit maintains information on database when an
error occur on computer or database or when we need copy the
database on more than one computer. The famous manner of
back up and recovery is export and import.
The steps of export and import are illustrated in figure (5).
83
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
Ifrrior CtfriWiiaTul
Xton^t
I Erd^E" C CiXlTltlXul. bwij't
Eriio: I£i^(xriiuaxLaiiu^4f jvciri
SpcoCy £0.c:Lcic:xi.ciTL:iTLil£i±9i£kiTL= DMT
Choose Eifort Entire Datakue Or Ufa:* Or TiHe* |
IfrrieriiYiEoitiiFeErL»ftje^a£Fvwi4.
Sp «i£y GLe IxicaAfxii ami EdHuaim = JXMT
Erii^r Bi-rflar Sue
i
Ouoos e Content* rf IH.e Orilv
■ T ,
Choose Create Enroi: >lu.e to Olject Existence
1
I Specify Imfc-xt Grant* Or Not I
SfHiCy Imfc-xt Tal-le Ddi Or Hot
HI
I Sjiedfy frnjioxt EritK'e Ejifoit iuc
Figure (5) : a- Export Data algorithm, b- Import Data algorithm
VIII. SYSTEM TEST
The system has been tested on Alrafidain bank. Many
transactions were take into consideration, (add customer , close
customer account, draw money, transfer funds, currency
change, etc).
. Transfer Fund Transaction: If we transfer fund between
customers the system demand many steps illustrated in the
following figure (6). The figure also shows that every step (or a
number of steps) of the unit within a system ADRTDBS. The
time parameters for this transaction are illustrated in table (1).
Table (1) Deadline Computation of Transfer Fund Transaction
No of Tables That Effected
6
No of Record That Effected
6
No of Fields That Effected
64
Type of Operation
Updating & Adding
T_operation
0.2778 mill sec.
T_update
0.1388 mill sec.
T_cc
2 mill sec.
N_op
6 mill sec.
Pr_Ex(T) = (T_operation + T_update) *
N_op+ T_cc
4.5 mill sec.
RL(T)
mill sec.
SF
2.8
TD = RL(T) + Pr_Ex(T) * SF
12.6 ~ 12 mill sec.
IX. PERFORMANCE EVALUATIONS
In conventional distributed database systems, performance
is primarily measured by the number of transactions completed
within a unit time. In distributed real time database systems,
timing and criticality characteristics of transactions must be
taken into account. So performance depends on many other
criteria, which are related to real time. Some of these criteria
are the number of transactions that missed their deadline,
average tardy time, etc. In this work, the performance metric
employed is the percentage of transactions that missed their
deadlines (%miss) in the total number of transactions that were
submitted to ARTDDBS system during the session period :
Miss ratio = No. of missed deadline transactions/
Total transactions * 100
Also we measure the total CPU utilization. And develop
performance measurement take into consideration database
reside in main memory consisting of 72 organism data. (Table
2) shows the model parameters and their baseline values.
We take a sample of 70 transactions of this application
distributed as (add new customer, updating customer
information, close account, deposit money, query about
account, transfer fund, display customer information). The time
of execute this transactions are 293 millisecond, and the
transactions missed their deadline 14.285%.
Miss Ratio = 10/70*100 = 14.285
84
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
Figure (6) : Transfer Fund Execution in ADRTDBS
Figure (6) : Cont. Transfer Fund Execution in ADRTDBS
85
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
Table (2) Performance Evaluation
Parameter
Base line value
System
Number of servers
2
Number of clients for each server
2
Communication cost
2 mill sec.
Database
Number of local databases in each site
1 Database
Number of objects in local database in
each site
11 tables, and 5 views
Database size
16 data object per local
database
Concurrency control
2PL (two phase locking)
Fraction of temporal data object
0.1
Transaction
Transaction size
3 to 5 operation uniform
distributed
Proportion of write operations
0.35
Slack Factor Range
1...15 the slack factor is
uniformly distributed in the
slack range (we use 2.8)
CPU
CPU scheduling
EDTDF (earliest data or
transaction deadline first)
CPU time to process an operation
1... 6 mill sec.
CPU time to commit changes on
database
1 mill sec.
CPU time to rollback changes on
database
1 mill sec.
CPU time to update a data object
6 mill sec.
The figure (7) illustrates measuring of Miss Ratio of this
system and figure (8) illustrates the CPU utilization.
s
~7 14 ^T~
Figure: (7) Miss Ratio of this system.
no. of transaction
Figure (8) CPU utilization
CONCLUSIONS AND FURTHER WORKS
In this paper a model of proposed distributed real-time
database system was designed, as this system has the ability to
execute real time transactions in distributed environment. The
proposed system uses full replication method, replication of the
same database on all sites.
The replication of the whole database at every site in this
proposed system improves availability remarkably because the
system can continue to operate as long as at least one site is
functioning well. It also improves performance of retrieval for
global queries, because the result of such a query can be
obtained locally from any site, hence a retrieval query can be
processed at the local site where it is submitted, if that site
includes a server module. Replication of data from the remote
site to the local site makes the data available on the local site
and minimizes the response execution time which very suitable
to distributed real time databases environment. Also, by
maintaining multiple copies, it becomes possible to provide
better protection against corrupted data.
using 2PL (Two Phase Locking) protocol help to control
synchronize execution of transactions in this proposed system
in two cases:
- When there is a request to execute same transactions from
some clients on one server at same time.
- When there is a request to execute same transactions from
some servers on one server at same time.
As a further work we suggest to use new encryption
algorithms to increase the security of system. Use another
distributed manner of database like hybrid and then see how it
is compatible with distributed real time database.
REFERENCES
[1] Al-Kinany, E. A., "Heterogeneous System Design for
Distributed Database Systems", Msc, Alrasheed College,
University of Technology, 2005.
[2] Bouzefrane, S. S.; Kaiser, C, "Distributed Overload Control
for Real Time Replicated Database Systems", Cedric
86
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
Laboratory, Conservatoire National des arts et Metiers,
FRANCE, 2002.
[3] Brohede, M; Andler, S. F., "Using Distributed Active Real
Time Database Functionality in Information Fusion
Infrastructures", University of Skode, SE-54128, Sweden.
http://cgi.omg.org/docs/ptc/01-08-34.pdf, 2005.
[4] Coulouris, G.; Dollimore, J.; Kindber g, T., "Distributed
Systems Concepts and Design", Fourth Edition, ADDISON-
WESELY PUBLISHING, 2005.
[5] Deborah, S.; Henry, R., "Oracle9i Net Services
Administrator's Guide", Release 2 (9.2), Oracle
Corporation, 2002.
[6] Deng, G.; Schmidt, D. C, Nechypurenko A.; Gokhale, A.,
"Developing Distributed Real Time and Embedded Systems
with Modeling Tools and component Middleware: A Case
Study", Department of EECS, Vanderbilt University,
Siemens Corporate Technology, Germany, 2005.
[7] Graham, M. H., "Issues in Real Time Data management",
Technical Report, CMU/SEI-91-TR-017, ESD-91-TR-017,
1991.
[8] Lindstrom, J.; Niklander, T.; Raatikainen, K., "A
Distributed Real Time Main Memory Database
Telecommunication",Department Computer Science,
University of Helsinki.
www.imtrg.me.metu.tr/publications/paper.pdf, 1998.
[9] Peddi, P.; Dipippo, L. C, "A Replication Strategy for
Distributed Real Time Object Oriented Database",
Proceedings of the Fifth IEEE International Symposium
on Object-Oriented Real-Time Distributed Computing,
Washington, D.C., 2002.
[10] Ramamritham, K., "Real Time databases", International
Journal of Distributed and Parallel Databases 1(2), 199-
226, 1993.
[11] Ravindran, B.; Anderson, J. S.; Jensen, E. D., "On
Distributed Real Time Scheduling in Network Embedded
Systems in the Presence of Crash Failures", ECE Dept.
Virginia Tech, VA 24061, USA, The MITRE
Corporation, Bedford, mA01730, USA, 2006.
[12] Shanker, U., " Some performance issues in distributed real
time database systems", M.sc, Department of Computer
Science, University of Virginia, TesiOnline. USA, 2000.
[13] Soparker, N.; Lery E.; Korth, H. F.; Silberschatz, A.,
"Adaptive Commitment for Distributed Real Time
Transactions", Work Partially Supported By NSF Grant
IRI-8805215, and by a Grant from IBM Corporation,
USA, 1994.
[14] Syberfeldt, S., "Optimistic Replication with forward
Conflict Resolution in Distributed Real Time Database",
Ph.D., Department of Computer and Information Science,
Linkopings University, Sweden. Printed by LiU-tryck,
Linkoping, 2007.
[15] Tao, J.; Williams, J. G., "Concurrency Control and Data
Replication Strategies for Large Scale and Wide
Distributed Database",University of Pittsburgh, IEEE, 0-
7695-0996-7/01,10.00, 2001.
[16] Wei, Y.; Prasad, V.; Son, S. H., "Qos Management of Real
Time Data Stream Queries in Distributed Environments",
Department of Computer Science, University of Virginia.
www.cs.virginia.edu/~radb/sch.html, 2006.
[17] Wei, Y.; Son, S. H.; Stankovic, J. A ., " Maintaining Data
Freshness in Distributed Real Time Databases", USA,
This Work was Supported in Part by NSF Grant IIS-
0208758,CCR-032,60, and CCR 0098269, 2003.
[18] Wolfe, V. F.; Black, J.; Thuraisingham, B.; Krupp, P.,
"Real Time Method Invocation in Distributed
Environment", USA, The MITRE Corporation Before
MA,USA, This Work is Partially supported by the US
National Science Foundation, Us NAVAL Undersea
Warfare Center, 1994.
[19] Woochul K., Sang H., and John A., "DRACON: QoS
Management for Large-Scale Distributed Real-Time
Databases", JOURNAL OF SOFTWARE, VOL. 4, NO. 7,
2009.
[20] Zharkov, A., "Performance Evaluation of Transaction
Handling Policies on Real Time Database Management
System Prototype", Proceeding of the Spring Yong
Researcher's Colloquium On Database and Information
Systems SYRCODIS, MOSCOW, Russia, 2007.
[21] Zharkov, A., "On Using Materialized Views for Query
Execution in Distributed Real Time Database
Management System", Proceedings of the Spring Young
Researcher's Colloquium On Database and Information
Systems SYRCoDIS, St. -Petersburg, Russia, 2008.
Dr. Dhuha Basheer Abdullah Albazaz /Asst.
Prof / computers Sciences Dept. / College of
Computers and Mathematics / University of
Mosul. She has a Ph.D. degree in Computer
Sciences since 2004. Specific Specialist in
Computer Architecture and Operating System.
Supervised many Master degree students in
operating system, computer architecture, dataflow machines,
mobile computing, real time, distributed databases. She has
three Phd. Students in FPGA field, distributed real time
systems, and Linux clustering. She also leads and teaches
modules at both BSc, MSc, and Phd. levels in computer
science. Also she teaches many subjects for Ph.D. and master
students.
87
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol 8, No. 6, September 2010
Person Identification System using Static-dyamic
Signatures Fusion
Dr. S.A Daramola
Department of Electrical and Information Engineering
Covenant University
Ota, Nigeria
Prof. T.S Ibiyemi
Department of Electrical Engineering
University of Ilorin
Ilorin, Nigeria
Abstract — Off-line signature verification systems rely on static
image of signature for person identification. Imposter can easily
imitate the static image of signature of the genuine user due to
lack of dynamic features. This paper proposes person identity
verification system using fused static-dynamic signature features.
Computational efficient technique is developed to extract and
fuse static and dynamic features extracted from offline and online
signatures of the same person. The training stage used the fused
features to generate couple reference data and classification stage
compared the couple test signatures with the reference data
based on the set threshold values. The system performance is
encouraging against imposter attacker in comparison with
previous single sensor offline signature identification systems.
Keywords- fused static -dynamic signature; feature extraction;
forgeries
I. Introduction
Person identity verification is a problem of authenticating
individual using physiological or behavioral characteristics
like face, iris, fingerprint, signature, speech and gait. Person
identification problem can be solved manually or
automatically. Biometric systems automatically use biometric
trait generated from one or two sensors to validate the
authenticity of a person. Automatic person identity
verification based on handwritten signature can be classified
into two categories: on-line and off-line, differentiated by the
way signature data is acquired from the input sensor. In off-
line technique, signature is obtained on a piece of paper and
later scanned to a computer system while in on-line technique,
signature is obtained on a digitizer thus making dynamic
information like speed, pressure available while in offline only
the shape of the signature image is available [1] [2].
In this paper, combination of offline and online signatures
are used for person identification. The process involves
verification of a signature signed on both paper and electronic
digitizer concurrently. Therefore the physical present of the
signer is required during the period of registration and
verification. This type of system is useful particular in the
bank while the physical present of the holders of saving
current are required before money can be withdrawn. In
Nigeria many banks manually identify holder of saving current
using face and static signature, in the process genuine users
are rejected as imposters because of high intra-class variation
in signatures. Frauds as result of signature forgeries must be
prevented particular among closely resemble people. Fusion
of dynamic and static signature will strengthen the
identification of people physically in paper documentation
environment.
A detailed review of on-line signature verification
including summary of off-line work until the mid 1990's was
reported in [1] [2]. Alessandro et al. [3] proposed a hybrid
on/off line handwritten signature verification system. The
system is divided into two modules. The acquisition and
training module use online-offline signatures, while the
verification module deals with offline signatures. Soltane et al
[4] presented a soft decision level fusion approach for a
combined behavioral speech- signature biometrics verification
system. And Rubesh et al [5] presented online multi-parameter
3D signature and cryptographic algorithm for person
identification. Ross et al [6] presented hand-face multimodal
fusion for biometric person authentication while Kiskus et al
[7] fused biometrics data using classifiers.
From the approaches mentioned above, some of the
authors used online signature data to strengthen the system
performance either at registration or training stage, while
others combined online signature data with other biometric
modalities data like speech, face as means of person
identification. The system proposes in this novel frame work
is based on fusion of static and dynamic signature data at
feature level for person identification. The universal
acceptance of signature, compatibility of offline and online
signature features make the proposed system more robust,
accurate and friendly in comparison with other previous multi
biometric modalities systems or single sensor offline system
for person identification.
Section 2 provides the description of the system, the
signature preprocessing and feature extraction and fusion
technique. Also in section 2, the signature training, threshold
selection and classification are presented. Section 3 shows the
experimental results and finally, conclusions are drawn in
section 4.
88
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol 8, No. 6, September 2010
II. PROPOSED SYSTEM
The system block diagram is shown in Fig. 1 . The offline
and online data are collected at the same time from the same
user during registration/training and verification exercises.
Also the offline signature are preprocessed to remove
unwanted noise introduced during scanning process whereas,
the online signatures are not preprocess in order to preserve
the timing characteristics of the signature. Discriminative
static and dynamic features are extracted separately from
offline and online signature respectively. At the feature level
the two signatures are fused together to obtain a robust static -
dynamic features. These features are used to generate couple
reference data during training and for signatures classification.
A. Data Acquisition
The signature database consists of a total number of 300
offline and 300 online handwritten genuine signature images
and 100 forged signatures. The genuine signatures are
collected from 50 people. Each of the users contributed 6
offline and 6 online signature samples. The 100 skilled
forgeries consist offline and online signatures, they are
collected from 25 forgers and each of the forgers contributed 4
samples. The raw signature data available from our digitizer
consists of three dimensional series data as represented by (1).
S(t) = [x(t),y(t),p(t)] T t = 0,1,2,.
(1)
where (x(t), y(t)) is the pen position at time t , mdp(t)
{0,1,... ,1024} represents the pen pressure.
Input offline
signature on paper
Input online
signature on digitizer
Preprocessing
Offline signature
feature extraction
i
On-line signature
feature extraction
Fusion of online and offline features
Couple test feature
Training
Couple reference feature
Classification stage
T
Result
Figure 1: System block diagram
B. Offline Signature Preprocessing
The scanned offline signature images may contain noise
caused by document scanning and it has to be removed to
avoid errors in further processing steps. The gray-level image
is convolved with a Gaussian smoothing filter to obtain a
smoothed image. The smoothed gray image is converted into
binary image and then thinned to one pixel wide.
C. Offline Feature Extraction
The feature extraction algorithm for the static signature is
stated as follows:
(1) Locate signature image bounding box.
(i) Scan the binary image from top to bottom to
obtain the signature image height.
(ii) Scan the binary image from left to right to obtain
the signature image width.
(2) Centralization of the signature image.
(i) Calculate centre of gravity of the signature image
using (2).
1 ^
^ N
(2)
(ii)Then move the signature image centre to coincide
with centre of the predefined image space.
(3) The image is partitioned into four sub-image parts.
(i) Through point X make a horizontal splitting
across the signature image.
(ii) Through point y make a vertical splitting across
the signature image.
(4) Partition each of the sub-image parts into four rectangular
parts.
89
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(i) Locate the centre of each of the sub-image parts
using (2).
(ii) Repeat step 3(i) and 3(ii) for each of the sub-
image parts in order to obtain a set of 16 sub-image
parts.
(5) Partition each of the 16 sub-image parts into four signature
cells
(i) Locate the centre of each of the sub-image parts
using (2).
(ii) Repeat step 3(i) and 3(ii) for each of the sub-
image parts in order to obtain a set of 64 sub-image
cells.
(6) Calculate the angle of inclination of each sub-image centre
in each cell to the lower right corner of the cell.
(i) Locate the centre of each of the 64 sub-image
cells using (2)
(ii) Calculate the angle that each centre point makes
with the lower right corner of the cell.
The feature extracted at this stage constitutes the offline
feature vector, which is represented as: F = f b f 2 , f 3 ,
f 4 f 64 . The details and diagrams of the feature extraction
are given in [8].
D. Online Signature Extraction
Three on-line signature features are extracted at each
sampling point from the raw data. The features are Ap/Ax,
Ap/Ay and v. Ax corresponds to change of x between two
successive sampling points, Ay corresponds to change of y
between two successive sampling points, Ap corresponds to
change of p between two successive sampling points, Ap/Ay
corresponds to ratio of Ap to Ay, Ap/Ax corresponds to ratio
of Ap to Ax and v corresponds to change of speed between
two successive sampling points [9]. These features are
obtained using (3), (4) and (5).
(IJCSIS) International Journal of Computer Science and Information Security,
Vol 8, No. 6, September 2010
done at decision level. While in [6] fusion of hand and face
was done at feature level. In this work offline feature is
combined with online feature at feature level. The
compatibility of the signature data from the same person, from
different sensors made the fusion possible without any lost of
information. The steps involve are stated as follows: given that
extracted static feature is F = f b f 2 , f 3 f 4 f 64 . The mean
and variance of the feature vector are calculated using (6) and
(7) respectively
v = JiAxf+(Ay) 2
Ap = pjt)-pjt-\)
Ay y(t)-y(t-l)
Ap = pit)- pit -\)
Ax x(t) - xit - 1)
(3)
(4)
(5)
I N
I N
(6)
(7)
The fused features are obtained by normalized each of the
, Ap Ap x
extracted online features (v, , ) using the variance
Ay Ax
of the offline feature (& ofr ) • The three fused features (SF1,
SF2 and SF3) become:
Ap Ap
<j rr Aver r, Axcj
off j off (
off
F. Training and Threshold Setting
Each of the registered users submitted 12 genuine
signatures to the system, out of which 8 signatures are fused
together to generate 4 couple reference features. These
features are used to generate 6 distance values by cross-
aligned the couple reference features to the same length using
Dynamic Time Warping (DTW). These distance values are
used to measure the variation within each of the user's
signatures, so as to set user-specific threshold for accepting or
rejecting a couple test signatures. Given four couple reference
signature samples Rl, R2, R3 and R4, these features are cross
aligned to obtain 6 distance values as shown in Fig.2. The
mean (m k ) and standard deviation (o k ) of the distances: d i2 ,
di3, di4, d 2 3, d 24 and d 34 are calculated and used to set the
threshold (t k ) for each of the users based on each of the fused
features as given in (8).
0>t k <m k +a k
(8)
E. Fusion of Offline and Online features
This technique is designed to compute fused feature
vector, which contains information from both the offline and
online signatures and used this feature vector for subsequence
processing. Information from two input sensors can be
combined at data acquisition stage, feature extraction stage or
at decision level stage. The accuracy of the system also
depends on the level of fusion and the discriminative ability of
fused data. In [4] the fusion of voice and signature data was
90
Figure2. Cross-alignment of Couple reference
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol 8, No. 6, September 2010
G. Classification of Couple Signature Features
Whenever a couple test signatures (offline and online)
come into the system, the fused feature vector of the couple
test signatures is pair-wise aligned with each of the four
couple reference features using DTW. Four distance values are
obtained as shown in Fig.3. The distance (d t ) of the couple test
feature (FT) from the four couple reference features Rl, R2,
R3 and R4 is calculated using (9).
TABLE 1: OFFLINE FEATURES IN COMPARISON WITH THE
PROPOSED FUSED FEATURES.
i f _ JT\ + JT2 + JT3 + JT4
(9)
If d t is within the assigned threshold value then the fused test
signature is assigned a pass mark otherwise it has no pass
mark. Finally a decision by the system in accepting or
rejecting a couple test signatures is based on total pass mark it
obtained based on the three fused features.
III. Experimental Results
Experiments have been conducted to evaluate the
discriminative ability of each of the fused features against
forgers attack. Also the proposed system is tested based on the
three new fused features. Total number of 150 fused signatures
made up of 100 genuine signature features and 50 skilled
forgery features are collected from 75 people are tested. The
performance evaluation is based on False Acceptance Rate
(FAR) and False Rejection Rate (FRR). Table 1 shows the
results of the performance of these fused features in
comparison with previous single offline features. Table 2
shows the proposed system FAR for skilled forgeries and the
FRR for genuine signatures.
Fused test features (FT)
Type
Feature
FRR
FAR
Some
previous
related
offline
features in
[8][10]
Pixels normalized angle relative
to the cell lower right corner
1.250
2.500
Image centre angle relative to
the cell lower right corner
2.500
2.500
Vertical centre points
7.500
8.750
Horizontal centre points
6.250
7.500
Proposed
fused
offline-
online
features
SF1
0.150
0.120
SF2
0.120
0.052
SF3
0.100
0.080
TABLE2: FAR AND FRR RESULTS OF THE PROPOSED SYSTEM
Type
Total
Accepted
Rejected
FAR
FRR
Individual
genuine
fused
features
100
95
5
-
0.05
Individual
skilled fused
forgeries
50
1
49
0.02
-
Figure 3. Distance between couple reference features and couple
test feature
IV. CONCLUSION
This paper has proposed a person identification system
using fused signature features from two biometric sensors.
Fused signature feature is used to strengthen verification
system in paper documentation environment like banks where
the present of the account holders are required for transaction.
Signature is universally accepted, this make the proposed
system more friendly and acceptable in comparison with
91
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
others biometric traits combination. The experimental results
have shown that fused signature identification method is more
accurate in comparison with previous single sensor offline
signature identification techniques.
References
[1] R. Plamondon and S.N. Srihari "On-line and off-line handwriting
recognition: a comprehensive Survey", IEE trans, on Pattern Analysis and
Machine Intelligence, vol. 22, No.l, pp. 63-84, 2000.
[2] F. Leclerc and R. Plamondon, "Automatic verification and writer
identification: the state of the art 1989-1993", International Journal of Pattern
Recognition and Artificial Intelligence, vol. 8, pp. 643 - 660, 1994.
[3] A. Zimmer and L.L. Ling "A hybrid on/off line handwritten signature
verification system", Proc. of the seventh International Conference on
Document Analysis and Recognition, 2003.
[4]. S. Mohamed, G. Noureddine and D. Noureddine "Soft decision level
fusion approach to a combined behavioral speech-signature biometics
verification", International journal of Biometrics &Bioinformatics, vol. 4,
issue 1.2009.
[5] P.M. Rubesh, G. Bajpai and V. Bhaskar "Online multi-parameter 3D
signature verification through Curve fitting", International Journal of
Computer Science and Network Security, vol.9, No. 5, pp 38-44. 2009.
[6] A. Ross and R. Govindarajan "Feature level using hand and face
biometrics", Proc. of SPIE Conference on Biometric Technology for Human
Identification, vol. 5779, pp. 196-204, 2005.
(IJCSIS) International Journal of Computer Science and Information Security,
Vol 8, No. 6, September 2010
[7] D. R. Kisku, P. Gupta and J. K. Sing "Offline signature identification by
fusion of multiple classifiers using statistical learning theory", International
Journal of Security and Its Applications, vol.4, No. 3, 2010.
[8] S. Daramola and S. Ibiyemi "Novel feature extraction technique for offline
signature verification", International Journal of Engineering Science and
Technology, vol (2)7, pp 3137-3143, 2010.
[9] S. A. Daramola and T.S Ibiyemi "Efficient On-line Signature verification"
International Journal of Engineering &Technology, vol 10, No. 4. pp 48-52.
2010.
[10] M. Banshider, R .Y Santhosh and B .D Prasanna "Novel features for off-
line signature verification" International Journal of Computers,
Communications & Control, vol. 1 , No. 1, pp. 17-24. 2006.
AUTHORS PROFILE
Dr. S.Adebayo Daramola obtained Bachelor of Engineering from University
of Ado-Ekiti, Nigeria, Master of Engineering from University of Portharcourt,
Nigeria and PhD from Covenant University, Ota, Nigeria. His research
interests include Image processing and Cryptography.
Prof. T.S Ibiyemi is a Professor in Computer Engineering. He has more than
30 years teaching and research experience; he has many papers in local and
international journals. His research interests include Image processing,
Multimedia and Processors architecture.
92
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol 8, No. 6, 2010
Short term flood forecasting using RBF static neural
network modeling a comparative study
Rahul P. Deshmukh
Indian Institute of Technology, Bombay
Powai, Mumbai
India.
Abstract — The artificial neural networks (ANNs) have been
applied to various hydrologic problems recently. This research
demonstrates static neural approach by applying Radial basis
function neural network to rainfall-runoff modeling for the
upper area of Wardha River in India. The model is developed
by processing online data over time using static modeling.
Methodologies and techniques by applying different learning
rule and activation function are presented in this paper and a
comparison for the short term runoff prediction results
between them is also conducted. The prediction results of the
Radial basis function neural network with Levenberg
Marquardt learning rule and Tanh activation function indicate
a satisfactory performance in the three hours ahead of time
prediction. The conclusions also indicate that Radial basis
function neural network with Levenberg Marquardt learning
rule and Tanh activation function is more versatile than other
combinations for RBF neural network and can be considered
as an alternate and practical tool for predicting short term
flood flow.
Keywords-component; Artificial neural network; Forecasting;
Rainfall; Runoff;
I.
Introduction
The main focus of this research is development of
Artificial Neural Network (ANN) models for short term flood
forecasting, determining the characteristics of different neural
network models. Comparisons are made between the
performances of different parameters for Radial basis function
artificial neural network models.
The field engineers face the danger of very heavy flow
of water through the gates to control the reservoir level by
proper operation of gates to achieve the amount of water
flowing over the spillway. This can be limited to maximum
allowable flood and control flood downstream restricting river
channel capacity so as to have safe florid levels in the river
within the city limits on the downstream.
By keeping the water level in the dam at the optimum
level in the monsoon the post monsoon replenishment can be
conveniently stored between the full reservoir level and the
permissible maximum water level. Flood estimation is very
A. A. Ghatol
Former Vice-Chancellor
Dr. Babasaheb Ambedkar Technological University,
Lonere, Raigad, India.
essential and plays a vital role in planning for flood regulation
and protection measures.
The total runoff from catchment area depends upon
various unknown parameters like Rainfall intensity, Duration
of rainfall, Frequency of intense rainfall, Evaporation,
Interception, Infiltration, Surface storage, Surface detention,
Channel detention, Geological characteristics of drainage
basin, Meteorological characteristics of basin, Geographical
features of basin etc. Thus it is very difficult to predict runoff
at the dam due to the nonlinear and unknown parameters.
In this context, the power of ANNs arises from the
capability for constructing complicated indicators (non-linear
models). Among several artificial intelligence methods
artificial neural networks (ANN) holds a vital role and even
ASCE Task Committee Reports have accepted ANNs as an
efficient forecasting and modeling tool of complex hydrologic
systems[22].
Neural networks are widely regarded as a potentially
effective approach for handling large amounts of dynamic,
non-linear and noisy data, especially in situations where the
underlying physical relationships are not fully understood.
Neural networks are also particularly well suited to modeling
systems on a real-time basis, and this could greatly benefit
operational flood forecasting systems which aim to predict the
flood hydrograph for purposes of flood warning and
control[16].
A subset of historical rainfall data from the Wardha
River catchment in India was used to build neural network
models for real time prediction. Telematic automatic rain
gauging stations are deployed at eight identified strategic
locations which transmit the real time rainfall data on hourly
basis. At the dam site the ANN model is developed to predict
the runoff three hours ahead of time.
In this paper, we demonstrate the use of Radial basis
function neural network (RBF) model for real time prediction
of runoff at the dam and compare the effectiveness of different
learning rules and activation function. Radial basis function
neural network is having a feed-forward structure consisting of
hidden layer for a given number of locally tuned units which
are fully interconnected to an output layer of linear units.
At a time when global climatic change would seem to
be increasing the risk of historically unprecedented changes in
93
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, 2010
river regimes, it would appear to be appropriate that
alternative representations for flood forecasting should be
considered.
II. Methodology
In this study different parameters like learning rule and
activation function are employed for rainfall-runoff modeling
using Radial basis function neural network model of artificial
neural network.
Radial basis functions networks have a very strong
mathematical foundation rooted in regularization theory for
solving ill-conditioned problems.
The mapping function of a radial basis function
network, is built up of Gaussians rather than sigmoids as in
MLP networks. Learning in RBF network is carried out in two
phases: first for the hidden layer, and then for the output layer.
The hidden layer is self-organising; its parameters depend on
the distribution of the inputs, not on the mapping from the input
to the output. The output layer, on the other hand, uses
supervised learning (gradient or linear regression) to set its
parameters.
Figure 1. The Radial basis function neural network
In this study we applied different learning rules to the
RBF neural network and studied the optimum performance
with different activation function. We applied Momentum,
Deltabar Delta, Levenberg Marquardt , Conjugate Gradient,
Quick prop learning rule with activation function Tanh, Linear
Tanh, Sigmoid and Linear Sigmoid.
Performance Measures:
The learning and generalization ability of the estimated
NN model is assessed on the basis of important performance
measures such as MSE (Mean Square Error), NMSE
(Normalized Mean Square Error) and r (Correlation
coefficient)
MSE (Mean Square Error):
The formula for the mean square error is:
MSE = ^°"°
NP
... (1)
Where
P = number of output PEs,
N = number of exemplars in the data set,
ij = network output for exemplar i at PE j,
ij = desired output for exemplar i at PE j.
NMSE (Normalized Mean Square Error):
The normalized mean squared error is defined by
the following formula:
PNMSE
NMSE = -
V
Z i=0 y/=o
N
Iy ... (2)
Where
P = number of output processing elements,
N = number of exemplars in the data set,
MSE = mean square error,
ij = desired output for exemplar i at processing
element j.
r (correlation coefficient):
The size of the mean square error (MSE) can be used
to determine how well the network output fits the desired
output, but it doesn't necessarily reflect whether the two sets of
data move in the same direction. For instance, by simply
scaling the network output, the MSE can be changed without
changing the directionality of the data. The correlation
coefficient (r) solves this problem. By definition, the
correlation coefficient between a network output x and a
desired output d is:
zU - x \\ d i ~ d
N
x — — Yx
N
where i=1
... (3)
d=-Yd
and N ^
94
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
The correlation coefficient is confined to the range [-1, 1].
When r = 1 there is a perfect positive linear correlation between
x and d, that is, they co-vary, which means that they vary by
the same amount.
III. Study Area and Data Set
The Upper Wardha catchment area lies directly in the path
of depression movements which originates in the Bay of
Bengal. When the low pressure area is formed in the Bay of
Bengal and cyclone moves in North West directions, many
times this catchment receives very heavy intense cyclonic
precipitation for a day or two. Occurrence of such events have
been observed in the months of August and September.
Rainfall is so intense that immediately flash runoff, causing
heavy flood has been very common feature in this catchment.
For such flashy type of catchment and wide variety in
topography, runoff at dam is still complicated to predict. The
conventional methods also display chaotic result. Thus ANN
based model is built to predict the total runoff from rainfall in
Upper Wardha catchment area for controlling water level of the
dam.
In the initial reaches, near its origin catchment area is hilly
and covered with forest. The latter portion of the river lies
almost in plain with wide valleys.
The catchment area up to dam site is 4302 sq. km. At
dam site the river has wide fan shaped catchment area which
has large variation with respect to slope, soil and vegetation
cover.
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, 2010
Data: Rainfall runoff data for this study is taken from the
Wardha river catchment area which contains a mix of urban
and rural land. The catchments is evenly distributed in eight
zones based on the amount of rainfall and geographical survey.
The model is developed using historical rainfall runoff data ,
provided by Upper Wardha Dam Division Amravati,
department of irrigation Govt, of Maharashtra. Network is
trained by rainfall information gathered from eight telemetric
rain-gauge stations distributed evenly throughout the catchment
area and runoff at the dam site.
The data is received at the central control room online through
this system on hourly basis. The Upper Wardha dam reservoir
operations are also fully automated. The amount of inflow,
amount of discharge is also recorded on hourly basis. From the
inflow and discharge data the cumulative inflow is calculated.
The following features are identified for the modeling the
neural network .
Table I- The parameters used for training the network
Month I RG1 I RG2 I RG3 I RG4 I RG5 I RG6 I RG7 I RG8 I CIF
Month
Rainl to Rain8
Cum Inflow
- The month of rainfall
- Eight rain gauging stations.
- Cumulative inflow in dam
Seven years of data on hourly basis from 2001 to 2007 is
used. It has been found that major rain fall (90%) occurs in the
month of June to October Mostly all other months are dry
hence data from five months. June to October is used to train
the network
Figure 2- Location of Upper Wardha dam on Indian map
IV. Result
The different structures of neural network are
employed to learn the unknown characterization of the system
from the dataset presented to it. The dataset is partitioned into
three categories, namely training, cross validation and test. The
idea behind this is that the estimated NN model should be
tested against the dataset that was never presented to it before.
This is necessary to ensure the generalization. An experiment is
performed at least twenty five times with different random
initializations of the connection weights in order to improve
generalization.
The data set is divided in to training , testing
and cross validation data and the network is trained for all
models of Radial basis function neural network for 5000
epochs.
The performance results obtain on parameters by
applying learning rules Momentum, Deltabar Delta, Levenberg
Marquardt , Conjugate Gradient, Quick prop with activation
function Tanh, Linear Tanh, Sigmoid, Linear Sigmoid are
listed in Table II through Table VI.
Figure 3- The Wardha river catchment
95
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, 2010
Table II- RBF network performance with Momentum learning rule Table V -RBF network performance with Conjugate Gradient
LEARNING RULE
Param
MSE
N
Min
Max
eter
MSE
Abs
Abs
r
1
error
error
Tanh
0.106
0.124
0.034
0.465
0.534
Linear
0.097
0.105
0.024
0.212
0.639
Tanh
Sigmo
0.089
0.093
0.047
0.421
0.678
id
Linear
0.094
0.132
0.041
0.381
0.689
Sigmo
id
Param
eter
4
MSE
N
MSE
Min
Abs
error
Max
Abs
error
r
Tanh
0.094
0.165
0.051
0.312
0.646
Linear
Tanh
Sigmo
id
Linear
Sigmo
id
0.089
0.092
0.094
0.094
0.134
0.124
0.059
0.041
0.064
0.215
0.474
0.541
0.633
0.701
0.732
Table III - RBF network performance with De l t a b a r Delta
LEARNING RULE
Table VI -RBF network performance with Quick prop. learning
RULE
Param
eter
2
MSE
N
MSE
Min
Abs
error
Max
Abs
error
r
Tanh
0.093
0.141
0.051
0.564
0.651
Linear
Tanh
Sigmo
id
Linear
Sigmo
id
0.190
0.143
0.086
0.241
0.215
0.095
0.041
0.032
0.067
0.412
0.495
0.315
0.591
0.543
0.603
Table IV- RBF network performance with L.M. learning rule
MSE
N
Min
Max
eter
MSE
Abs
Abs
r
3
error
error
Tanh
0.076
0.064
0.018
0.143
0.854
Linear
0.086
0.094
0.028
0.298
0.732
Tanh
Sigmo
0.083
0.094
0.020
0.228
0.634
id
Linear
0.089
0.095
0.034
0.469
0.758
Sigmo
id
MSE
N
Min
Max
eter
MSE
Abs
Abs
r
5
error
error
Tanh
0.133
0.245
0.042
0.465
0.584
Linear
0.169
0.212
0.054
0.514
0.601
Tanh
Sigmo
0.106
0.256
0.059
0.329
0.563
id
Linear
0.098
0.112
0.046
0.311
0.609
Sigmo
id
The parameters and performance for RBF model with
different learning rule and activation function are compared on
the performance scale and are listed in the Table VII shown
below. The comparative analysis of the MSE and r (the
correlation coefficient) is done.
Table VII- Comparison of performance parameters
Tanh
Linear Tanh
Sigmoid
Linear
Sigmoid
MSE
r
MSE
r
MSE
r
MSE
r
1
Momentum
0.106
0.534
0.0975
0.639
0.089
0.678
0.094
0.689
2
Deltabar
Delta
0.0931
0.651
0.1906
0.591
0.143
0.543
0.086
0.603
3
L.M.
0.07629
0.854
0.0861
0.732
0.083
0.634
0.0894
0.758
4
Conjugate
Gradient
0.0946
0.646
0.0894
0.633
0.0921
0.701
0.0945
0.732
5
Quickprop
0.1331
0.584
0.1691
0.601
0.106
0.563
0.0986
0.609
After training the network the optimum performance is studied
and it is found that Levenberg Marquardt learning rule and
Tanh activation function produce optimal result. In the Table-
VIII the parameters and the best performances for Radial basis
function neural network are listed.
96
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
Table VIII- RBF network parameters
Parameter
Performance
MSE
0.07629
NMSE
Min Abs Error
0.06431
0.01943
Max Abs Error
0.14387
r
0.85437
Fig 4 shows the plot of actual Vs predicted optimum values for
Radial basis function neural network found with Levenberg
Marquardt learning rule and Tanh activation function.
Actual Vs Predicted Runoff by RBF NN Model
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, 2010
have fewer weights, these networks train extremely fast and
require fewer training samples.
V. Conclusion
An ANN-based short-term runoff forecasting system is
developed in this work. A comparison between five different
learning rules with four activation function for optimal
performance for Radial basis function neural network model is
made. We find that Radial basis function neural network with
Levenberg Marquardt learning rule and Tanh activation
function is more versatile than other approaches studied. Radial
basis function neural network with Levenberg Marquardt
learning rule and Tanh activation function is performing better
as compare to other approaches studied as far as the overall
performance is concerned for forecasting runoff for 3 hrs lead
time. Other approaches studied are also performing optimally.
Which means that static model of Radial basis function neural
network with Levenberg Marquardt learning rule and Tanh
activation function is powerful tool for short term runoff
forecasting for Wardha River basin.
-Actual Runoff -
-Predicted Runoff
Figure 4.- Actual Vs. Predicted runoff by RBF for L.M. and Tanh
The error found in the actual and predicted runoff at the
dam site is plotted for RBF network as shown in the Figure 5.
Error in prediction for RBF NN Model
l
2j£Tl
^S^
24x( (WW
UJ/vA 5
22A//VVM
i^xvA 6
21 1 C3^g^
|fy^| f] ] —♦—error
20 — \^HmI
\^fc^ W 8
is V X* f
\\ VX Yio
17< /\h~
-T\Ml
i>yj
-V 12
14
Fig 5 - Error graph of RBF Model for L.M. and Tanh
The main advantage of RBF is that it finds the input to output
map using local approximators. Each one of these local pieces
is weighted linearly at the output of the network. Since they
Acknowledgment
This study is supported by Upper Wardha Dam Division
Amravati, department of irrigation Govt, of Maharashtra, India
References
[1] P. Srivastava, J. N. McVair, and T. E. Johnson, "Comparison of process-
based and artificial neural network approaches for streamflow modeling
in an agricultural watershed," Journal of the American Water Resources
Association, vol. 42, pp. 545563, Jun 2006.
[2] K. Hornik, M. Stinchcombe, and H. White, "Multilayer feedforward
networks are universal approximators," Neural Netw., vol. 2, pp. 359-
366,1989.
[3] M. C Demirel, A. Venancio, and E. Kahya, "Flow forecast by SWAT
model and ANN in Pracana basin, Portugal," Advances in Engineering
Software, vol. 40, pp. 467-473, Jul 2009.
[4] A. S. Tokar and M. Markus, "Precipitation-Runoff Modeling Using
Artificial Neural Networks and Conceptual Models," Journal of
Hydrologic Engineering, vol. 5, pp. 156-161,2000.
[5] S. Q. Zhou, X. Liang, J. Chen, and P. Gong, "An assessment of the VIC-
3L hydrological model for the Yangtze River basin based on remote
sensing: a case study of the Baohe River basin," Canadian Journal of
Remote Sensing, vol. 30, pp. 840-853, Oct 2004.
[6] R. J. Zhao, "The Xinanjiang Model," in Hydrological Forecasting
Proceedings Oxford Symposium, 1ASH, Oxford, 1980 pp. 351-356.
[7] R. J. Zhao, "The Xinanjiang Model Applied in China," Journal of
Hydrology, vol. 135, pp. 371-381, Jull992.
[8] D. Zhang and Z. Wanchang, "Distributed hydrological modeling study
with the dynamic water yielding mechanism and RS/GIS techniques," in
Proc. of SPIE, 2006, pp. 63591M1-12.
[9] J. E. Nash and I. V. Sutcliffe, "River flow forecasting through
conceptual models," Journal ofHydrology, vol. 273, pp. 282290,1970.
[10] D. Zhang, "Study of Distributed Hydrological Model with the Dynamic
Integration of Infiltration Excess and Saturated Excess Water Yielding
Mechanism." vol. Doctor Nanjing: Nanjing University, 2006, p. 190.529
97
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, 2010
[11] E. Kahya and J. A. Dracup, "U.S. Streamflow Patterns in Relation to the
EI Nit'lo/Southern Oscillation," Water Resour. Res., vol. 29, pp. 2491-
2503,1993.
[12] K. J. Beven and M. J. Kirkby, "A physically based variable contributing
area model of basin hydrology," Hydrologi cal Science Bulletin, vol. 43,
pp. 43-69,1979.
[13] N. J. de Vos, T. H. M. Rientjes, "Constraints of artificial neural
networks for rainfall-runoff modelling: trade-offs in hydrological state
representation and model evaluation", Hydrology and Earth System
Sciences, European Geosciences Union, 2005, 9, pp. 111-126.
[14] Holger R. Maier, Graeme C. Dandy, "Neural networks for the perdiction
and forecasting of water resources variables: a review of modeling issues
and applications", Environmental Modelling & Software, ELSEVIER,
2000, 15, pp. 101-124.
[15] T. Hu, P. Yuan, etc. "Applications of artificial neural network to
hydrology and water resources", Advances in Water Science, NHRI,
1995, 1, pp. 76-82.
[16] Q. Ju, Z. Hao, etc. "Hydrologic simulations with artificial neural
networks", Proceedings-Third International Conference on Natural
Computation, ICNC, 2007, pp. 22-27.
[17] G. WANG, M. ZHOU, etc. "Improved version of BTOPMC model and
its application in event-based hydrologic simulations", Journal of
Geographical Sciences, Springer, 2007, 2, pp. 73-84.
[18] K. Beven, M. Kirkby, "A physically based, variable contributing area
model of basin hydrology", Hydrological Sciences Bulletin, Springer,
1979, 1, pp.43-69.
[19] K. Thirumalaiah, and CD. Makarand, Hydrological Forecasting Using
Neural Networks Journal of Hydrologic Engineering. Vol. 5, pp. 180-
189, 2000.
[20] G. WANG, M. ZHOU, etc. "Improved version of BTOPMC model and
its application in event-based hydrologic simulations", Journal of
Geographical Sciences, Springer, 2007, 2, pp. 73-84.
[21] H. Goto, Y. Hasegawa, and M. Tanaka, "Efficient Scheduling Focusing
on the Duality of MPL Representatives," Proc. IEEE Symp.
Computational Intelligence in Scheduling (SCIS 07), IEEE Press, Dec.
2007, pp. 57-64.
[22] ASCE Task Committee on Application of Artificial Neural Networks in
Hydrology, "Artificial neural networks in hydrology I: preliminary
concepts", Journal of Hydrologic Engineering, 5(2), pp. 115-123, 2000
Rahul Deshmukh received the B.E. and
M.E. degrees in Electronics Engineering from
Amravati University. During 1996-2007, he
stayed in Government College of Engineering,
Amravati in department of Electronics and
telecommunication teaching undergraduate
and postgraduate students. From 2007 till now
he is with Indian Institute of Technology (IIT)
Bombay, Mumbai. His area of reserch are
artificial intelligence and neural networks.
A. A. Ghatol received the B.E. from
Nagpur university foallowed by M. Tech
and P.hd. from IIT Bombay. He is best
teacher award recipient of government of
Maharastra state. He has worked as
director of College of Engineering Poona
and Vice-Chancellor, Dr. Babasaheb
Ambedkar Technological University,
Lonere, Raigad, India. His area of
research is artificial intelligence, neural
networks and semiconductors.
98 http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
Analysis of impact of Symmetric Encryption
Algorithms in Data Security Model of Grid Networks
N. Thenmozhi
Department of Computer Science
N.K.R. Govt. Arts College for Women
Namakkal-637 001, India.
M. Madheswaran
Department of Electronics and Communication Eng
Muthayammal Engineering College
Rasipuram-637 408, India.
Abstract— The symmetric and asymmetric encryption algorithms
are commonly used in grid software to provide necessary
security. The use of symmetric encryption algorithm will
significantly affect the network communication performance.
In this paper, the impact of using different popular and
commonly used symmetric key cryptography algorithms for
encrypting data in a typical grid commuting environment is
analyzed. It is obvious that the use of encryption and decryption
at application layer will certainly have an impact in the
application layer performance in terms of speed. In this work,
we have studied its impact at network layer performance in a
typical grid computing environment in the algorithms such as
DES, Triple DES, AES, Blow Fish, RC2 and RC6. The
performances are measured through simulation studies on ns2
by simulating these algorithms in GARUDA Grid Network
Topology.
Keywords- Grid Security; Encryption; ECGIN; ERNET;
GARUDA; PPlive; GridFTP;
I.
INTRODUCTION
Internet and Grid computing applications are growing
very fast, so the needs to protect such applications have
increased. Encryption algorithms play a main role in
information security systems. On the other side, those
algorithms consume a significant amount of computing
resources such as CPU time, memory, and battery power.
The Globus Toolkit is the very commonly used software
for Grid computing. It provides different kinds of security for
grid computing. The Grid Security Infrastructure (GSI) of
Globus and a Public Key Infrastructure (PKI) provide the
technical framework (including protocols, services, and
standards) to support grid computing with five security
capabilities: user authentication, data confidentiality, data
integrity, non-repudiation, and key management.
A Security Issues
Authentication is the process of verifying the validity of a
claimed individual and identifying who he or she is.
Authentication is not limited to human beings; services,
applications, and other entities may be required to
authenticate also. Basic authentication is the simplest web-
based authentication scheme that works by sending the
username and password within the request. Generally
authentication is achieved through the presentation of some
token that cannot be stolen (forged). This can be either peer-
to-peer relationship (password for client and server) or
through a trusted third party (certification authority or
Kerberos server). Biometrics characteristics can also be used
to a service for authentication purpose, since a unique
identification of human being can give more security for
example a finger print scanner can be used to log into a local
machines. Trust can be defined as the assured reliance on the
character, ability, strength, or truth of someone or something.
Access Control is the ability to limit and control the
access to host systems and applications via communications
links. The process of authorization is often used as a synonym
for access control, but it also includes granting the access or
rights to perform some actions based on access rights.
Data integrity assures that the data is not altered or
destroyed in an unauthorized manner. Integrity checks are
provided primarily via hash functions (or "message digests").
Data confidentiality, Sensitive information must not be
revealed to parties that it was not meant for. Data
confidentiality is often also referred to as privacy. The
standard approach to ensure confidentiality is through
encryption, which is the application of an algorithm that
transforms "plaintext" to "cipher text" whose meaning is
hidden but can be restored to the original plaintext by another
Algorithm (the invocation of which is called decryption).
Key management deals with the secure generation,
distribution, authentication, and storage of keys used in
cryptography. Nonrepudiation refers to the inability of
something that performed a particular action such as a
financial transaction to later deny that they were indeed
responsible for the event.
Basically, security requires at least three fundamental
services: authentication, authorization, and encryption. A grid
resource must be authenticated before any checks can be done
as to whether or not any requested access or operation is
99
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
allowed within the grid. Once the grid resources have been
authenticated within the grid, the grid user can be granted
certain rights to access a grid resource. This, however, does
not prevent data in transit between grid resources from being
captured, spoofed, or altered [18]. The security service to
insure that this does not happen is encryption. Obviously, use
of data encryption certainly will have its impact on
application layer performance. But, in this work we will
examine its impact on total network performance. In this
paper, we will study the impact of four symmetric encryption
algorithms in a typical grid network.
The use of cryptography will certainly have an impact on
network performance in one way or another. So we decided to
model an application layer encryption -decryption scenario in
a typical grid computing environment and study its impact on
network performance through network simulations.
B. Security Methods
Symmetric encryption: Using the same secret key to
provide encryption and decryption of data. Symmetric
cryptography is also known as secret-key cryptography.
Asymmetric encryption: Using two different keys for
encryption and decryption. The public key encryption
technique is the primary example of this using a "public key"
and a "private key" pair. So it is referred as public-key
cryptography.
Secure Socket Layer/Transport Layer Security
(SSL/TLS): These are essentially the same protocol, but are
referred to one another differently. TLS has been renamed by
the IETF, but they are based on the same RFC.
Public Key Infrastructure (PKI): The different
components, technologies, and protocols that make up a PKI
environment. Grid security implementations are
predominantly built on public key infrastructure (PKI)
(Housely et al., 2002; Tuecke et al, 2004). In a PKI each
entity (e.g. user, service) possesses a set of credentials
comprised of a cryptographic key and a certificate.
Mutual Authentication: Instead of using an Lightweight
Distribution Access Protocol (LDAP) repository to hold the
public key (PKI), two parties who want to communicate with
one another use their public key stored in their digital
certificate to authenticate with one another.
C. The symmetric key Encryption Algorithms
Data Encryption Standard(DES), was the first encryption
standard to be recommended by NIST (National Institute of
Standards and Technology). It is based on the IBM proposed
algorithm called Lucifer. DES became a standard in 1974.
Since that time, many attacks and methods were recorded that
exploit the weaknesses of DES, which made it an insecure
block cipher[22].
Advanced Encryption Standard(AES), is the new
encryption standard recommended by NIST to replace DES.
Rijndael (pronounced Rain Doll) algorithm was selected in
1997 after a competition to select the best encryption
standard. Brute force attack is the only effective attack known
against it, in which the attacker tries to test all the characters
combinations to unlock the encryption. Both AES and DES
are block ciphers[20].
Blowfish is a variable length key, the block size is 64 bits,
and the key can be any length up to 448 bits block cipher.
This algorithm can be optimized in hardware applications
though it's mostly used in software applications. Though
it suffers from weak keys problem, no attack is known
to be successful against [8][23].
RC2 is a block cipher with a 64-bits block cipher
with a variable key size that range from 8 to 128 bits. RC2 is
vulnerable to a related-key attack using 234 chosen plaintexts
[20].
Authentication and authorization has been a basic and
necessary Service for internet transactions. Several new
standards have merged which allow dynamic access control
based on exchanging user attributes. Unfortunately, while
providing highly secure and flexible access mechanisms are a
very demanding task. Authentication and Authorization
Infrastructures (AAIs) can provide such integrated federations
of security services. They could, in particular, provide
attribute based access control (ABAC) mechanisms and
mediate customers' demand for privacy and vendors' needs
for information [10].
II.
LITERATURE SURVEY
The Globus Security Infrastructure (GSI) is one of the
most famous security architecture. GSI is based on Public
Key Infrastructure (PKI), which performs mutual
authentication via X.509 certificates. The author describes
present a password-based grid security infrastructure
(PBGSI), which authenticates clients by authenticated key
exchange (AuthA) methods and uses improved Chaffing and
Winnowing for secure data transfer. By using password-based
methods in authentication, authorization and delegation,
PBGSI provides convenient interface for the user. At the
same time, encryption-less secure data transfer improves the
performance; and mechanisms used in our scheme (time-
stamp etc.) enhance the security of the whole grid [11].
A grid environment is built to verify the feasibility and the
efficiency of the extended OCSP protocol. The paper deals
with the running requirement and the data description of the
client and each extended OCSP responder in detail. It
describes the processing algorithm of each responder. In order
to improve the efficiency of the system, the path length
constraint and time constraint of request transmitting are
designed specially. Theory and experiments all prove that the
extended OCSP system improves the efficiency of certificate
verification effectively [12].
Recently, Authentication protocol has been recognized as
an important factor for grid computing security. This paper
[20] described a new simple and efficient Grid authentication
system providing user anonymity. It is based on hash
function, and mobile users only do symmetric encryption and
100
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
decryption and it takes only one round of messages exchange
between the mobile user and the visited network, and one
round of message exchange between the visited network and
the corresponding home network.
There are number of projects investigating attribute-based
authentication such as the VO Privilege Project, GridShib,
and PERMIS. However, there are quite a few decision
dimensions when it comes about designing this scheme in
grid computing [10].
Authentication in the grid environment can be performed
in two ways either in the application layer part or in the
communication part. Cryptography plays a major role to
implement authentication. It is obvious that the use of
encryption and decryption at application layer will certainly
have an impact in the application layer performance in the
grid environment. In this paper, we have simulated the
encryption algorithms in a typical grid network scenario using
the results from the paper [1].
A Europe-China Grid Internetworking (EC-GIN) Project
The Internet communication infrastructure (the TCP/IP
protocol stack) is designed for broad use; as such, it does not
take the specific characteristics of Grid applications into
account. This one-size-fits-all approach works for a number
of application domains, however, it is far from being optimal
general network mechanisms, while useful for the Grid, and
cannot be as efficient as customized solutions. While the Grid
is slowly emerging, its network infrastructure is still in its
infancy. Thus, based on a number of properties that make
Grids unique from the network perspective, the project EC-
GIN (Europe-China Grid Internetworking) will develop
tailored network technology in dedicated support of Grid
applications. These technical solutions will be supplemented
with a secure and incentive-based Grid Services network
traffic management system, which will balance the conflicting
performance demand and the economic use of resources in
the network and within the Grid [30].
By collaboration between European and Chinese partners,
EC-GIN parallels previous efforts for real-time multimedia
transmission across the Internet: much like the Grid, these
applications have special network requirements and show a
special behavior from the network perspective.
B. The ERNET Project
ERNET[26] (Education and Research Network) was the
first dedicated and integrated step taken towards to enable the
research and education community in India to leverage the
benefits of ICTs. ERNET India aims at developing, setting up
and operating nationwide state-of-the-art computer
communication infrastructure and providing services to the
users in academic and research institutions, Government
organizations, and industry, in line with technology
developments and national priorities. Dissemi- nation,
training and knowledge transfer in the field of computer
communication and information technology are an integrating
part of ERNET mission.
ERNET also acts as a bridge for co-operation with other
countries in the area of computer com- munications,
information technology, computer networking and other
related emerging technologies.
The ERNET network has 15 Points of Presence spread
throughout India serving 1389 institutions, including 152
universities, 284 agricultural universities and many other
research organizations. It has 14 points of peering for Internet
bandwidth connectivity using submarine cables.
The network comprises a mix of terrestrial and satellite-
based wide area networks. It provides a wide range of
operation and application services. As of today, universities,
academic institutions, R&D labs and schools, etc. use ERNET
for a variety of applications and services including email, file
transfer, database access, world wide web , web hosting, mail
relaying, security solutions, distant learning and grids.
ERNET is the first network in the country to provide dual
stack access of Internet protocol version 6 (IPv6) and Internet
protocol version 4 (IPv4) test beds to its users to develop, test
and implement IPv6 based mail, Domain name Services, Web
applications and products.
ERNET has deployed many overlay networks over its
terrestrial and satellite network under different schemes.
Some examples are GARUDA (see below), UGC-Infonet,
interconnecting Indian universities, ICAR-Net,
interconnecting Agricultural Research centers, Universities
and Stations, and several pilot projects aiming at
interconnecting schools. Separate networks were
implemented to allow DAE institutes to connect to the
GEANT network and to participate in LHC activities.
ERNET Backbone and POPs
• Backbone Nodes & POPs
■ Proposed POPS
— ft' 34 Mbps
— 2/SMbps
Eh^- international Gateway
£ SATWAN Hub
Figure 1. The ERNET Topology [18]
C. Overview of GAR UDA Project
GARUDA[27] initiative is a collaboration of science
researchers and experimenters on a nation- wide grid of
computational nodes, mass storage and scientific instruments
that aims to provide the technological advances required to
101
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
enable data and compute intensive science of the 21st century.
One of GARUDA's most important challenges is to strike the
right balance between research and the daunting task of
deploying that innovation into some of the most complex
scientific and engineering endeavours being undertaken
today.
The Department of Information Technology (DIT) has
funded the Center for Development of Ad- vanced
Computing (C-DAC[27]) to deploy the nation-wide
computational grid 'GARUDA' which today connects 45
institutions across 17 cities in its Proof of Concept (PoC)
phase with an aim to bring "Grid" networked computing to
research labs and industry. In pursuit of scientific and
technological excel- lence, GARUDA PoC has also brought
together the critical mass of well-established researchers.
GARUDA Grid Component Architecture
Figure 2. GARUDA Grid Component Architecture [29]
C. Present Network Architecture
The GARUDA network is a Layer 2/3 MPLS Virtual
Private Network [VPN] connecting selected institutions at
10/100 Mbps with stringent quality and Service Level
Agreements. The network has been contracted as a man- aged
service with a peak capacity of 2.43 Gbps across 17 cities.
This network is a pre-cursor to the next genera- tion Gigabit
speed nation-wide Wide Area Network with high
performance computing resources and scientific instruments
for seamless collaborative research and ex- periments. The
PoC network was established at all the GARUDA partner
institutes in close collaboration with ERNET who are
responsible for the operation, mainte- nance and management
of this network.
D. Computational Resources in GARUDA
In this collaborative grid project, various resources such
as high performance computing sys- terns (HPC) and satellite
based communication systems have been committed by
different centers of C-DAC and GARUDA partners. It may
be noted that since the resources are diverse in nature, one of
the major challenges of GARUDA is to deploy appropriate
tools and middleware to enable applications to run seamlessly
across the grid.
<S1*
<S 3
^ST^
Figure 3. GARUDA topology - EU-I NDIA GRID [18]
E. Network Simulator
The Grid Computing paradigm has been widely adopted
within the research community for scientific computing. Grid
Computing is used as a method by which access is seamlessly
given to a set of heterogeneous computational resources
across a dynamic set of physical organizations, supplying
massive computing and storage capabilities. Within a Grid
environment, computational jobs are submitted to and run on
suitable resources and data is stored and transferred
transparently without knowing its geographic location. All of
this behavior will obviously show its impact on the
underling network infrastructure and the data generated
within a Grid environment may substantially affect the
network performance due to the volume involved.
We will use NS2 to simulate the network, but it is
well known that NS2 doesn't implement any security
features. Till now, there is no option for simulating security
things in NS2. The reasons for lack of security features in ns2
are:
> Security is a subtle thing related to many aspects, which
is much different from other kinds of network protocols.
> Generally there will not be any real data or packet to
encrypt or decrypt in ns2.
> The scope of a simulation will be minimizing the overall
simulation time. But if we do real encryption or
decryption in simulator, then it will go beyond the
concept of a simulator.
> Lack of support for sending real pay load in ns2.
> Lack of support for handling socket connection like real
TCP/IP scenario.
> Ns2 simulator has limitation in simulating simultaneous
threaded processes to mimic real socket connections.
Ns2 [16] is an object oriented simulator, written in C++,
with an OTCL interpreter as a frontend. The simulator
supports a class hierarchy in C++, and a similar class
hierarchy within the OTcl interpreter. The root of this
102
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
hierarchy is the class TclObject. Users create new simulator
objects through the interpreter. Applications sit on top of
transport agents in ns and there are two basic types of
applications: traffic generators and simulated applications.
Currently, there are four C++ classes derived from the traffic
generator class [20]. Traffic Generator: EXPOO_Traffic,
POO_Traffic, CBR_Traffic, TrafficTrace.
However, none of these classes match the traffic
characteristics of PPLive, and of GridFTP. So we decided to
simulate encryption in ns2 at application layer, by modeling a
new encrypted traffic generator.
III.
MODELING GRID AND GRID TRAFFIC IN NS2
Though there are different kinds of security requirements
or models for grid computing systems, the role of a symmetric
key encryption algorithm and its impact will be a significant
one when implemented in application layer that will affect the
performance in terms of time. In this work, we have simulated
the workload of different Symmetric Key Encryption
algorithms such as DES, Triple DES, AES, Blow Fish, RC2
and RC6 at application layer using Network Simulator tool.
The proposed traffic model is based on the model used in
ECGIN for symmetric key encryption and GridFTP as a cross
traffic. The proposed model is implemented on the Indian grid
network topology GARUDA, to study the impact of the
encryption based traffic model.
A Modeling Encrypted PPLive Traffic
Along with the rapid development of P2P file sharing and
IPTV video services, P2P streaming services have become
a core multi-user video sharing application on the
Internet. The focus of grid technology in the video area is
generally on the resource scheduling and replica
management aspects, while the service traffic
characteristics are still similar to the traditional video service.
In depth work has already been carried out in the areas of
monitoring and modeling video traffic[25]. Therefore,
exploring the developing trends of grid systems, video
sharing, monitoring and the analysis of P2P IPTV traffic are
interesting and promising topics of research.
The time interval between two packets and the size of each
packet waiting for sending out is very important when
modeling actual traffic. Therefore if the model can accurately
match these two characteristics, it can be said to generate
traffic that is similar to the actual data. The EC-GIN project
built a new traffic generator to model the actual traffic called
Lognormal Traffic, which is primarily responsible for
controlling the packets time interval and the packet sizes.
In this work, we extended the traffic model of PPLive
(Lognormal Traffic) to support a simulated encryption-
decryption scenario.
Based on traffic model of EC-GIN, an algorithm has been
put forward to control the packet generation sequence. First,
data initialization is performed as follows:
• Send a video packet when simulation begins.
• Compute the next video packet sending time. Put it
into a variable NextT.
Next, the time needed to send the next packet is computed.
To account for different packet sizes, different parameters are
used to calculate inter-video packet time (variable NextT) and
the inter-control packet time (array t_i). The values of t_l to
t_n are summed to variable SmallT. As long as the value of
SmallT is less than NextT, t_i is used as the inter- packet time
for sending small packets (control packets). Otherwise, a
large packet(video packet) is sent immediately with an inter-
packet time of NextT - (SmallT - t_i).
In addition to the above process, we have delayed the
packet transmission with respect to the size of the packet to
be sent and the selected encryption algorithm.
So the new Scheduled Transmission Time will be equal to
the sum of inter-packet time and the time taken for encrypting
the packet by the selected algorithm.
In our implementation we have simulated the encryption
algorithms in a typical grid network scenario just by including
the encryption delay at the traffic generator using the results
from the paper [1]. In the traffic model of ECGIN, they used
UDP in their design. We have decided to use TCP in our
design, because, TCP is the most commonly used transport
protocol in grid network communication.
B. Modeling GridFTP
The GridFTP tool of Globus Toolkit is one of the most
important components provided by Globus for moving large
amounts of data in bulk. GridFTP is based on FTP, the
highly- popular Internet file transfer protocol. Given the
characteristics of Grid traffic - often a mixture of short,
sporadic service calls and bulk data transfers - a GridFTP
simulation scenario differs from other traffic models and is
therefore important for testing Grid-specific network
mechanisms. The GridFTP simulator of EC-GEN was
developed with the OTCL language to mimic this GridFTP
traffic. The EC-GEN GridFTP is embedded in a gridftp.tcl
file. In this work we just used GridFTP as a background cross
traffic during evaluation the impact of encrypted PPLive
traffic. The three major parameters defined for the GridFTP
simulator are:
• Bandwidth: this parameter is used to set the total
bandwidth of the link. By default, this parameter is set to
1.0Mbps. With this and the ratio parameter, we can
determine the "rate_" parameter for each FTP instance.
• Parallel: this parameter is used to set the parallel
GridFTP streams. By default, this is set to 4. Since each
GridFTP stream can be simulated by FTP, this parameter
will actually set the number of FTP instances for the
GridFTP simulator.
• Ratio: this parameter is used to set the throughput ratio
among the parallel streams. By default, this is set to
1:1:1:1 which means each stream will transmit packets at
an equal speed.
103
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
The GridFTP simulator consists of two classes. One is the
GridFTP class and the other is the GridFTPSink class. We
also override two methods for the basic Simulator class,
attach-agent and connect, with which the GridFTP instance
can be attached to the network node and be connected to the
GridFTPSink instance.
C. The Simulation of GARUDA Network in ns2
The following NAM (Network Animator) output shows
the model of GARUDA network simulated on ns2. The
topology was derived from the information provided by the
ERNET and GARUDA projects [26][27].
Figure 5. The Simulated GARUDA Topology
> The links shown in green are 8/34Mbps links
> The links shown in red are 2/8 Mbps links
> Nodes shown as red hexagon are backbones and POPs
> Nodes shown as blue circles are the connected institutes
IV. SIMULATION RESULTS AND DISCUSSION
A simple model of GARUDA grid network has been
simulated in ns2 and the impact of different encryption
schemes on network performance has been evaluated. A
normal 2 GHz Pentium IV computer with 1 GB RAM was
used for this simulation.
A Traffic models
In order to create the different traffic scenarios files we
used different types of grid traffics mentioned in ECGIN
project. They are GridFTP Traffic and PPLive Traffic.
Some of the simulation parameters are
Number of Backbone and POP nodes 12
Number of Simulated Institution Nodes 36
Routing Protocol DV
Backbone Link Capacity 8/34 Mbps
Institution to Backbone Links 2/8 Mbps
Queue Type DropTail
We have simulated a encrypted PPLive traffic from one
node to another (in this topology, from Madras to Delhi) and
used some GridFTP cross traffic.
B. Performance
The following graph shows the performance of the
network with respect to different cryptography algorithms
used in application layer.
The Throughput
The following graphs show the comparison of throughput
in different encryption schemes over time.
UtouijIifiuUlO 3
Time VS Throughput ■ Comparison
5.0000 10.0000 15.0000
Figure 6. Time VS Throughput- Comparison
The following graph shows the average throughput. The
throughput in the case of Blowfish based scheme was good.
10.0000 15.0000 20.0000 25.QC
Figure 7. The Average Throughput
The Received Packets comparison
The following graphs show the comparison of time and
received packets in different encryption schemes.
104
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
[ciose] | Hdcpy | [^ouT|
Recieved Bytes ;«;1CP
Time VS Recieved Bytes - Comparison
Blown*
RC2
Figure 8. The Time VS Received Packets - Comparison
The End to End Delay
The following graphs show the comparison of end to end
delay in different encryption schemes over time
E2Edelay k1Ct 3
Time VS E2Edelay ■ Comparison
NONE
DES
AES
BlowFish
RC2
2.0000 3.0000 4.0000 5.0000 6.0000 7.0000 8.0000
Figure 9. Time VS E2Edelay - Comparison
The Average Delay
The following graphs show the average delay in different
encryption schemes.
The Average Delay
Encryption Method
Even though all the transmitted packets were received
successfully, the throughput and delay was much affected by
the retransmission of the packets during the packet loss or
drop. This retransmission of packet had an impact on
throughput. Faster the encryption algorithm, higher the
bandwidth it will try to use. So it will increase delay, packet
loss as well as drop at intermediate nodes.
V.
CONCLUSION
Figure 10. The Average Delay
The security is a very important issue in grid network
design. Apart from authentication and authorization, the use
of symmetric encryption algorithm for grid data security is
also having significant impact on the design and performance
of grid networks. A model for grid security infrastructure has
been implemented on network simulator ns2 and the impact
of use of encryption algorithms in network performance has
been measured. We have simulated a simplified model of
GARUDA grid network in ns2 and simulated some of the
basic traffic types of grid network (proposed in ECGIN). As
shown in the graphs in previous section, the use of
cryptography at application layer has obvious impact on
network performance. Depending on the cryptographic
algorithms, the delay in delivery of packet is proportional
with respect to time. Due to queuing delay at the intermediate
node, the faster algorithm provides better throughput with a
little bit of delay in packet delivery.
Future works may address the issues of impact of
asymmetric encryption algorithms used in a grid network for
authentication and other purposes. Further, the work may be
extended for implementing some other traffic types of grid
network.
REFERENCES
[1] Diaa Salama Abd Elminaam, Hatem Mohamed Abdual Kader,
and Mohiy Mohamed Hadhoud, "Evaluating The Performance
of Symmetric Encryption Algorithms" International Journal of
Network Security, Vol.10, No.3, PP.216-222
[2] D. S. Abdul. Elminaam, H. M. Abdul Kader and M. M.
Hadhoud, Performance Evaluation of Symmetric
Encryption Algorithms, Communications of the IBIMA
Volume 8, 2009 ISSN: 1943-7765.
[3] Aamer Nadeem, "A Performance Comparison of Data
Encryption Algorithms", IEEE 2005.
[4] Earle, "Wireless Security Handbook,". Auerbach Publications
2005
[5] Priya Dhawan., "Performance Comparison: Security Design
Choices", Microsoft
[6] Edney, " Real 802.11 Security: Wi-Fi Protected Access and
802.11i",
Addison Wesley 2003.
[7] Hardjono, " Security In Wireless LANS And MANS ", Artech
House
Publishers 2005
[8] Bruce Schneier, "Applied Cryptography", John Wiley & Sons,
Inc 1996
[9] Ronald L. Rivest, M.J.B. Robshaw, R. Sidney, and Y.L. Yin, "
The
105
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
RC6TM, Block Cipher", Version 1.1 - August 20, 1998.
[10] Christian Schlager, Manuel Sojer, Bjorn Muschall, and Giinther
Pernul
, "Attribute-Based Authentication and Authorisation
Infrastructures for E-Commerce Providers", K. Bauknecht et al.
(Eds.): EC-Web 2006,
LNCS 4082, pp. 132 - 141, 2006.
[11] Zhun Cai,"A Password-based Grid Security Infrastructure"
10.1109 /ICDS.2008.39, Second International Conference on
The Digital Society, Institute of Digital Technology
AISINOInc.
[12] Shaomin Zhang, Baoyi Wang, Hebei Province, "Research on
An Extended OCSP Protocol for Grid", Proceedings of
the 7 th World Congress on Intelligent ontrol and
Automation, 25 - 27, 2008, China.
[13] Ronghui Wu, Renfa Li, Fei Yu ,guangxue, Cheng Xu,
"Research on User Authentication for Grid Computing
Security", Proceedings of the Second International Conference
on Semantics, Knowledge, and Grid (SKG'06) 0-7695-2673-
X/06 $20.00 © 2006.
[14] Anna Cinzia Squicciarini, Elisa Bertino and Sebastien
Goasguen, "Access Control Strategies for Virtualized
Environments in Grid Computing Systems", Proceedings of
the 11th IEEE International Workshop on Future Trends of
Distributed Computing Systems (FTDCS'07) 0-7695-2810-
4/07 $20.00 © 2007.
[15] Marty Humphery, Mary R. Thomson, and Keith R.Jackson,
"Security for Grids", Proceeding of the IEEE, Vol 93, No. 3,
pp.644-650, March 2005.
[16] Europe-China Grid InterNetworking, European Sixth
Framework STREP FP6-2006-IST-045256, Deliverable
D2.1, Ns2 code for Grid network simulation. The EC-GIN
Consortium, Europe-China Grid InterNetworking, Survey of
Grid Simulators, Network-level Analysis of Grid Applications,
The EC-GIN Consortium.
[17] International Technical Support Organization, "Introduction to
Grid Computing with Globus", September 2003, IBM
Corporation.
[18] http://partners.euindiagrid.eu/deliverables/D3. 1. html
[19] http://www.faqs.org/rfcs/rfc2828.html
[20] http://msdn2.microsoft.com/en-us/library/ms978415.aspx,
Developer Network October 2002.
[21] http://en.wikipedia.org/wiki/Block_cipher
[22] http://www.tropsoft.com/strongenc/des.htm
[23] http://www.eskimo.com/~weida^enchmarks.html
[24] Coder's Lagoon, http://www.hotpixel.net/software.html
[25] http://www.ec-gin.eu
[26] http://www.eis.ernet.in
[27] www.garudaindia.in
[28] http://www.euindiagrid.eu/
[29] www.cdac.in
[30] ttp://www.euindiagrid.eu/index.php/documents/doc_downloa
d/1 1- einfrastructures-across-europe-and-india
Mrs. N. Thenmozhi is working as Assistant Professor,
Department of Computer Science in N.K.R. Govt. Arts
College for Women, Namakkal. She obtained her Bachelor
degree in Statistics from Saradha College, Salem under
Madras University, Master's degree in Computer
Applications from Bharathiar University, Coimbatore,
Master's degree in Software Systems from BITS, Pilani,and
M.Phil From Manonmaniam Sundaranar University. She is
currently pursuing Ph.D. under Mother Teresa Women's
University, Kodaikanal. She has 18 years of Teaching
Experience and 2 years of Industrial experience. She has
published number papers in various national and international
conferences. She is life member of ISTE. Her field of interest
includes Grid Computing, Network Security and Image
Processing.
M.Madheswaran received the BE Degree from Madurai
Kamaraj University in 1990, ME Degree from Birla Institute
of Technology, Mesra, Ranchi, India in 1992, both in
Electronics and Communication Engineering. He obtained his
PhD degree in Electronics Engineering from the Institute of
Technology,Banaras Hindu University, Varanasi, India, in
1999. At present he is a Principal of Muthayammal
Engineering College, Rasipuram, India. He has authored over
Seventy five research publications in International and
National Journals and Conferences. Currently he is the
chairman of IEEE India Electron Devices Society Chapter.
His areas of interest are theoretical modeling and simulation
of high-speed semiconductor devices for integrated
optoelectronics application, Bio-optics and Bio-signal
Processing. He was awarded the Young Scientist Fellowship
(YSF) by the State Council for Science and
Technology,TamilNadu, in 1994 and Senior Research
Fellowship (SRF) by the Council of Scientific and Industrial
Research (CSIR), Government of India in 1996. Also he has
received YSF from SERC, Department of Science and
Technology, Govt, of India. He is named in Marquis Who's
Who in Science and engineering in the year 2006. He is a
Member of Institute of Electrical and Electronics Engineers,
Fellow of Institution of Electronics and Telecommunication
Engineers, Member of Indian Society for Technical Education
and Member of Institution of Engineers.
AUTHORS PROFILE
106
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
Low Power and Area Consumption Custom Networks-On-Chip
Architectures Using RST Algorithms
^.Ezhumali 2 Dr.C.Arun
Professor, Dept of Computer Science Engineering
2 Asst. Professor, Dept of Electronics and Communication
Ralalakshmi Engineering College, Thandalam-602 105, Chennai, India
Abstract: Network-on-Chip (NoC)
architectures with optimized topologies have
been shown to be superior to regular
architectures (such as mesh) for application
specific multiprocessor System-on-Chip
(MPSoC) devices. The application specific NoC
design problem takes, as input the system-level
floorplan of the computation architecture .The
objective is to generate an area and power
optimized NoC topology. In this work, we
consider the problem of synthesizing custom
networks-on-chip (NoC) architectures that are
optimized. Both the physical links and routers
determine the power consumption of the NoC
architecture. Our problem formulation is based
on the decomposition of the problem into the
inter-related steps of finding good flow
partitions, and providing an optimized network
implementation for the derived topologies. We
used Rectilinear-Steiner-Tree (RST)-based
algorithms for generating efficient and
optimized network topologies. Experimental
results on a variety of NoC benchmarks showed
that our synthesis results were achieve reduction
in power consumption and average hop count
over different mesh implementations. We
analyze the quality of the results and solution
times of the proposed techniques by extensive
experimentation with realistic benchmarks and
comparisons with regular mesh-based NoC
architectures.
Index Terms — Multicast routing, network-on-
chip (NoC), synthesis, system-on-chip (SoC),
topology.
l.Introduction
Network-on-Chip (NoC) is an emerging
paradigm for communications within large
VLSI systems implemented on a single silicon
chip. The layered-stack approach to the design
of the on-chip intercore communications is the
Network-on-Chip (NOC) methodology. In a
NoC system, modules such as processor cores,
memories and specialized IP blocks exchange
data using a network as a "public
transportation" sub-system for the information
traffic. A NoC is constructed from multiple
point-to-point data links interconnected by
switches (a.k.a. routers), such that messages
can be relayed from any source module to any
destination module over several links, by
making routing decisions at the switches.
A NoC is similar to a modern
telecommunications network, using digital bit-
packet switching over multiplexed links.
Although packet switching is sometimes
claimed as necessity for a NoC, there are several
NoC proposals utilizing circuit-switching
techniques. This definition based on routers is
usually interpreted so that a single shared bus, a
single crossbar switch or a point-to-point
network is not NoCs but practically all other
topologies are. This is somewhat confusing
since all above-mentioned are networks (they
enable communication between two or more
devices) but they are not considered as network-
on-chips. Note that some erroneously use NoC
as a synonym for mesh topology although NoC
paradigm does not dictate the topology.
Likewise, the regularity of topology is
sometimes considered as a requirement, which
is, obviously, not the case in research
concentrating on "application-specific NoC
topology synthesis".
107
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
1F=%=%
■Core
^»=^=^=^i
$=
Network Adapter
Routing Node
-Link
1<H=^=^H=^
figure. 1 Topological illustration of a
4-by-4 grid structured NoC.
The wires in the links of the NoC are shared
by many signals. A high level of parallelism
is achieved, because all links in the NoC can
operate simultaneously on different data
packets. Therefore, as the complexity of
integrated systems keeps growing, a NoC
provides enhanced performance (such as
throughput) and scalability in comparison
with previous communication architectures
(e.g., dedicated point-to-point signal wires,
shared buses, or segmented buses with
bridges). Of course, the algorithms must be
designed in such a way that they offer large
parallelism and can hence utilize the
potential of NoC.
the complexity of designing wires for
predictable speed, power, noise, reliability,
etc., because of their regular, well-controlled
structure. From a system design viewpoint,
with the advent of multi-core processor
systems, a network is a natural architectural
choice. A NoC can provide separation between
computation and communication; support
modularity and IP reuse via standard
interfaces, handle synchronization issues,
serve as a platform for system test, and, hence,
increase engineering productivity.
Although NoCs can borrow concepts and
techniques from the well-established domain
of computer networking, it is impractical to
blindly reuse features of "classical" computer
networks and symmetric multiprocessors. In
particular, NoC switches should be small,
energy-efficient, and fast. Neglecting these
aspects along with proper, quantitative
comparison was typical for early NoC
research but nowadays they are considered in
more detail. The routing algorithms should
be implemented by simple logic, and the
number of data buffers should be minimal.
Network topology and properties may be
application-specific. Research on NoC is now
expanding very rapidly, and there are several
companies and universities that are involved.
Figure 1 shows how a NoC, in comparison
with shared buses, could be occupied with
various components as resources
Traditionally, ICs have been designed with
dedicated point-to-point connections, with one
wire dedicated to each signal. For large
designs, in particular, this has several
limitations from a physical design viewpoint.
The wires occupy much of the area of the chip,
and in nanometer CMOS technology,
interconnects dominate both performance and
dynamic power dissipation, as signal
propagation in wires across the chip requires
multiple clock cycles. NoC links can reduce
2.EXISTING RELATED WORKS
So far, the communication problems faced
by System on chip were tackled by making use
of regular Network on chip architectures. The
following are the list of popular regular NoC
architectures:
Mesh Architecture.
Torus Architecture.
Butterfly Fat Tree Architecture.
108
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
Extended Butterfly Fat Tree Architecture
The NoC design problem has received
considerable attention in the literature. Towles
and Dally [1] and Benini and De Micheli [2]
motivated the NoC paradigm. Several existing
NoC solutions have addressed the mapping
problem to a regular mesh-based NoC
architecture [3], [4]. Hu and Marculescu [3]
proposed a branch-and-bound algorithm for
the mapping of computation cores on to mesh-
based NoC architectures. Murali et ah [4]
described a fast algorithm for mesh-based NoC
architectures that considers different routing
functions, delay constraints, and bandwidth
requirements. On the problem of designing
custom NoC architectures without assuming
existing network architecture, a number of
techniques have been proposed [5]-[10]. Pinto
et ah [7] presented techniques for the
constraint-driven communication architecture
synthesis of point-to-point links by using
heuristic-based -way merging. Their technique
is limited to topologies with specific structures
that have only two routers between each
source and sink pair. Ogras et ah [5], [6]
proposed graph decomposition and long link
insertion techniques for application-specific
NoC architectures. Srinivasan et ah [8], [9]
presented NoC synthesis algorithms that
consider system-level floor planning, but their
solutions only considered solutions based on a
slicing floorplan where router locations are
restricted to corners of cores and links run
around cores. Murali et ah [10] presented an
innovative deadlock-free NoC synthesis flow
with detailed backend integration that also
considers the floorplanning process. The
proposed approach is based on the min-cut
partitioning of cores to routers. This work
presents a synthesis approach based on a set
partitioning formulation that considers
multicast traffic, Although different in
topology and some other aspects, all the above
papers essentially advocate the advantages of
using NoCs and regularity as effective means
to design high performance SoCs. While these
papers mostly focus on the concept of regular
NoC architecture (discussing the overall
advantages and challenges), to the best of our
knowledge, our work is better than previous
custom NoC synthesis formulations and
efficient way to solve it.
PROPOSED SYSTEM
3.1 PROBLEM DEFINITION
• We consider the problem of synthesizing
custom networks-on-chip (NoC)
architectures that are optimized for a
given application.
• We divide the problem statement into
the flowing interrelated steps:
Physical topology Construction.
Power and Area Comparisons
3.2 SYSTEM ARCHITECTURE
Input
Specification J
Networks-on-
chip Synthesis
♦
NoC Power -\
and Area
Estimation J
Topology
Generation
Figure. 2 Proposed System Architecture
Our NoC synthesis design flow is depicted in
Figure 2. The major elements in the design
flow are elaborated as follows.
Input Specification: The input specification
to our design flow consists of a list of
modules. As observed in recent trends, many
modern SoC designs combine both hard and
soft modules as well as both packet-based
network communications and conventional
109
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
wiring. Modules can correspond to a variety
of different types of intellectual property (IP)
cores such as embedded microprocessors,
large embedded memories, digital signal
processors, graphics and multimedia
processors, and security encryption engines,
as well as custom hardware modules. These
modules can come in a variety of sizes and
can be either hard or soft macros, possibly as
just black boxes with area and power
estimates and constraints on aspect ratios. To
facilitate modularity and interoperability of
IP cores, packet-based communication with
standard network interfaces is rapidly gaining
adoption. Custom NoC architectures are
being advocated as a scalable solution to
packet-based communication. In general, a
mixture of network-based communications
and conventional wiring may be utilized as
appropriate, and not all inter-module
communications are necessarily over the on-
chip network. For example, an embedded
microprocessor may have dedicated
connections to its instruction and data cache
modules. Our design flow and input
specification allow for both interconnection
models. Below is an example of a
communication demand graph:
Figure 3 Sample Input Specification
NoC Synthesis: Given input specification
information, the NoC synthesis step then
proceeds to synthesize a NoC architecture
that is optimized for the given specification.
Consider the above diagram that depicts a
small illustrative example. It only shows the
portion of the input specification that
corresponds to the network-attached modules
and their traffic flows. The nodes represent
modules, edges represent traffic flows, and
edge labels represent the length of the two
vertices. The NoC Synthesis generates
topologies based on the communication
demand graph and comparing with
parameters like power consumption and area
usage chooses the best architecture. Below is
an example of two architectures generated
based on the given CDG.
r\ 200
>v5
\j
v2 I
r ^00
3v6
J PL
200 1
200 1
200 v ^
Figure 4 Sample Topologies Generated
NoC Power and Area Estimation: To
evaluate the power and area of the
110
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
synthesized NoC architecture, we use a state-
of the- art NoC power-performance simulator
called Orion that can provide detailed power
characteristics for different power
components of a router for different
input/output port configurations. It accurately
considers leakage power as well as dynamic
switching power, which is important since it
is well known that leakage power is
becoming an increasingly dominating. Orion
also provides area estimates based on a state-
of-the-artrouter microarchitecture.
MODULE DESCRIPTION
s~ ~ x
Flow Partitioning
Steiner Tree Based Topology Construction
Implementation Optimization
Figure 5 Formulation of Synthesis Problem
4.1 Flow Partitioning
Flow partitioning is performed in
the outer loop of our synthesis formulation to
explore different partitioning of flows to
separate subnetworks. We make use of the
following algorithm to implement flow
partitioning:
4.2 STEINER TREE BASED
TOPOLOGY CONSTRUCTION
For each flow partition considered, physical
network topologies must be decided. In current
process technologies, layout rules for
implementing wires dictate physical topologies
where the network links run horizontally or
vertically. Thus, the problem is similar to
Rectilinear Steiner Tree (RST) problem that has
been extensively studied for the conventional
VLSI routing problem. Given a set of nodes, the
RST problem is to find a network with the
shortest edge lengths using horizontal and
vertical edges such that all nodes are
interconnected. The RST problem is well
studied with very fast implementations
available. We create an RST solver in the inner
loop of flow partitioning to generate topologies
for the set partitions considered.
[ ii put: G"(V ? E. TV. A): communication demand graph
G: specified evaluation function for implementation cost
Output: T: synthesized network architecture
I: initialize P Q =
2: for all e k £ E
4: ccEtCK}) = EvaluatcOD5t(r({e fc }) ? C)
5: end for
& t -
7: while F*\ > 1 do
S: for all g u ,g v fc P* do
A«. = At U 0„
T{ff uv ) = SolvcRST^,)
cestui,) = EvaluafccCcB^X^,), C)
P($u>9v) = «*tCffi«0 + 5^e*» *?*„*, «»tCft)
end for
(n : j>) = airsmin^,^/* }%$^gj)
# = *+!
IS: end while
19: foralU G [0,^-1] do
20: <F*) = E, u ^™t(^)
21: #Ghi[P i ]=X) 9ue i-A9u)
22: end for
23: t — itfgMm^emn-ii rfP'-)
24: T ~ £ol?i[F t \
23: return T
9:
JO:
11:
\2:
13:
14:
15:
J 0.
J7;
ena lor
Figure 6 Flow Partitioning Algorithm
IMPLEMENTATION RESULTS
5.1. EXPERIMENTAL SETUP
We have implemented our
proposed algorithm in C. In our
implementation, we have designed a
Rectilinear Steiner Tree solver to generate
the physical network topologies in the inner
loop of the algorithm. Simulator ORION 2.0
does the power and area estimates. The
Results obtained are shown in a line chart for
mere comparisons. A snapshot of the all the
results have been shown later in this chapter.
All experimental results were obtained on a
3.06-GHz Intel P4 processor machine with
512 MB of memory running Linux.
111
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
5.2. EXPERIMENTAL RESULTS
ALL FSTs: 64 Points
Figure 7 Snapshot of ALL The FSTs
Generated
— -i
V. L
1"
■^-p
^
^r
■ .
1
i ,
L-' ■
Steiner Minimal Tree: 64 Points, length =
56729
Figure 8 Steiner Minimal Tree Generated
Method of Evaluation: In all our experiments,
we aim to evaluate the performance of the
proposed algorithms. On all benchmarks with
the objective of minimizing the total area as
well as power consumption of the synthesized
NoC architectures. The total area as well as
power consumption includes all network
components. We applied the design parameters
of 1 GHz clock frequency, 4-flit buffers, and
128-bit flits. For evaluation, fair direct
comparison with previously published NoC
synthesis results is difficult in part because of
vast differences in the parameters assumed. To
evaluate the effectiveness of our algorithms, we
have the full mesh implementation for each
benchmark for comparison from previous
published papers have been taken. These
comparisons are signified to show the benefits
of custom NoC architectures.
Table 1. NOC Power Comparisons
S.No
V ertices
Custom Power
Mesh Power
Opt MeshPower
1
6
0.0416
0.0990
0.0430
2
7
0.0432
0.1000
0.0500
3
8
0.0494
0.1780
0.0600
4
11
0.0617
0.2570
0.1220
5
12
0.0663
0.2720
0.1520
6
14
0.0848
0.3760
0.1310
7
20
0.0987
0.4950
0.1540
S
24
0.1034
0.6330
0.2020
9
25
0.1034
0.6420
0.2600
10
36
0.1203
1.0080
0.3000
11
44
0.1353
1.4250
0.3440
Power Comparison
£ o.eooo
: :::
/
.* — -♦
__— — ~
_..-*■ ■
:=I_"II~~-»^-~i^-i
■— ■— ■
-tuaom Mnwar
Figure 9 NoC Power Comparisons
The area results, power results, the execution
times, and area as well as power
improvements of that algorithm are reported.
The results show the algorithm can
efficiently synthesize NoC architectures that
minimize power and area consumption as
compared with regular topologies such as
mesh and optimized mesh topologies.
112
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
Table 2. NoC Area Comparisons
S.No
1
2
Vertice
s
6
7
Custom
Area
0.1543
0.1557
Opt.
Mesh
Area
0.31
0.43
3
8
0.1810
0.41
4
12
0.2252
0.48
5
14
0.3060
0.91
6
20
0.3563
1.03
1
24
0.3732
1.19
3
25
0.3753
1.81
9
36
0.4332
1.81
10
44
0.493(5
2.01
NoC Area Comparisons
-Custom Area
-opt. Mesh Area
6 7 8 12 14 20 24 25 36 44
Vertices
Figure 10. NoC Area Estimates
Thus, the above two line charts in
figure 9 and 10 clearly show a reduction in
power and area estimates of custom NoC
with mesh and optimized mesh topologies.
Mesh topologies was explained in chapter 2.
Eliminating router ports and links that are not
used forms optimized mesh topologies. The
power reduction is at an average of 83.43
percent and 50 percent as compared to mesh
and optimized mesh topologies respectively.
The area reduction is at an average of 70.95
percent as compared to optimized mesh
topologies.
6.CONCLUSION AND FUTURE WORK
In this research Works have been carried out
in context related to Regular topologies like
mesh, torus and etc. This work presented an
idea on building customizing network on
chip with the better flow partitioning and
also considered power and area reduction as
compared to the already presented Regular
topologies, we proposed a formulation of the
custom NoC synthesis problem based on the
decomposition of the problem into the inter-
related steps of deriving a good physical
network topology, and providing an
comparison in terms of area and power with
the well established regular topologies. We
used the algorithm called CLUSTER for
systematically examining different possible
set partitioning of flows, and we proposed
the use of RST algorithms for constructing
good physical network topologies. Our
solution framework enables the decoupling
of the evaluation cost function from the
exploration process, thereby enabling
different user objectives and constraints to be
considered. Although we use Steiner trees to
generate a physical network topology for
each group in the set partition, the final NoC
architecture synthesized is not necessarily
limited to just trees as Steiner tree
implementations of different groups may be
connected to each other to form non-tree
structures.
This work does not differentiate the
routers/switches (communication modules)
with the operating modules present in the
chip. In near future, the work of identifying
the best placement of routers and minimizing
the number of routers and also the effectives
of the customized Network on Chip in terms
of other parameters like throughput, latency.
Link utilization and buffer utilization can be
taken into account.
113
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
REFERENCES
[1] Shan Yan, Bill Lin, " Custom Networks-
on-Chip Architectures With Multicast
Routing," IEEE transactions on very large
scale integration (VLSI) systems, vol. 17, no.
3, march 2009.
[2] K. Srinivasan, K. S. Chatha, and G.
Konjevod, "Linear-programming based
techniques for synthesis of network-on-chip
architectures," IEEE Trans. Very Large Scale
Integr. (VLSI) Syst., vol. 14, no. 4, pp. 407-
420, Apr. 2006.
[3] K. Srinivasan, K. S. Chatha, and G.
Konjevod, "Application specific network-on-
chip design with guaranteed quality
approximation algorithms," in Proc.
ASPDAC, 2007, pp. 184-190.
[4] S. Murali, P. Meloni, F. Angiolini, D.
Atienza, S. Carta, L. Benini, G . De Micheli,
and L. Raffo, "Designing application-specific
networks on chips with floor plan
information," in Proc. ICCAD, 2006, pp.
355-362.
[8] D. Greenfield, A. Banerjee, J. -G. Lee,
and S. Moore, "Implications of rent's rule for
NoC design and its fault-tolerance," in Proc.
NOCS, May 2007, pp. 283-294.
[9] S. Yan and B. Lin, "Application-specific
network-on-chip architecture synthesis based
on set partitions and Steiner trees," in Proc.
ASPDAC, 2008, pp. 277-282.
[10] Xilinx, San Jose, CA, "UMC delivers
leading-edge 65 nm FPGAs toXilinx," Des.
Reuse, Nov. 8, 2006 [Online]. Available:
http://www.design-
reuse.com/news/14644/umc-edge-65nm-
fpgas-xilinx.html
[11] P. Gratz, K. Sankaralingam, H. Hanson,
P. Shivakumar, R.McDonald, S. W. Keckler,
and D. Burger, "Implementation and
evaluation of a dynamically routed processor
operand network," in Proc. NOCS, May
2007, pp. 7-17.
[12] N. Enright-Jerger, M. Lipasti, and L.-S.
Peh, "Circuit-switched coherence," IEEE
Computer. Arch. Lett. vol. 6, no. 1, pp. 193-
202, Mar. 2007.
[5] L. Zhang, H. Chen, H. Chen, B. Yao, K.
Hamilton, and C.-K. Cheng, "Repeated on-
chip interconnect analysis and evaluation of
delay, power, and bandwidth metrics under
different design goals," in Proc. ISQED,
2007, pp. 251-256.
[6] R. Mullins, "Minimizing dynamic power
consumption in on-chip networks," in Proc.
Int. Symp. Syst.-on-Chip, 2006, pp. 1-4.
[7] C. -W. Lin, S. -Y. Chen, C. -F. Li, Y. -
W. Chang, and C. -L. Yang, "Efficient
obstacle-avoiding rectilinear Steiner tree
construction," in Proc. Int. Symp. Phys. Des.
2007, pp. 127-134.
[13]. Shan Yan, Student Member, IEEE, and
Bill Lin, Senior Member, IEEE "Custom
Networks-on-Chip Architectures With
Multicast Routing" IEEE Transactions On
Very Large Scale Integration (VLSI)
Systems, Vol. 17, No. 3, Pp 342-355, March
2009.
9
S .J
(M.Tech.,)
Engineering
Hyderabad,
Ezhumalai Periyathambi
received the B.E degree in
Computer Science and
engineering from Madras
University, Chennai , India in
1992 and Master Technology
in computer science and
from J N T University,
India in 2006. He is currently
114
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
working towards the Ph.D degree in
Department of Information and
Communication, Anna University, Chennai,
India. He is working as Professor in the
Department of Computer Science and
Engineering , Rajalakshmi Engineering
College, Chennai, Tamilnadu, India. His
research in reconfigurable architecture, Multi-
Core Technology CAD - Algorithms for VLSI
Architecture. Theoretical Computer Science.
And mobile computing.
1 1 5 http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
Prediction of Epileptic form Activity in Brain
Electroencephalogram Waves using
Support vector machine
Tavithra Devi S T
M.Phil Research Scholar
PSGR Krishnammal College for Women
Coimbatore Tamilnadu, INDIA
2 VijayaMS
Assistant Professor and Head
GRG School of Applied Computer
Technology
PSGR Krishnammal College for Women
Coimbatore Tamilnadu, INDIA
ABSTRACT
Human brain is a highly complex structure composed of
millions of nerve cells. Their activity is usually well organized
with mechanisms for self-regulation. The neurons are
responsible for a range of functions, including consciousness
and bodily functions and postures. A sudden temporary
interruption in some or all of these functions is called a
seizure. Epilepsy is a brain disorder that causes people to have
recurring seizures. Electroencephalogram (EEG) is an
important diagnostic test for diagnosing epilepsy because it
records the electrical activity of the brain. This paper
investigates the modeling of epilepsy prediction using Support
Vector Machine, a supervised learning algorithm. The
prediction model has been employed by training support
vector machine with evocative features derived from EEG
data of 324 patients and from the experimental results it is
observed that the SVM model with RBF kernel produces 86%
of accuracy in predicting epilepsy in human brain.
Keywords
Support Vector Machine, Epilepsy, Prediction, Supervised
Learning.
1. INTRODUCTION
Epilepsy is a disorder characterized by recurrent
seizures of cerebral origin, presenting with episodes of
sensory, motor or autonomic phenomenon with or without
loss of consciousness. Epilepsy is a disorder of the central
nervous system, specifically the brain [1]. Brain is one of
the most vital organs of humans, controlling the
coordination of human muscles and nerves. Epileptic
seizures typically lead to an assortment of temporal
changes in perception and behavior. Based on the
physiological characteristics of epilepsy and the
abnormality in the brain, the kind of epilepsy is determined.
Epilepsy is broadly classified into absence epilepsy, simple
partial, complex partial and general epilepsy. Absence
epilepsy is a brief episode of staring. It usually begins
between ages 4 and 14. It may also continue to adolescence
or even adulthood. Simple partial epilepsy affects only a
small region of the brain, often the hippocampus. It can
also include sensory disturbances, such as smelling or
hearing things that are not there, or having a sudden flood
of emotions. Complex partial epilepsy usually starts in a
small area of the temporal lobe or frontal lobe of the brain.
In general epilepsy the patient becomes unconscious the
patient has a general tonic contraction of all their muscles,
followed by alternating colonic contractions. It affects the
entire brain.
Various diagnostic techniques like Computed
Tomography (CT), Magnetic Resonance Imaging (MRI),
Electroencephalogram (EEG), and Positron Emission
Tomography (PET) are commonly presented.
Electroencephalography (EEG) is the recording of
electrical activity along the scalp produced by the firing of
neurons within the brain. In clinical contexts, EEG refers to
the recording of the brain's spontaneous electrical activity
over a short period of time, usually 20^10 minutes, as
recorded from multiple electrodes placed on the scalp. The
Electroencephalograph (EEG) signal is one of the most
widely signal used in the bioinformatics field due to its rich
information about human tasks for epilepsy identification
because of its characteristics like frequency range, spatial
distributions and peak frequency. EEG waves are observed
by neurologists based on spectra waveform of the signal to
identify the presence of epilepsy.
Machine learning provides methods, techniques and
tools, which help to learn automatically and to make
accurate predictions based on past observations. Current
empirical results prove that machine learning approach is
well-matched for analyzing medical data and machine
learning techniques produce promising research results to
medical domains.
Forrest Sheng Bao carried out the work and developed
a neural network based model for Epilepsy diagnosis using
EEG [1]. Piotr Mirowski carried out the work and
implemented a model based on classification of patterns of
EEG synchronization for seizure prediction using neural
network [2]. Suleiman A.B. R. proposed a new approach
for describing and classifying the EEG brain natural
oscillations such as delta, theta, alpha, and beta frequencies
116
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
using Wigner-Ville analysis with Choi-Willians filtering
and neural network [3].
The motivation behind the research reported in this
paper is to predict the presence of epilepsy in human brain.
Supervised learning technique, a kind of machine learning
algorithm is used to model the epilepsy prediction problem
as classification task to assist physician for accurate
prediction of epilepsy in patients.
In this paper, the prospective benefits of supervised
learning algorithm namely support vector machine are
made use of for the computerized prediction of epilepsy.
The proposed SVM based epilepsy prediction model is
shown in Figure 1 .
Feature Extraction using
Wavelet Toolbox in
MATLAB
SVM Training
SVM Based
Prediction model
Prediction
Figure 1. Proposed SVM based epilepsy prediction model
2. DATA ACQUISITION
EEGs show continuous oscillating electric activity. The
amplitude and the patterns are determined by the overall
excitation of the brain which in turn depends on the activity
of the reticular activating system in the brain stem.
Amplitudes on the surface of the brain can be up to 10 mV,
those on the surface of the scalp range up to 100 mV.
Frequencies range from 0.5 to 100 Hz. The pattern changes
markedly between states of sleep and wakefulness. Distinct
patterns are seen in epilepsy and five classes of wave
groups are described as alpha, beta, gamma, delta and
theta.
• Alpha waves contain frequencies
between 8 and 13 Hz with amplitude less than 10
mV. It found in normal people who are awake and
resting quietly, not being engaged in intense
mental activity. Their amplitude is highest in the
occipital region. When the person is asleep, the
alpha waves disappear. When the person is alert
and their attention is directed to a specific activity,
the alpha waves are replaced by asynchronous
waves of higher frequency and lower amplitude.
• Beta waves have a frequency range of 14
to 22 Hz, extending to 50 Hz under intense mental
activity. It has their maximum amplitude (less than
20 mV) on the parietal and frontal regions of the
scalp. There are two types: beta I waves, lower
frequencies which disappear during mental
activity, and beta II waves, higher frequencies
which appear during tension and intense mental
activity.
• Gamma waves have frequencies between
22 and 30 Hz with amplitude of less than 2 mV
peak-to-peak and are found when the subject is
paying attention or is having some other sensory
stimulation.
• Theta waves have a frequency range
between 4 to 7 Hz with amplitude of less than 100
mV. It occurs mainly in the parietal and temporal
regions in sleep and also in children when awake,
and during emotional stress in some adults,
particularly during disappointment and frustration.
Sudden removal of something causing pleasure
will cause about 20 s of theta waves.
• Delta waves have frequency content
between 0.5 and 4 Hz with an amplitude less than
100 mV. It occurs during deep sleep, during
infancy and in serious organic brain disease. They
will occur after transactions of the upper brain
stem separating the reticular activating system
from the cerebral cortex. They are found in the
central cerebrum, mostly the parietal lobes.
Five sets of images namely Normal Epilepsy, Absence
Epilepsy, Simple Partial Epilepsy, Complex Partial
Epilepsy and General Epilepsy are taken into
consideration.
3. FEATURE EXTRACTION
Feature extraction process plays a very important role
on the classification. Fourier transformation method,
discrete transformation method and continuous
transformation methods are normally available to extract
features that characterize EEG signals. The wavelet
transform (WT) provides very general techniques which
can be applied to many tasks in signal processing. Wavelets
are ideally suited for the analysis of sudden short-duration
signal changes.
In the proposed model, EEG signal analysis and feature
extraction have been performed using Discrete Wavelet
Transform (DWT). The DWT is a extraordinary case of the
WT that provides a compact representation of a signal in
time and frequency that can be computed efficiently.
117
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
The DWT is defined by the following equation:
The energy is computed using E is given by
¥ (a , b) (t) = 2 a/2 ¥ (2 a/2 (t-b))
(1)
E=2X7N
(3)
where a is a scales and b is positions of the wavelet
mother v|/ (t) is a time function with finite energy. Choosing
scales and positions are based on powers of two, which are
called dyadic scales and positions (a j=2 J; bj k =2 J k) ( j and
k integers). Equation (1) shows that it is possible to build a
wavelet for any function by dilating a function \|/ (t) with a
coefficient 2 j, and translating the resulting function on a
grid whose interval is proportional to 2-j.
The selection of appropriate wavelet and the number of
decomposition levels is very important in analysis of
signals using the WT. The number of decomposition levels
is chosen based on the dominant frequency components of
the signal. The levels are chosen such that those parts of the
signal that correlate well with the frequencies required for
classification of the signal are retained in the wavelet
coefficients. The smoothing feature of the Daubechies
wavelet of order 2 (db2) made it more suitable to detect
changes of the signals. Thus, the wavelet coefficients are
computed using db2. The frequency bands corresponding
to different levels of decomposition for db2 with a
sampling frequency of 256 Hz. The discrete wavelet
coefficients are computed using the MATLAB wavelet
toolbox.
The purpose of feature extraction is to reduce the size
of the original dataset by measuring certain properties or
features that distinguish one input pattern from another.
The various measurements based on statistical features
from EEG are extracted. The extracted features provide the
characteristics of the input type to the classifier by
considering the description of the relevant properties of the
signals into a feature space.
The statistical feature of the wavelet coefficients in
each subband such as energy, entropy, Minimum subband,
maximum subband, mean, and standard deviation are used
to investigate the adequacy for the discrimination of normal
and abnormal patients. The following statistical features
have been derived using the following.
Entropy is the diminished capacity for spontaneous
changes in signals.
where Xi is signal value, values are present in waves is
denoted as n. Total number of signal is N
Maximum Subband - It generate maximum of the
wavelet coefficients in each subband is calculated using
M ax =Max( Xl ) (4)
where max (xi) is maximum number of signal value.
Mean - It is defined as average value of a distribution
of the wavelet coefficients in each subband which is given
by
E=£xi/N
(5)
where xi is signal and total number of signal is present
in the wavelet is N
Minimum Subband - calculate minimum of the
wavelet coefficients in each subband is defined as
M in =Min(xO (6)
Where min (xi) is minimum number of signal value.
Standard deviation - standard deviation of each
subband is defined as o .This feature provide information
about the amount of change of the frequency distribution.
°=I(x-u) 2
(7)
Entropy = J] P(i, j) log P(i 9 j)
(2)
Where P(i, j) reflects the distribution of the probability
of occurrence of each signal (i , j are integer).
Energy - Provides the sum of squared elements in the
wavelet. This is also known as uniformity or the angular
second moment.
where £ is sum of squared elements in the wavelet,x is
signal value and |i is a mean of the corresponding
signal(xi).
Thus a total of 21 statistical feature are extracted from
EEG signal for each subband for preparing dataset.
118
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
4. SUPPORT VECTOR MACHINE
Support Vector Machine (SVM) is a kind of learning
machine based on statistical learning theory. SVM is
basically applied to model pattern classification task. SVM
first, maps the input vectors into feature vectors in feature
space with a higher dimension, either linearly or non-
linearly. Then, within the feature space SVM constructs a
hyperplane which separates two classes. SVM training
always seeks a global optimized solution and avoids over-
fitting, thus it has the ability to deal with a large number of
features. The machine is presented with a set of training
examples, (xi, yi) where the xi is the real world data
instances and the yi are the labels indicating which class the
instance belongs to. For the two class pattern recognition
problem, yi = +1 or yi = -1. A training example (xi, yi) is
called positive if yi = +1 and negative otherwise. SVMs
construct a hyperplane that separates two classes and tries
to achieve maximum separation between the classes.
Separating the classes with a large margin minimizes a
bound on the expected generalization error.
The simplest model of SVM called Maximal Margin
classifier, constructs a linear separator (an optimal
hyperplane) given by w T x - y= between two classes of
examples. The free parameters are a vector of weights w
which is orthogonal to the hyper plane and a threshold
value. These parameters are obtained by solving the
following optimization problem using Lagrangian duality.
Ill II 2
Minimize = —\\w\\
2 II H
Subject to D ii(w T x i -y)>l 9 i = 1, , /.
(8)
where D^ corresponds to class labels +1 and -1. The
instances with non-null weights are called support vectors.
In the presence of outliers and wrongly classified training
examples it may be useful to allow some training errors in
order to avoid over fitting. A vector of slack variables £i
that measure the amount of violation of the constraints is
introduced and the optimization problem referred to as soft
margin is given below. In this formulation the contribution
to the objective function of margin maximization and
training errors can be balanced through the use of
regularization parameter C.
The following decision rule is used to correctly predict
the class of new instance with a minimum error.
f(x)= sgnfwVy]
The advantage of the dual formulation is that it permits
an efficient learning of non-linear SVM separators, by
introducing kernel functions. Technically, a kernel function
calculates a dot product between two vectors that have
been (non- linearly) mapped into a high dimensional
feature space. Since there is no need to perform this
mapping explicitly, the training is still feasible although the
dimension of the real feature space can be very high or
even infinite. The parameters are obtained by solving the
following non linear SVM formulation (in Matrix form),
Minimize L D( u)=l/2u T Qu - e T u (9)
d T u=0 0<u<Ce
where and K - the Kernel Matrix. Q = DKD.
The Kernel Function K (AAT) (polynomial or
Gaussian) is used to construct hyperplane in the feature
space, which separates two classes linearly, by performing
computations in the input space.
f(x)= sgn(K(x,Xi T )*u-y)
Where u - the Lagrangian multipliers. In general the
larger the margin the lower the generalization error of the
classifier.
5. EXPERIMENTAL SETUP
The data investigation and epilepsy prediction is carried
out using SVMlight 1 for machine learning. Five categories
of feature vectors are labeled as 1 for Absence, 2 for
General, 3 for Complex Partial Epilepsy, 4 for Normal
Epilepsy and 5 for Simple Partial Epilepsy, The training
dataset used for epilepsy prediction modeling consists of
about 324 images, where each category consists of about
65.
The dataset has been trained using SVM with linear,
polynomial and RBF kernel and with different parameter
settings for d, gamma and C-regularization parameter. The
parameters d and gamma are related with polynomial
kernel and RBF kernel respectively.
The 10 fold cross validation method is used for
evaluating the performance of the SVM based trained
models. The performance of the models is evaluated based
on prediction accuracy of the models and learning time.
6. RESULTS AND DISCUSSION
The cross validation outcome of the trained models
based on support vector machine with linear kernel is
shown Table I.
Table I.
SVM Linear kernel
Linear SVM
C=0.1
C=0.2
C=0.3
C=0.4
Accuracy (%)
70
72
76
78
Time(secs)
0.01
0.02
0.02
0.03
1 SVMlight is an open source tool.
http ://www. cs. Cornell. edu/people/tj/svm_light/
119
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
The outcome of the model based on SVM with
polynomial kernel and with parameters d and C are shown
in Table II.
Table II. SVM Polynomial kernel
d
C=0.1
C=0.2
C=0.3
C=0.4
1
2
1
2
1
2
1
2
Accuracy (%)
70
80
82
80
80
81
74
75
Time(secs)
0.2
0.1
0.2
0.6
0.3
0.1
0.3
0.4
The predictive accuracy of the non-linear support
vector machine with the parameter gamma (g) of RBF
kernel and the regularization parameter C is shown in
Table III.
Table III. SVM RBF kernel
g
C=0.1
C=0.2
C=0.3
C=0.4
1
2
1
2
1
2
1
2
Accuracy (%)
80
83
83
81
83
86
85
77
Time(secs)
0.2
0.3
0.4
0.4
0.5
1.5
1.6
1.2
The average and comparative performance of the SVM
based prediction model in terms of predictive accuracy and
learning time is given in Table IV and shown in Figure 1
and Figure 2.
Table IV. Overall performance of three models
Kernel type
Accuracy
Learning time
Linear
84.96%
0.027 sees
Polynomial
90.12%
0.362 sees
RBF
93.87%
0.787 sees
Prediction Accuracy
86
I Accuracy(%)
Linear Polynomial RBF
Figure 2. Prediction Accuracy
Learning Time
0.9
0.8
0.7
0.6
0.5
0.4
0.3
0.2
0.1
./
./
S-
I Learning
Time(secs)
Linear Polynomial RBF
Figure 3 : Learning Time
As far as the epilepsy predictions task is anxious,
accuracy plays major role in determining the performance
of the epilepsy trained model than considering the learning
time. From the above results, it is found that the predictive
accuracy shown by SVM with RBF kernel with parameters
C=0.2 and g=2 is higher than the SVM with linear and
polynomial kernel.
7. CONCLUSION
This paper elucidates the modeling of the epileptic
seizure prediction task as multi-class classification problem
and the implementation of supervised learning algorithm,
support vector machine. The performance of SVM based
epilepsy prediction models is evaluated using 10 fold cross
validation and the results are analyzed. The results indicate
that the support vector machine with RBF kernel provide
the high prediction accuracy compared to other kernels.
SVM is better than conventional methods and show good
performance in all experiments it is very flexible and more
powerful because of its robustness. It is hoped that more
interesting results will follow on further exploration of
data.
8. ACKNOWLEDGMENT
The author would like to thank the Management and
Hospital, Coimbatore for providing the EEG data.
9. REFERENCES
[1] Forrest Sheng Bao , Jue-Ming Gao, Jing Hu , Donald Y. C.
Lie , Yuanlin Zhang , and K. J. Oommen. "Automated
Epilepsy Diagnosis Using Interictal Scalp EEG". 31st
Annual International Conference of the IEEE EMBS
Minneapolis, Minnesota, USA, September 2-6, 2009.
[2] Piotr Mirowski MSc*, Deepak Madhavan Yann Le
Cun, uben Kuzniecky " Classification of Patterns of
EEG Synchronization for Seizure Prediction".
[3] A. R.Sulaiman, " Joint Time - Frequency Analysis and
Its pplication for Non - Stationary Signals", Ph.D. Thesis
Elect. Eng. Dept, University of Mosul, 2001.
[4] Webster, J. G., "Medical Instrumentation Application and
esign", 2 nd ed., New York: Wiley, 1995.
120
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
[5] Nello Cristianini and John Shawe - Taylor. "An
Introduction to Support Vector Machines and other
kernel - based earning methods" Cambridge University
Press, 2000.
[6] K. Crammer and Y. Singer. "On the Algorithmic
implementation of Multi - class SVMs, JMLR,
2001. Vojislav Kecman: "Learning and Soft Computing —
Support Vector Machines, Neural Networks, Fuzzy Logic
Systems", The MIT Press, Cambridge, MA, 2001.
[7] Chui, C.K. (1992a), "Wavelets: a tutorial in theory and
applications", Academic Press.
[8] Ian H. Witten, Eibe Frank, Len Trigg, Mark Hall, Geoffrey
Holmes, Sally.
[9] Ian H. Witten, Eibe Frank. : Data Mining - Practical
Machine Learning Tools and Techniques. 2nd edn. Elsevier.
(2005).
[10] Joachims T, SchoTkopf B, Burges C, Smola A,"Making
large-Scale SVM Learning Practical. Advances in Kernel
Methods - Support Vector Learning", 1999, MIT Press,
Cambridge, MA, USA.
[11] John Shawe-Taylor, Nello Cristianini, "Support Vector
Machines and other kernel-based learning methods", 2000,
Cambridge University Press, UK.
[12] Soman K.P, Loganathan R, Ajay V, "Machine Learning with
SVM and other Kernel Methods", 2009, PHI, India.
[13] Crammer Koby, Yoram Singer,"On the Algorithmic
Implementation of Multi-class Kernel-based Vector
Machines", Journal of Machine Learning Research, MIT
Press, Cambridge, MA, USA, 2001, Vol.2 Page 265-292.
121 http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
Deployment of Intelligent Agents in
Cognitive Networks
Huda Fatima
Dept. ofCS
Jazan University
Jazan, K.S.A
Dr.Sateesh Kumar Pradhan
Dept. of Comp.Engineering
King Khalid University
Abha, K.S.A
Mohiuddin Ali Khan Dr. G.N.Dash
Dept. of Comp. Networks Dept. of Comp. Science
Jazan University Sambalpur University
Jazan, K.S.A Orissa, India
Abstract — Every organization faces challenging
task in the designing of the communication
network in order to make its efficiency smoother
by the increasing complexities. Therefore, we
have to proposed a concept of cognitive network
and how the intelligent agents are deployed to
overcome the challenges.With the tremendous
expansion of networks across the globe, the
deployment of intelligent agents in cognitive
networks contributes as an efficient, reliable and
challenging task for the researchers. In this
paper, we survey the existing research work on
cognitive networks and later we provide the
artificial intelligent techniques that are
potentially suitable for the development of
cognitive networks.
Keywords: Artificial Intelligence, Cognitive
network, Intelligent agents.
I.
INTRODUCTION:
One of the fastest growing areas is the information
and communication technologies. These changes
have an immediate impact on diverse aspects of the
modern society, which includes inter-human
relations, economy, education & entertainment. In
this respect, the development of reliable, flexible and
future-proof infrastructure should be capable of
increasing the users' quality of life by providing
services such as e-health, e-learning and e-payments.
In order to meet the demand of the increased
complexity, future networks should be easily
maintainable and their capabilities should be
continuously improved and upgraded by relying as
little as possible on human intervention. Therefore
the network research community proposed a new
concept of networking: The Cognitive Network.
What is a Cognitive Network and how are the
intelligent agents deployed is what we have
presented here.
Cognitive networks
In this section, we analyze several existing
definitions for cognitive networks, and we argue that
two elements are essential for developing a cognitive
network (CN): the knowledge representation and the
cognition loop. Next, we discuss the framework
proposed in [2] for introducing cognition to
communication networks. The main part of the
section focuses on methods from AI that seem
applicable for developing CNs. We provide a
summary of several types of intelligent agents (IAs),
map them to the functional states of the cognitive
loop. As we go along, we also refer to existing
research on CNs which makes use of the respective
type of IA, where available. How it started? The
word cognitive refers to an entity that is able to
perform some kind of conscious intellectual activity
such as thinking, reasoning, learning or remembering
in order to make sense of its surroundings. This word
was first used in communication networks to refer to
a technology by Mitola as he introduced the
cognitive radio [4].
122
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
We would like to emphasize that, according to the
dictionary [9], the word cognitive used as an
adjective to a noun means: of, relating to, being, or
involving conscious intellectual activity (as thinking,
reasoning, or remembering); based on or capable of
being reduced to empirical factual knowledge.
In [2], the authors define the CN as a network with a
cognitive process that can perceive current network
conditions, plan, decide, act on those conditions,
learn from the consequences of its actions, all while
following end-to-end goals. This loop, the cognition
loop, senses the environment, plans actions
according to input from sensors and network
policies, decides which scenario fits best its end-to-
end purpose using a reasoning engine, and finally
acts on the chosen scenario as discussed in the
previous section. The system learns from the past
(situations, plans, decisions, actions) and uses this
knowledge to improve the decisions in the future.
This definition of CN does not explicitly mention the
knowledge of the network; it only describes the
cognitive loop and adds end-to-end goals that would
distinguish it from CR or so called cognitive layers
[2]. We consider this definition of CN incomplete
since it lacks knowledge which is an important
component of a cognitive system as discussed so far
in this paper and also in [2,4,6,8].
The cognitive process can operate in a centralized
way, spanning over a large network, or in a totally
distributed manner at a device level. In the first case,
it might be too expensive to centralize all the
network specific information that the cognition loop
requires, while in the second case there might be too
little knowledge available to pursue end-to-end
network goals. In reality, the deployment of the
cognitive functionality in a network will depend on
the network specific problems and will be an
engineering decision. However, it is important that
the cognitive framework is designed in such way as
to be modular, easily upgradeable and scalable in
order to be able to accommodate existing as well as
next generation technologies and applications.
The capabilities of a Cognitive Network can be
highly distributed or extremely centralized. In
general, a Cognitive network is formed of a set of
distributed cognitive entities (agents) which are
somehow "smart" as they have certain reasoning
capabilities to be connected to the network. The
entities in such a network interact with each other,
they can cooperate, act selfishly or a combination of
the two. While functioning in this environment, the
entities are able to learn and take decisions in such
way as to reach an end-to-end goal. These end-to-
end goals are dictated by the business and user
requirements [2,4]. Developing and maintaining such
a network is an extremely challenging task and has
enormous potential, especially in the area of network
management.
A Cognitive Network needs to evolve overtime: its
set of technologies has to be updated by removing
deprecated and adding new ones; its set of tools that
help managing complexity should be added and
removed in a plug and play fashion. Thus, the
architecture of cognitive network should be flexible
and should lead to a modular and highly scalable
infrastructure. Furthermore, the cognitive network
must be self aware : it should be able to determine
appropriate actions to achieve goals and to learn
while doing all these. It should be self-configuring,
self-optimizing, self-healing and self-protecting in a
cognitive way.
In this paper, we analyze some recent trends in the
development of communication networks and
investigate in more detail the concept of cognitive
network. Cognitive networks are promising to be the
major step towards efficient and automatic
management of increasing complexity of
communication networks.
Cyclic Process in Cognitive Network.
All systems that are able to adjust their functioning
according to changes in their environment are based
on feedback information. Cognitive networks are no
exception in this respect, so they will also use a
control loop, also called cognition cycle [7, p. 7],
feedback loop [2], context based adaptation loop [8].
According to Thomas et al. [2], the loop employed
by a cognitive network should be based on the
concept of the Observe-Orient-Decide-Act loop
originally used in the military, augmented by
learning and following end-to-end goals to achieve
cognition. In [8], the loop also has a communicating
capability for communicating with other loops in a
distributed environment.
123
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
The cognition cycle as described by Mitola [7, p. 8]
features the following states: observe, orient, plan,
decide, act and learn. It uses the orient module for
classifying stimuli and does not explicitly encompass
policies.
Cycle management:
In [10], the authors investigate a cognitive agent for
wireless network selection which is designed to hide
the complexity of the wireless environment from the
user. The selection problem is decomposed
into four elements that enhance the agent to select
the network which is most suitable to user
preferences. First, user's feedback that the decision
making process will be used is captured. Second, the
available services are evaluated against learned user
preferences. Third, the agent decides when to change
services and which new service to select based on
user's preferences, context and goals.
Fourth, the value of previously unseen services is
predicted. Using this approach, the agent
continuously monitors the wireless environment and
selects the best service according to the current
model of user preferences. However, when the user
is unsatisfied (or changes preferences), The model is
updated and a new selection is made to satisfy
preferences.A Cognitive Resource Manager (CRM)
and its conceptual architecture are introduced in
[14]. The CRM's functioning is based on a cognition
cycle adapted from Mitola [7] and aims at enabling
autonomic optimization of the communication stack
as a whole, thus acting as an intelligent vertical
calibration(Fig.l). The intelligence would be based
upon methods from the field of AI.
rr[~
I!
U
»2
Fig 1. Open Systems Interconnection (OSI)
model
Loop for security
The CycSecure application [12] makes use of an
incomplete cognitive loop. It uses daemons installed
on machines in the network that collect local
information and send it to the server when polled. A
human operator can examine and modify the
network model, query and view network statistics.
The system is able to generate possible attack plans
based on the information gathered from the system
and the internal knowledge base. Based on these
attack plans, the human operator can decide for
remedy measures to increase the security of the
system.
Communication Requirements and research
directions
In the history of telecommunications, development
has always been driven by humans need to
communicate, i.e. reliably transmit ever increasing
amount of information across increasing distances.
However, communication networks became
increasingly complex and more difficult to manage,
requiring increasingly specialized tools and human
operators for their maintenance, configuration and
optimization.From the user's point of view the
necessities in the world of telecommunications, as it
is today, are : higher bandwidth or alternative
solutions capable of accommodating the traffic .
These necessities derive from the user's thirst for
digital content.
From the network operators' point of view, some of
the main necessities are: complexity, management,
security, scalability, fault tolerance, fast integration
of new technologies and a good business model [6].
The network operator has to create adequate
premises for delivering the digital content.
These user's and network operators necessities are
actually forming the basis for research activities
currently underway in the area of cognitive
networks. In general, research directions in
communications can be classified in 8 broad
categories: theory, signal processing, networks,
software, user satisfaction, security, management
and next generation protocols and architectures. In
an attempt to obtain an objective big picture of the
trends in research areas as well as quantitative
estimation of the ongoing work, we used ontogeny, a
semi-automatic ontology editor [6] to analyze the
124
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
conference proceedings of IEEE Globecom 2006 and
2007, totaling 201 1 papers
Artificial intelligence:
Artificial intelligence is concerned with intelligent
behavior in artifacts. Intelligent behavior, in turn,
involves perception, reasoning, learning,
communicating, and acting in complex environments
.Artificial Intelligence has as one of its long-term
goals the development of machines that can do these
things as well as humans can, or possibly even
better. Another goal of AI is to understand behavior
whether it occurs in machines or in humans or other
animals
Intelligent agents developed in a couple of streams
of work, among them is cybernetics [Wiener 198],
cognitive psychology, Computational linguistics
[Chomsky 1914], and adoptive control theory
[Widrow & Hoff 1960], also contributed to the
intellectual matrix developed by Artificial
intelligence.
Intelligent Agents:
Intelligent agents in Artificial intelligence react,
plan, reason and learn in an environment more or
less compatible with its abilities and goals. Here we
shall see how the actions of other agents can be
anticipated in each agents own planning, and indeed,
how an agent can even affect the actions of other
agents in the service of its own goals. To predict
what another agent will do , we need methods for
one agent to model another ; to affect what another
agent will do. There are two kinds of models used by
agents, iconic and feature based. An iconic model of
the environment attempts to simulate relevant
aspects of the environment; a feature-based model
attempts to describe the environment-perhaps by
formulas in the predicate calculus. The agents that
we deploy can use either an iconic or a feature-based
model of the other agents cognitive structure. And
the other agent itself might be presumed to be using
either an iconic or feature-based model. The four
possibilities are shown in table 1 along with the
modeling strategy each one provokes.
Figure 2. Intelligent agents for Cognitive Networks
The starting point towards developing a CN is the
intelligent agent (IA). This section presents existing
and emerging AI techniques that can prove useful for
developing agents for CNs. According to Russell and
Norvig [13, p. 42], an agent is central to AI. It is an
entity that perceives the environment through
sensors and acts upon that environment through
actuators. This is the so called "weak" definition of
agency while "stronger" definitions take into
account functions and characteristics of the agent
[14, p. 8,13, p. 42]. Among different classifications
of agents, we will consider as a reference, the one
established at IBM, which uses three dimensions to
describe agents (see Fig.6). The first dimension is
the Agency, which determines the degree of
"autonomy and authority vested in the agent". The
second dimension is the Intelligence, which
describes the degree of reasoning and learned
behavior. Finally, the third dimension is Mobility,
which specifies the degree to which agents travel
through the network [14, p. 9]. Current networks
operate via message passing (i.e. IP packets between
two routers or primitives between TCP and IP)
where the receiver takes an action as a consequence
of the received message. This type of operation is
asynchronous and is characteristic to expert systems
[14, p. 9,15]. This approach permitted loose coupling
of complex systems (e.g. communication networks).
However, this approach permits the lowest degree of
autonomy according to Fig. 1. On the Intelligence
axis, some of the current communication systems do
not even reach the lowest level as they do not even
allow specification of preferences (e.g. QoS
specifications). In this respect, CNs are expected to
enhance the level of intelligenceof current
communication systemsby incorporating so called
125
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
Intelligent Agents (IAs) in the KP. On the Agency
axis, IAs can perform actions on behalf of the user,
more specifically they can interact with data,
applications or services. On the Intelligence axis, IAs
can hold a model(i.e. user, system, environment,
etc.), perform reasoning, planning and learning.
These actions are exactly the same as the ones
desired from CN and can be found in the states of the
cognition loop (see Plan, Decide, Act, Learn and
Policy Fig.4).
Mg. i The ccrap rt ion loop.
Networks of the future will make use of agents to
improve their performance with respect to all three
axes in Fig 1 .
*+
Eervic* inlermctivity
Application interactmty
Dala iMGraclMty
Representation of us&f
Asyncshrcny
■■£
^
^
ssss
Intelligence
In the case of CNs, the main improvement is
achieved with respect to the Intelligence axis.
Therefore, in the remainder of the section we focus
on describing utility of Intelligent Agents for these
networks. We also emphasize the correspondence
between Intelligent Agents and the states of the
cognition loop. From the intelligence point of view,
the minimal requirement for an Intelligent Agents in
general is to hold a model and be able to reason
based on this model. These IAs (Intelligent Agents
)are also called knowledge-based agents. Reasoning
can take place upon two types of knowledge: certain
(true, false and unknown) and uncertain. Reasoning
under certain knowledge is accomplished by logical
agents. In this respect, agents "can form
representations of the world, use a process of
[logical] inference to derive new representations
about the world, and use these new representations to
deduce what to do" [13, p. 191]. Logical agents use
symbolic knowledge representations, so called
artificial languages, and typically first-order logic to
infer new facts. These representations also support
semantic querying. Agents that have incomplete or
uncertain information use decision theory and are
also called decision theoretic agents. These agents
use knowledge representations specific for uncertain
domains (i.e. full joint distributions can constitute
the knowledge base) to reason. Then they perform
probabilistic inference, which is the computation of
posterior probabilities from the observed evidence.
Conclusions
The recently emerging CN concept is promising to
be the right answer to emerging challenges of the
network management. In this paper we surveyed
existing work on CNs. We first analyzed recent
research trends in communications. We mapped
existing AI techniques to the states of the cognition
loop and identified challenges for research in AI
from which CNs could benefit. We concluded the
paper with identification of standardization activities
related to or potentially benefiting from the research
in the area of CNs.
The discussions in this paper indicate that the way
forward in developing CNs is to bring together the
experts from the areas of communication networks
and AI. Communication networks are faced with
great complexity challenges and several AI
techniques proved to handle complexity well.
Furthermore, AI is searching for areas of
applications, and communication networks are
underexploited in this respect. However, due to the
vastness in Artificial Intelligence field, we hope to
upgrade more in terms of Cognitive Networks and
other methods & tools of AI.
126
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
References :
[I] Nils J. Nilsson, Artificial Intelligence A New Synthesis.
[2] R.W. Thomas, L.A. DaSilva, A.B. MacKenzie, Cognitive networks, in: Proceedings of the First IEEE
International Symposium on New Frontiers in Dynamic Spectrum Access Networks, Baltimore, MD,
USA, November 8-11, 2006.
[3] J. Mitola, Cognitive Radio - An Integrated Agent Architecture for Software Defined Radio, Ph.D.
Dissertation, Royal Institute of Technology, Kista, Sweden, May 8, 2000
[4] Q. Mahmoud, Cognitive Networks - Towards Self-Aware Networks, John Wiley and Sons, 2007, ISBN
9780700999.
[5] D.D. Clark, C. Partrige, J.C. Ramming, J.T. Wroclawski, A knowledge plane for the internet, in:
Proceedings of the SIGCOMM 2004, Karlsruhe, Germany, August 26-29, 2004.
[6] R.W. Thomas, Cognitive Networks, Ph.D. Dissertation, Virginia Polytechnic and State University,
Blacksburg, VA, June 16, 2007.
[7] J. Mitola, Cognitive Radio - An Integrated Agent Architecture for Software Defined Radio, Ph.D.
Dissertation, Royal Institute of Technology, Kista, Sweden, May 8, 2000.
[8] P. Balamuralidhar, R. Prasad, A context driven architecture for cognitive nodes, Wireless Personal
Communications 16 (2008) 124- 110.
[9] FCC, ET Docket No. 04-422, Notice of Proposed Rule Making and Order, December 2004.
<http://www.scribd.com/doc/112914/ Federal-Corn munications-Commission-FCC04422Al>.
[10] QWL-QoS Ontology. <http://www4.ntu.edu.sg/home6/PG0487868/ OWLQoSOntology.html>
(visited on August 2008).
[II] P. Mahonen, M. Petrova, J. Riihijarvi, M. Wellens, Cognitive wireless networks: your network just
became a teenager, in: Proceedings of the INFOCOM 2006, Barcelona, Spain, April 24-29, 2006.
[12] B. Shepard, C. Matuszek, C.B. Fraser, W. Wechtenhiser, D. Crabbe, Z. Gundordu, J. Jantos, T.
Hughes, L. Lefkowitz, M. Witbrock, D. Lenat, E. Larson, A knowledge-based approach to network
security: applying Cyc in the domain of network risk assessment, in: Proceedings of the Innovative
Applications of Artificial Intelligence Conference, Pittsburgh, PA, USA, July 9-14, 2006.
[13] S. Russell, P. Norvig, Artificial Intelligence: A Modern Approach, second ed., Prentice Hall, 2002,
ISBN 0147901262.
[14] J. Bradshaw, Software Agents, AAAI Press/The MIT Press, 1997, ISBN 0262622412.
[15] P. Jackson, Introduction to Expert Systems, Addison-Wesley International Computer Science Series,
1986, ISBN 0201142246.
AUTHORS PROFILE
I am currently employed in Jazan University, Jazan, K.S.A Department of Computer Networks. My area of
Research is Artificial Intelligence, Data Mining, Network Security. I have published few papers in International
Journals. I wish to do research more into these fields.
1 27 http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol 8, No. 6, September 2010
A Performance Study on AES Algorithms
B.D.C.N.Prasad 1
Dept. of Computer Applications,
P V P Siddardha Institute of Technology
Vijayawada, India
P E S N Krishna Prasad 2
Dept. of Computer Science & Engineering
Aditya Engineering College
Kakinada, India
P Sita Rama Murty
Dept. of Computer Science & Engineering
Sri Sai Aditya Institute of Science & Technology
Kakinada, India
K Madhavi 4
Dept. of CSE
Dadi Institute of Technology
Anakapalli, India
Abstract — The Aim of this project is to find the performance
comparative analysis of AES algorithms such as MARS, RC6,
Rijndael, Serpent, Twofish algorithms in terms of speed,
memory, time, encryption and decryption, key setup time,
number of rounds, key sizes and also hardware considerations.
Most of the AES algorithms, especially symmetric block
ciphers, are based on the principle of substitution and
transposition to encrypt a plain-text message and to produce a
cipher-message. Those transformations are based on well-
understood Mathematical problems using non-linear functions
and linear modular algebra.
Implementation of cryptographic algorithms mainly uses bit-
level operations and table look-ups. Bit-wise operators (XORs,
AND/OR, etc.), substitutions, logical shifts and permutations
are quite common operations. Such operations are well suited
for their fast execution in hardware platforms. Furthermore,
currently abundant memory resources in hardware platforms
enhance encryption speed for the operations like substitution.
These operators play an important role in analysis and
comparison of the performance of the above mentioned AES
algorithms, to evaluate simple, effective and efficient outcomes
and also the information might be more secure.
Keywords-AES algorithms; Mars; RC6; Rijndeal; Sarpent;
Two fish;
I.
Introduction
Security is a broad topic and covers a multitude of sins,
in its simplest form. It is concerned with making sure that
nosy people cannot read, or worse yet, modify message
intended for other recipients. It is concerned with people
trying to access remote services that they are not authorized
to use. Security also deals with people trying to deny that
they sent certain message.
Network security problems can be divided roughly into
four intertwined areas:
• Confidentiality,
• Authentication and Integrity control
• Denial of service
Cryptography, over the ages, has been practiced by many
who have devised ad-hoc techniques to meet some of the
information security requirements. The last twenty years
have been period of transition as the discipline to a broader
area. There are now several international scientific
conferences denoted exclusively to cryptography and also
and International Association for Crypto-logic Research
(IACR), aimed at fostering research in the area.
There are two general types of cryptographic algorithms.
1. Symmetric algorithms.
2. Asymmetric algorithms.
The current Digital Encryption Standard (DES) does no
longer satisfy the need for data security because of its short
56-bit key. Such short keys can today be broken by brute
force attacks. We are looking for newer and more flexible
algorithms to replace DES. Some of the candidates for the
Advanced Encryption Standard (AES) are MARS encryption
algorithm, RC6, Serpent, Rijndael, and Twofish. These are
symmetric key block ciphers use 128 bit blocks and supports
variable key sizes (from 128 to 1248 bits). These use
addition and subtractions, S -boxes, fixed and data dependent
rotations, and multiplications.
The final AES selection was made on the basis of several
additional characteristics:
• computational efficiency and memory
requirements on a variety of software and
hardware, including smart cards
• flexibility, simplicity
implementation
and
ease
of
The existing system consisted of files with literally no
file security standards like AES algorithms such as MARS,
RC6, Rijndael, Serpent, and Twofish. AES algorithms are
symmetric cipher algorithms which are far better than DES
algorithms, since DES algorithms are limited key size with
fixed number of blocks. So, we have chosen for finding the
comparison of AES algorithms to provide the security for
Data as well as networks and files. AES algorithms are to be
128
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
implemented due to the following factors against which
several security measures had to be taken up:
1. Reading data
2. Manipulating and modifying data
3. Illegal use of files
4. Corrosion of data files
5. Distortion of data transmission
The main issue of (1) is secrecy and confidentiality.
Confidentiality has always played an important role in
diplomatic and military matters. Often Information must
store or transferred from one place to another without being
exposed to an opponent or enemy. Key management is also
related to Confidentiality. This deals with generating,
distributing and storing keys. Items (2-4) are primarily
concerned with reliability. Often the expression integrity is
used as a measure of genuineness of data. Also computer
files and networks must be protected against intruders and
Unauthorized. Item 5 is different aspect of the security of the
information.
Vol 8, No. 6, September 2010
CBC mode within ESP. This mode requires an Initialization
Vector (IV) that is the same size as the block size. Use of a
randomly generated IV prevents generation of identical
cipher text from packets which have identical data that spans
the first block of the cipher algorithm's block size.
The IV is XOR'd with the first plaintext block before it is
encrypted. Then for successive blocks, the previous cipher
text block is XOR'd with the current plaintext, before it is
encrypted. For the use of CBC mode in ESP with 64-bit
ciphers.
2) Key Size
Some cipher algorithms allow for variable sized keys,
while others only allow specific, pre-defined key sizes. The
length of the key typically correlates with the strength of the
algorithm; thus larger keys are usually harder to break than
shorter ones. This article stipulates that all key sizes MUST
be a multiple of 8 bits.
The default key size that implementations MUST support
128 bits. In addition, all of the ciphers accept key sizes of
192 and 256 bits.
A. AES Algorithms
AES algorithms are symmetric cipher algorithms
with variable key sizes and blocks, also with number of
rounds to encrypt and decrypt the data than DES algorithms.
There are numerous algorithms in AES. From them we have
chosen the following algorithms for finding the performance
analysis on time, memory, key sizes, key setup time,
encryption, and decryption and so on.
The Chosen algorithms are as:
• MARS encryption algorithm
• RC6 Algorithm
• Rijndael Algorithm
• Serpent Algorithm
• Twofish Algorithm
TABLE II.
Key sizes
TABLE I.
General Structure
Cipher
Type
Rounds
Using
MARS
Extended
Feistel
32
Variable Rotation,
Multiplication
Non Cryptic Rounds
RC6
Feistel
20
Variable Rotation,
Multiplication
Rijndael
Square
10,12,14
Serpent
SP Network
32
Bitslice
Twofish
Feistel
16
I) Mode
No operational modes are currently defined for the
AES cipher. The Cipher Block Chaining (CBC) mode is
well-defined and well-understood for symmetric ciphers, and
is currently required for all other ESP ciphers. This article
specifies the use of the AES cipher and the other finalists in
Algorithm
Key Sizes(bits)
Default
MARS
128 - 448*
128
RC6
Variable up to 2040
128
Rijndael
128,192,256
128
Serpent
Variable up to 256**
128
Two fish
Variable up to 256***
128
MARS key lengths must be multiples of 32 bits.
** Serpent keys are always padded to 256 bits. The
padding consists of a "1" bit followed by "0" bits.
*** Twofish keys, other than the default sizes, are always
padded with "0" bits up to the next default size.
3) Weak Keys
Some cipher algorithms have weak keys or keys that
MUST not be used due to their interaction with some aspect
of the cipher's definition. If weak keys are discovered for the
AES or any of the other finalists, then weak keys SHOULD
be checked for and discarded when using manual key
management. When using dynamic key management, weak
key checks SHOULD NOT be performed as they are seen as
an unnecessary added code complexity that could weaken the
intended security.
4) Block Size and Padding
All of the algorithms described in this document use a
block size of sixteen octets (128 bits), mandatory for the
AES. Some of the algorithms can handle larger block sizes
as well. Padding is required by the algorithms to maintain a
16-octet (128-bit) blocksize. Padding MUST be added, such
that the data to be encrypted has a length that is a multiple of
16 octets. Because of the algorithm specific padding
requirement, no additional padding is required to ensure that
the cipher text terminates on a 4-octet boundary (i.e.
maintaining a 16-octet blocksize guarantees that the ESP Pad
129
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
Length and Next Header fields will be right aligned within a
4-octet word
5) Rounds
This variable determines how many times a block is
encrypted. While this variable MAY be negotiated, a default
value MUST always exist when it is not negotiated.
Algorithm
Negotiable
Default of Rounds
MARS
yes
32
RC6
yes
20
Rijndael
yes
10,12,14
Serpent
yes
32
Twofish
yes
16
(IJCSIS) International Journal of Computer Science and Information Security,
Vol 8, No. 6, September 2010
1) RC6
RC6 is the submission of MIT (Massachusetts Institute of
Technology) and the RSA-Laboratories. Similar to MARS it
splits the 128 bit blocks into four words in the algorithm, but
the algorithm is designed in a way that you can easily change
to two 64 bit words in newer architectures. RC6 is also a
Feistel network. It uses the same type of operations except
from look-up tables and fixed rotations. The algorithm is
more flexible than MARS about blocksize and number of
rounds. The AES submission is optimized for 128 bit blocks
and 20 rounds. Several performance test showed that RC6 is
slower than MARS for encryption and for the key setup. But
it uses less memory because there are no look-up tables.
B. MARS Algorithm
MARS is a shared-key block cipher that works with
a block size of 128 bit and a variable key size. The algorithm
is a type-3 Feistel network which is word (32 bit) oriented.
The word orientation should bring a performance for
software implementations on most computer architectures
available today. A fully optimized implementation is
expected to run at lOOMbit/second and hardware can achieve
an additional lOx speedup factor.
1) Operations
MARS algorithm uses a big variety of different
operations:
Additions, subtractions and xors: These simple
operations are used to mix data and key values together.
Because xors are interleaved with additions and subtractions
these operations do not commute with each other.
Table look-up: Similar to the S-boxes in DES has also
MARS cipher a table look-up. It uses a single table of 512
32-bit words, simple called S-box. A problem of the table
look-up is the slow software implementation (at least 3
instructions per look-up). That's why S-box look-up is only
used sparely in MARS where fast avalanche of the key bits is
needed.
Fixed rotations: Data-dependent rotations: Data
dependent rotations may lead to differential weaknesses.
This problem is solved in MARS by combining these
rotations with multiplications.
Multiplications: All multiplications in MARS are modulo
232 which suits most modern computer architectures.
Multiplications used to be a problem in cryptographic
algorithms because they were slow. Today is this no longer
the case. Most architecture can complete a multiplication in 2
clock cycles. MARS algorithms uses 16 multiplications per
block. This should not be a big deal. For hardware
realizations we have the problem that a multiplicator needs
much more chip-space than adders or logical units.
C. Comparison with other AES Candidates
There are 4 other candidates for AES in the last
round. So they are all 128 bit block ciphers with variable key
length from 128 bit to at least 192 bit. All designs claim to be
secure against all known attacks like differential, linear,
known plaintext or cipher text and other attacks.
2) Rijndael
Rijndael is the submission of the Belgium Proton World
Int. and the Katholieke Universities Leuven, Belgium. This
algorithm is quite different from MARS. It works with
Galois Field GF(128) arithmetic and handles the input blocks
as matrices of bytes. They define several operations to these
matrices as Byte Sub, ShiftRow, MixColumn and
AddRoundKey. For detailed information about these
operations consult [Rijndael99]. Several combinations of
these operations define a round. Depending on the key length
which is in the range from 128 to 256 bits a fixed number of
rounds has to be executed. This cipher is not a Feistel
network. Several performance tests showed that Rijndael is
about the same speed in encryption and decryption as
MARS. But the key expansion for keys of the same length is
significant slower.
3) Serpent
Serpent is a submission from three universities
(Cambridge University, England; Technion, Haifa, Israel;
University of Bergen, Norway). Therefore it's the only
algorithm where no company stands behind. The algorithm is
pretty similar to DES, it uses permutations, xors, fixed
rotations and shifts and constant table look-up's. The first
version of the algorithm even used the same S-boxes as DES.
The key can vary from 128 to 256 bit. The algorithm works
internally also with 4 words as RC6 and MARS.
Performance tests that the encryption of Serpent is about
25% faster than the MARS encryption. But the key
expansion is significant slower. An implementation of
Serpent also uses a lot of memory because of the look-up
tables.
4) Twofish
Twofish is the submission from a company called
Counterpane. It is a 16 round Feistel cipher that works with
key dependent 8x8 bit look-up tables, 4 by 4 matrices over
the Galois field GF(128), a pseudo-Hadamard transform,
permutations and rotations. The detailed description of these
functions can be found in [Twofish]. The key length varies
also from 128 bit to 256 bit as in most other AES candidates.
Performance tests showed that the encryption speed of
Twofish is about the same as for MARS, but the Twofish key
setup is significant faster.
130
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
D. Performance Analysis
The performance analysis can be done with various
measures such as speed comparison with encryption and
decryption cycles, key setup and key initialization, analysis
of various key sizes and fair speed/security comparisons. The
performance analysis will be presented in the form of tables
and figures below.
1) Speed Comparisons
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
TABLE III.
Speed
Key Setup
Cipher
Encrypt
(Cycles)
Decrypt
(Cycles)
Encrypt
Decrypt
Init
MARS
1600
1580
4780
5548
18
RC6
1436
1406
5186
5148
30
Rijndael
1276
1276
17742
18886
28
Serpent
1800
2102
13154
12648
14
TwoFish
1254
1162
18846
18634
20
2000
1500
1000 f-
500
□ Encryption(Cycles)
1 ■ Decryption(Cycles)
MARS RC6 Rijndael Serpent TwoFish
Figure 1. Graph for Encryption and Decryption (Cycles)
20000
18000
16000
14000
12000
10000
6000
4000 -j-
2000 -
□ Encryption
■ Decryption
MARS RC6 Rijndael Serpent TwoFish
Figure 2. Graph for Key setup Encryption and Decryption
Initialization
30 -
25
20
15
10 -
5
c
D Initialization
MARS RC6
Rijndael Serpent
TwoFish
Figure 3. Key Initialization
2) Analysis on various Key Sizes
a) Encryption
TABLE IV. Encryption
Algorithm
Encryl28
Encryl92
Encry256
MARS
3738
3707
3733
RC6
4698
4740
4733
Rijndael
4855
4664
4481
Serpent
1843
1855
1861
Twofish
1749
1749
1744
b) Decryption
TABLE V.
Decryption
Algorithm
Encryl28
encryl92
encry256
MARS
3965
3965
3936
RC6
4733
4698
4740
Rijndael
4819
4624
4444
Serpent
1873
1897
1896
Twofish
1781
1775
1761
Encryption
5000
4000
s \
— ♦ — Algorithm
B 3000
in
— m — Encryl28
encryl92
\
* 2000
1000
encry256
12 3 4 5 6
Figure 4. Encryption
131
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
Decryption
r^ J ^ === \
\
tfj 3000
\
* 2000
^ n
-Encryl28
-encryl92
encry256
#
** 4? .*
4* *
x^°
Figure 5. Decryption
c) Fair Speed/ Security Comparisons
TABLE VI. Fair Speed/ Security Comparisons
Cipher
Original
(cycles)
Rounds
Minimal Rounds
Time
(Cycles)
MARS
1600
32
20
1000
RC6
1436
20
20
1436
Rijndael
1276
10
8
1021
Serpent
1800
32
17
956
Twofish
1254
16
12
940
MARS RC6 Rijndael Serpent Twofish
Figure 6. Fair speed / security comparisons
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
algorithms. But MARS is one among the chosen algorithms
is some what better as considered reports.
It won't be an easy decision to choose one of the
finalists as AES. There is no known weakness in all these
algorithms, so other factors as performance, needed
hardware or flexibility must be used for the decision. MARS
cipher is for sure a good candidate. It has the largest
available key length of all of them and it is expandable to
larger block sizes than 128 bit. Another advantage of MARS
is that it comes from a well known company that is in this
business for a long time which means they have a lot of
experience and have proven their trustworthiness.
F. References
[I] Cryptography and Network Security -"William Stallings" /Third
Edition.
[2] The Laws of Cryptography with JAVA Code -"Neal R.Wagner".
[3] MARS: C.Burwick, D. Coppersmith, E.DAvignon, R.Gennaro,
S.Halevi, C.Jutla, S.Matyas, L. O'Connor, M.Peyravian, D.Safford,
N.Zunic, "MARS - a candidate cipher for AES", IBM Corporation,
September 1999.
[4] TweakIBM99 - Shai Halevi, "Detailed discussion of the MARS
"tweak" for Round 2", IBM Corporation, Mai 1999.
[5] RC6: Ronald L. Rivest, M.J.B. Robshaw, R. Sidney, Y.L. Yin, "The
RC6 Block Cipher", M.I.T. Laboratory for Computer Science, RSA
Laboratories.
[6] Rijndael: Joan Daemen, Vincent Rijmen, "AES Proposal: Rijndael",
Proton World Int.l, Belgium, Katholieke Universiteit Leuven,
Belgium, September 1999.
[7] Serpent: Ross Anderson, Eli Biham, Lars Knudsen, "Serpent: A
Proposal for the Advanced Encryption Standard", Cambridge
University, England; Technion, Haifa, Israel; University of Bergen,
Norway.
[8] Twofish: Bruce Schneier, John Kelsey, Doug Whiting, David
Wagner, Chris Hall, Niels Ferguson, "Two sh: A 128-Bit Block
Cipher", Counterpane Systems, University of California Berkeley.
[9] E Biham, " A Note Comparing AES Candidates, NIST,1999.
[10] P. Preneel, V Rijmen and A Bosselaers, " Principles and Performance
of Cryptographic Algorithms", Dr. Dobb's journal.
[II] B. Schneier, J Kelsey, D. Whiting, D Wagner, C. Hall and N
Ferguson, "Performance Comparison of the AES Candidate
conference,1999.
E. Conclusion
A performance comparison can be made among various
AES Algorithms such as MARS, RC6, Rijndael, Serpent,
Twofish. The Performance analysis reports were presented
in the specified contents. It is concluded that all the above
specified algorithms have almost similar speed rate and
timings while using Java tool for execution of these
AUTHORS PROFILE
Dr. B D C N Prasad, currently is a Professor & Head of
Department of Computer Applications at Prasad V. Potluri
Siddardha Institute of Science and Technology,
Vijayawada, Andhra Pradesh, India. He received Ph.D. in
Applied Mathematics from Andhra University,
Visakhapatnam, India in 1984. His research interests
132
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
includes Machine Intelligence, Data Mining, Rough Sets and Information Security in
Computer Science and Boundary value problems and Fluid Dynamics in Mathematics.
He has several publications in mathematics and computer science in reputed national
and international journals. He is a member of ISTAM , ISTE and also he is a national
executive member of Indian Society for Rough Sets.
Mr. P E S N Krishna Prasad, currently is a Research
Scholor under the guidance of Dr. BDCN Prasad in the
area of Machine Intelligence and Neural Networks. He is
working as Associate Professor in the Department of CSE,
Aditya Engineering College, Kakinada, Andhra pradesh,
India. He is a member of ISTE. He has presented and
published papers in several national and International conferences and journals. His
areas of interest are Artificial Intelligence, Neural Networks and Machine Intelligence.
i
Mr. P Sita Rama Murty, currently is a Research Scholor, in
the area of ATM networks and Information Secuirty. He is
working as Assistant Professor in the department of CSE,
Sri Sai Aditya Institute of Science and Technology,
Kakinada, Andhra Pradesh, India
133
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
Hybrid Fingerprint Image compression and
Decompression Technique
*Dr.R.Seshadri, ,B.Tech„M.E,Ph.D
**Yaswanth Kumar.Avulapti ,M.C.A, M.Tech, (PhD)
***Dr.M.Usha Rani M.C.A,PhD
^Director, S.V.U. Computer Center S.V.University, Tirupati
**Research Scholar, Dept of Computer Science, S.V.University, Tirupati
***Associate Professor, Dept. of Computer Science,, SPMVV, Tirupati
Abstract
In this paper a biometric authentication
system based on Fingerprint. A fingerprint is
the representation of the epidermis of a finger.
It consists of a pattern of interleaved ridges
and valleys.
Like every thing in the body fingerprint
ridges form through a combination of genetic
and environmental factors. Finger prints are
fully formed at about seven months of fetus
development. Fingerprint ridges don't change
throughout the life of an individual except
incase of accidents such as cuts on the
fingertip (or) burns on the fingertip. In this
paper we proposed a hybrid model to
compress the fingerprints.
Keywords :Biometrics,Enrollment
Authentication, compression, Decompression.
Introduction
The term Biometrics is derived from
the Greek word bio (life) and metrics (to
measure). Basically it is a method of
identifying a person based on his/her
physiological or behavioral characteristics
such as Fingerprints, Iris, Face, Hand
geometry, Retinal scan,,DNA,Signature,Key
Stroke, Voice,Gait,Ear,Palm print, Dental
radiographs. Among all the biometric
techniques, fingerprint recognition is the most
popular method and is successfully used in
many applications
Biometrics is a rapidly evolving
wonderful technology which has been widely
used in forensics applications such as criminal
identification and prison security.
Biometrics(Fingerprints) can be used to
prevent unauthorized access to a computer.
The main objective of the fingerprint
image compression is to reduce the number of
bits as much as possible by keeping the
resolution and quality of the fingerprint while
decompressed as same as to the original
fingerprint image.
Fingerprint Identification System
A finger print system works in two
different modes they are Enrollment mode
and Authentication mode as shown in figure. 1
134
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
Enrollment mode in which fingerprint system
is used to identify and collect the related
information about the person and his/her
fingerprint image.
Authentication mode in which fingerprint
system is used to identify the person who is
declared to be him/her.
Enrollment Mode
Finger print
Acquisition
Feature
Extraction
Template
Authentication Mode
Finger Print
Acauisition
Feature
Extraction
Matching
Film
Digital
Spatial
resolution
4
linepairs/mm
1024 by
1024 pixels
Data
capacity per
image
1MB
Data rate
30 images
per seconds
30 MB/Sec
Media
one film
4-CDs
Matching Score
Fig. 1. Enrollment and Authentication of a
Fingerprint system
Need for fingerprint compression
Let us see this scenario the spatial
resolution of the cinefilm is generally
assumed to be equivalent to a digital matrix of
at least 1000 by 1000 pixels, each with up to
256 gray levels (8 bit or one byte) of contrast
information.
The Figure. 2 derives from this principal
parameter some examples for requirements on
digital image communication and archival in a
catheterization laboratory with low to mediums
volume.
Fig.2.Replacement of cine film by digital
Imaging with high resolution
In this scenario the enormous data rate
of 30 Megabyte per second has to be
supported. This is much faster than even
advanced ATM networks (offering less than
20MB/s orl60Mbit/s).
Looking for existing off-line media real-time
display from CD-R would require a CD-R
player with a data rate of 200X, while the
fastest players available presently deliver 5 OX
(IX stands for a data rate of 150 KB/sec). The
total amount of data required in this
scenario is even more frightening. To store
these images and make them available over
network (e.g. the internet), compression
techniques are needed.
Steps involved in Fingerprint compression
1) Specifying the rate (bits available) and
distortion parameters for the target
fingerprint image
2) Dividing the fingerprint data into various
classes based on their importance
3) Dividing the available bit budget among
these classes such that the distortion is
minimum.
4) Quantize each class separately using the
bit allocation
5) Encode each class separately and write to
the file
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
Steps involved in Fingerprint decompression Steps in Proposed Fingerprint compression
1) Read the fingerprint quantized data from
file using the decoder
2)Dequantize the fingerprint data(Recerse of
step 4)
3)Rebuild the Fingerprint Image(Reverse of
step 2)
Proposed Fingerprint Image Compression
System
We proposed a Fingerprint image
compression system in which the fingerprint
images are compressed by Lossless
compression technique called Run length
Encoding and Decompressed by Huffman
coding. It is also called as Hybrid technique.
A Fingerprint compression system
consists of two blocks namely Encoder and
Decoder as shown in the fig. 3
Fingerprint
Compression
Fingerprint
Decompression
Original
Finger print
Run Length
Encoding
Quantizer
Encoder
Reconstructed
Finger print
i
k,
Huffman
Encoding
A
k,
De-Quantizer
A
k,
Decoder
1) Specifying the bits available and distortion
parameters for the target fingerprint image
2) Dividing the fingerprint data into various
classes based on their importance
3) Dividing the available bit budget among
these classes such that the distortion is
minimum.
4) Quantize each class separately using the bit
allocation
5) Encode each class separately using an Run
Length encoder and write to the file
Steps involved in Fingerprint decompression
1) Read the fingerprint quantized data from
file using the Huffman decoder
2)Dequantize the fingerprint data(Reverse of
step 4)
3 Reconstructed the Fingerprint
image(Reverseof step 2)
In the Run length encoding the
Fingerprint images with repeating grey values
along rows or columns can be compressed by
storing "runs" of identical grey values as
shown in the fig.4 & fig 4.b
Finger print
Grey value 1
Repetition 1
Finger print
Grey value 2
Repetition 2
Fig.4.aRun length encoding the Fingerprint
Images
This a very simple fingerprint
Fig.3.Fingerprint image"co!hpression using compression method used for chronological
Run length Encoding & Decompressed 36 data.The run leifff;^^^fflg /slt f^ces the
using Huffman coding sequences of identical pixels called the runs
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
by tiny symbols. The run length code for
fingerprint image is represented by a sequence
of {A,B} where a is the intensity of the pixel
and B is the number of consecutive pixels
with the intensity A as shown in the figure .If
both A and B are represented by 8-bits this
span of 12 pixels is coded using eight bytes
yielding a compression ratio 1:5.
For Example
23,23,23,23,23,67,67,89,89,78,78
{23,5}{67,2}{89,2}{78,2}
Fig.4.bRun length encoding the Fingerprint
Images
Huffman encoding is performed that is
mapping of the code words to the
corresponding symbols will result in a
compressed data. The original image is
reconstructed i.e. decompression is done by
using Huffman decoding.
Generate a tree equivalent to the
encoding tree. Read input character wise and
left table until last element is reached in the
table. Output the character encodes in the leaf
and returnto the root, and continues until all
the codes of corresponding symbols.
Advantages of Fingerprint Compression:
a) It reduces the storage space at the time of
processing the fingerprints
b) It not only reduces the storage requirements
but also overall execution time
c)It also reduces the probability of
transmission errors since fewer bits are
transferred
d) It also provides a level of security against
criminal monitoring
137
Conclusion
This paper presents different steps
involved in the development of fingerprint
authentication system. The proposed
fingerprint image compression and
decompression technique uses both the Run
length encoding and Huffman coding. This
hybrid fingerprint compression and
decompression technique are good for certain
applications like the security technologies.
These two compression and decompression
techniques are lossless ones. Using this hybrid
technique which leads to less storage of
memory and reducing the execution time.
References
l.A. K. Jain,Patrick Flynn,Arun A.Ross .
"Handbook of Biometrics".
2. The Henry Clas sification System Copyright
© 2003 International Biometric Group
3. Compression Using Fractional Fourier
Transform A Thesis Submitted in the partial
fulfillment of requirement for the award of the
degree of M.E Electronics &Communication.
By Parvinder Kaur.
4. RL-Huffman Encoding for Test
Compression and Power Reduction in Scan
Applications-MEHRDAD NOURANI and
MOHAMMAD H. TEHRANIPOUR, The
University of Texas at Dallas
5.A.B.Watson,"Image Compression using the
DCT" ,Mathematica Journal, 1995,pp.81-88.
6. DAVID A. HUFFMAN, Sept. 1991, profile
Background story: Scientific American, pp.
54-58.
7. Efficient Huffman decoding by MANOJ
Aggarwal and Ajai Narayana A NEW LOSSLESS
METHOD OF IMAGE COMPRESSION AND
DECOMPRESSION USING HUFFMAN CODING
TECHNIQUES by JAGADISH H. PUJAR, 2LOHIT
M. KADLASKAR Faculty, Department of EEE, B V
B College of Engg. & Tech. India
2 Student, Department of EEE, B V B College of
Engg. & Tech. India
8. D. Monro and ^MiWe-.c§^«ilpq uenc y
balance in biorthdpNMASWelets. transactions
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
of the IEEE Int.Conf. on Image Processing,
1:624(627, 1997.
9. A. Said and W. Pearlman. A new fast and
e_cient image codec based on set partitioning
in hierarchical trees. IEEE Transactions on
Circuits and Systems for Video Technology,
6:243(250, June 1996.
10. W. Sweldens. The lifting scheme: A
custom-design construction of biorthogonal
wavelets.
Authors Profile
T^»* i**f
I
Dr.R.Seshadri was
born in Andhra
Pradesh, India, in 1959.
He received his B.Tech
degree from Nagarjuna
University in
1981. He received his
M.E degree in Control
System Engineering
from PSG College of Technology,
Coimbatore in 1984. He was awarded with
PhD from Sri Venkateswara University,
Tirupati in 1998. He is currently Director,
Computer Center, S.V. University, Tirupati,
India. He has Published number of papers in
national and international conferences,
seminars and journals. At present 12 members
are doing research work under his guidance in
different areas
papers in national and international
conferences, seminars. He attends Number of
work shops in different fields.
Dr. M. Usha Rani is an
Associate Professor in the
Department of Computer
Science and HOD for
CSE&IT, Sri Padmavathi
Mahila
Viswavidyalayam(SPMVV
Womens' University),
Tirupati. She did her Ph.D. in Computer
Science in the area of Artificial Intelligence
and Expert Systems. She is in teaching since
1992. She presented more than 34 papers at
National and Internal Conferences and
published 19 articles in national &
international journals. She also written 4
books like Data Mining - Applications:
Opportunities and Challenges, Superficial
Overview of Data Mining Tools, Data
Warehousing & Data Mining and Intelligent
Systems & Communications. She is guiding
M.Phil and Ph.D. in the areas like Artificial
Intelligence, DataWarehousing and Data
Mining, Computer Networks and Network
Security etc.
Mr.YaswanthKumar
.Avulapati received
his MCA degree with
First class from Sri
Venkateswara
University, Tirupati. He
received his M.Tech
Computer Science and
Engineering degree
with Distinction from
Acharya Nagarjuna University, Guntur.He is a
research scholar in S.V.University Tirupati,38
Andhra Pradesh. He has presented number of
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
Punctured Self -Concatenated Trellis Codes
with Iterative Decoding
Labib Francis Gergis
Misr Academy for Engineering and Technology
Mansoura City, Egypt
Abstract-A special concatenated code structure
called self-concatenated trellis code (SCTC) is
presented. This scheme based on only one
recursive convolutional code(# SC), followed by
a mapping modulator. The union bounds of
SCTC are derived for communications over
Additive White Gaussian Noise (AWGN) and
Rayleigh fading channels. Asymptotic results
for large interleavers are extended to M-ary
bandwidth efficient modulation schemes by
puncturing process. The combination of self-
concatenated codes with powerful bandwidth-
efficient component codes leads to a
straightforward encoder structure, and allows
iterative decoding. The scheme has been
investigated for 4-PSK, 8-PSK, 16-PSK, and
16-QAM modulation schemes with varying
overall bandwidth efficiencies. The choice
based on the rate of RSC and puncturer
encoder component.
key words ;S elf -Concatenated codes, trellis-coded
modulation, uniform interleaved coding,
convolutional coding, iterative decoding
1. INTRODUCTION
Trellis coded modulation (TCM) [1] was
originally proposed for transmission over
AWGN and fading channels due to its
attractive bandwidth efficiency.
Concatenated trellis-coded modulation is an
alternative to TCM. Different approaches to
concatenated trellis-coded modulations were
presented in [2], and [3]. The main principle in
the concatenated coding schemes is to use two
codes in series (or parallel) joined through one
or more interleavers. This means that the
information sequence is encoded twice, the
second time after a scrambling of the
information bits.
Concatenated trellis codes are classified as
serially concatenated convolutional codes
(SCCC), these codes were analyzed in [4].
Using the same ingredients, another type of
concatenated codes named parallel
concatenated convolutional codes (PCCC), was
described in [5]. A third choice is defined as a
hybrid concatenation of convolutional codes
(HCCC) was described in [4] and [6]. Self-
concatenated convolutional codes proposed in
[7], [8], and [9] constitute another attractive
iterative detection aided code-family for their
low complexity, since they invoke only a single
encoder and a single decoder.
Puncturing is the process of deleting some
parity bits from the codeword according to a
puncturer code rate. The redundant bits in
coding decrease the bandwidth efficiency.
Puncturing increases code rate without
increasing complexity and decreases free
distances of code. The advantage of punctured
codes for binary transmission is that the
encoders and decoders for the entire class of
codes constructed easily by modifying the
single encoder and decoder for the rate 1/2
binary convolutional code from which the high
rate punctured code was derived [10].
The construction of self-concatenated trellis
codes (SCTC) is described in section 2. Section
3, derives analytical upper bounds to the bit-
error probability of SCTC using the concept of
uniform interleavers. Factors that affect the
performance of SCTC are described in section
4. Finally results for some examples depicted in
section 4, have been stated in section 5.
139
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
2. SCTC MODEL
The basic concept of self-concatenated
scheme is shown in Figure. 1, the input bit
sequence {b^ of the self-concatenated encoder
is interleaved to yield the bit sequence {b 2 }.
After the parallel-to-serial (P/S) conversion, the
information sequence is defined as b (1) = {b 1}1
b 2 ,i b 1)2 b 2 ,2 .... }• The resultant bit sequences
are input to a recursive systematic
convolutional (RSC) encoder. At the output of
the encoder the interleaved bit sequence is
punctured. The encoder output is composed of
the combined systematic bit sequence and
parity bit sequence.
where E b /N is the bit energy to noise density
ratio, A c w>h for block code C represents the
number of codewords of the block code with
output weight h associated with an input
sequence of weight w, and N is the size of the
interleaver. The A c w/l is the input-output weight
coefficient (IOWC). The function Q (^2R h
E b /N ) represents the pairwise error
probability which is a monotonic decreasing
function of the signal to noise ratio and the
output weight ft.
For a fading channels, assuming coherent
detection, and perfect Channel State
Information (CSI), the conditional pairwise
error probability is given by
7T
b!
b!
b (l)
RSC
Encoder
R^l/2
Puncturer
R 2 =l/2
,(i)
Fig 1. The Self -Concatenated
Code Encoder
The overall code rate, R, can be derived
based on [9] as:
R = R t /2R 2 = (1/2) / 2 (1/2) = 1/2
(1)
It can be observed that different codes can
be designed by changing R 2 .
3. PERFORMANCE OF
SELF-CONCATENATED
TRELLIS CODES
Consider a linear block code C with code
rate R, and minimum distance ft m . An upper
bound on the conditional bit-error probability
of the block code C over AWGN channels,
assuming coherent detection, maximum
likelihood decoding, can be obtained in the
form [4]
N/R N
Pb(e/p) < Z Z MV) A\ M •
h=d min w=l
Rh(E b /N )
(2)
Q(\ hREb/N^pt 2 )
(3)
i=l
The fading samples p are independent
identically distributed (i.i.d.) random variables
with Rayleigh density of the form
f(p)= 2 pe
P 2
(4)
The structure of a SCTC, as shown in
Figure .1, is composed of q-1 interleavers each
of size N bits, and a single systematic recursive
trellis code C with rate (bq/bq+1), where only
the b + 1 outputs of the encoder are mapped to
2 b+1 modulation levels.
The average input-output weight
coefficients A c w/l for SCTC with q-1
interleavers can be obtained by averaging
equation (2) over all possible interleavers. A
uniform interleaver is defined as a probabilistic
device that maps a given input word of weight
w into all its distinct pV ~1 permutations with
equal probability 1/ f^V~] •
Thus, the expression for IOWC of SCTC is
derived as [7]
A c
w,h
(5)
JV
w
where A C WW) i|W|h is the number of code words of
the trellis encoder of weight ft, which is
determined in [5], and
140
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
N
w
w!
(6)
Substituting equation (6) in equation (2) yields
[3]
P b (e/p) = B m W +1 Q(< 2R h m E b /N )
(7)
where the constant B m is independent of N, and
is derived in [8], and h m is the minimum
Euclidean distance of the SCTC scheme .
4. SCTC: PERFORMANCE
FACTORS
i .e-t
1.e-5 -
1.e-6 -
■
1.e-7 -
T *
^
LU
♦.. ■
V
' ■
m
1.e-8 -
1.e-9 -
'■♦.
T
'T .
''♦..
■
'▲
>
^»
■
•
N = 10
'♦.
"^
■
1.e-10-
■
▲
T
N = 50
N = 100
N = 200
' ' ♦ .
'♦..
A
>
1.e-10-
♦
N = 300
'''♦
I
I
I
I
It is shown from equation (7), that there
are many factors that affect the performance of
SCTC. The most influential parameter is the
interleaver size N. The bit error probabilities
for self-concatenated trellis code with overall
rate R-l/2, is shown in Fig. 2, with various
interleaver lengths N= 10, 50, 100, 200, and 300
are plotted versus the signal-to-noise ratio
E b /N . The systematic and parity bits, b and b ly
are mapped to 4-ary Phase Shift Keying
(QPSK) modulation. The figure shows the
beneficial gain that can be achieved through
increasing N.
HD}^[D}^]Hp-
TCM constituent encoder
J_
Mapper
dB
Fig. 3. Upper Bounds to the Bit Error
Probability for SCTC with QPSK
using different Interleaver Lengths
Applying the upper bound of equation (7),
we obtain the results reported in Fig. 3.1t is
also clear from equation (7) that, the minimum
Euclidean distance of the SCTC code (h m ) is an
another main parameter affecting the
performance of SCTC. Different values of h m
could be obtained by a variety of modulation
schemes. Puncturing is used in order to
increase the achievable bandwidth efficiency.
Different codes could be designed by changing
the rates Ri and R 2 . The output of the encoder
is then mapped to the Gray-code mapping
function. The various coding schemes
considered in this paper are characterized in
Table 1, that defines both R h R 2 , the overall
code rate R, and the associated mapped
modulation scheme to R. The BER versus
E b /N performance curves of the various
QPSK, 8-PSK, 16-PSK, and 16-QAM are
shown in Fig.4.
Fig. 2. Self-Concatenated Trellis Encoder with
rate R = 1/2
141
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
Ri
R 2
R
Modulation
Scheme
1/2
1/2
1/2
QPSK
1/3
1/4
2/3
8-PSK
1/3
2/3
1/4
16-PSK
16-QAM
Table 1.
Various Modulation Schemes Obtained
from Varying fij and R 2 .
1.e-6
1.e-7
a: 1 .e-8
LU
1.e-9 -
1.e-10- v
2
E h /N n
dB
Fig. 4. Upper Bounds to the Bit Error
Probability for SCTC versus
Different Modulation Schemes
The choice of decoding algorithm and
number of decoder iterations also influences
performance.
A functional diagram of the iterative
decoding algorithm for SCTC is presented in
Fig. 5.
Fig 5. Self -Concatenated Trellis Decoder
The decoder is a self-concatenated scheme
using a soft-input soft-output (SISO) maximum
aposteriori probability (MAP) algorithm [9]. It
first calculates the extrinsic logliklihood Ratio
(LLR) of the information bits, namely L e (bi)
and L e (b 2 ). Then they are appropriately
interleaved to yield the a priori LLRs of the
information bits, namely L> a (bi) and L a (b 2 )> as
shown, in Fig. 5.Self-concatenated decoding
proceeds, until a fixed number of iterations is
reached.
The performance of SCTC with QPSK
modulation schemes considered are shown in
Fig .6. The SCTC has an overall rate R = 111,
the interleaver length N of this code = 100 bits.
The performance after various numbers of
iteration is shown. It is clear that performance
improves as the number of decoder iterations
increases.
5. CONCLUSIONS
In this paper, a channel coding scheme
(SCTC) that is bandwidth efficient and allows
iterative decoding of codes built around
punctured codes together with higher order
signaling.
142
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
1.e-6 -
QPSK Scheme
1.e-7 -
V ^
D
N = 100
1.e-8 -
A
%
1.e-9 -
*
□
1.e-10-
V
A
□
LU
00
1.e-11-
1.e-12-
1.e-13-
1.e-14-
1.e-15-
1.e-16-
•
□
A
No Iteration
2 Iteration
3 Iteration
4 Iteration
V
V
□
□
V
V
I
I
I
I
C
)
1
2
3
4 e
E h /N n
dB
Fig. 6. Upper Bounds to the Bit Error
Probability for SCTC versus
Different Decoding Iterations
The SCTC schemes consists of binary RSC
codes and different puncturing rates. The
puncturer is used to increase the achievable
bandwidth efficiency. A search for good rates
was performed, taking into account the
puncturing at the transmitter. It is also
demonstrated the significant in the
performance and the decrease of the bit error
rate and probability of errors to SCTC within
increasing: the interleaver size N, and the
number of decoder iterations
REFERENCES
[1] G. Ungerboeck, "Channel coding
with multilevel phase signaling,"
IEEE Trans. Inf. Th., Vol. 25, pp. 55-
67, Jan. 1982
[2] F. Brannstrom, A. Amat, and L.
Rasmussen, " A General Structure for
Rate-Compatible Concatenated
Codes," IEEE Trans on Communication
Letters, Vol. 11, pp. 437-439, May 2007.
[3] A. Amat, G. Montorsi, and S.
Benedetto, " New High-rate
Convolutional Codes for Concatenated
Schemes", Proceeding of IEEE
International Conference on
Communication, ICC 2002,Vol. 3, pp.
1661-1666, April 2002.
[4] D. Divsalar, and F. Pollara, " Serial and
Hybrid Concatenated Codes with
Applications, " International Symposium
on Turbo Codes and Related Topics,
Brest, France 1997
[5] S. Benedetto, and G. Montorsi,"
Unveiling Turbo Codes: Some Results
On Parallel Concatenated Coding
Schemes, "IEEE Transactions on
Information Theory, Vol. 42, No. 2,
March 1996
[6] A. Amat, and E. Rosnes," Good
Concatenated Code Ensembles for the
Binary Erasure Channel", IEEE
Journal on Selected Areas in
Communications, Vol. 27, No. 6, pp.
928-943, August 2009.
[7] S. Ng, M. Butt, and L. Hanzo, " On the
Union Bounds of Self-Concatenated
Convolutional Codes", IEEE Signal
Processing Letters, Vol. 16, No. 16, No.
9, September 2009.
[8] S. Benedetto, D. Divsalar, G. Montorsi,
and F. Pollara, " Self-Concatenated
Codes with Self-Iterative Decoding for
Power and Bandwidth Efficiency"
International Symposium on
Information Theory, ISIT 1998,
Cambridge, MA, USA, August 1998.
[9] M. Butt, R. Riaz, S. Ng, and
L. Hanzo, Distributed Self-
Concatenated Codes for Low-
Complexity Power-Efficient
Cooperative Communication", IEEE
VTC 2009, Anchorage, Alasks, USA,
2009.
[10] R. Deshmukh, and S. Ladhake, "
Analysis of Various Puncturing
Patterns and Code Rates: Turbo
Code", International Journal of
Electronic Engineering Research
Volume 1, Number 2, pp.79-88, India,
2009.
143
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
AUTHOR PROFILE
Labib F.Gergis received the Bsc, Msc, and
Ph.D from faculty of engineering, Mansoura
University, Egypt, in 1980, 1990, and 2000,
respectively. He is presently in Misr Academy
for Engineering and Technology, Egypt. His
areas of interest include digital
communications, Coding, and Multiple Access.
144 http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8„ No. 6, 2010
Application of Fuzzy Composition Relation For DNA
Sequence Classification
Amrita Priyam
Dept. of Computer Science and Engineering
Birla Institute of Technology
Ranchi, India.
Abstract — Abstract — This paper presents a probabilistic
approach for DNA sequence analysis. A DNA sequence
consists of an arrangement of the four nucleotides A, C, T
and G. There are various representation schemes for a
DNA sequence. This paper uses a representation scheme in
which the probability of a symbol depends only on the
occurrence of the previous symbol. This type of model is
defined by two parameters, a set of states Q, which emit
symbols and a set of transitions between the states. Each
transition has an associated transition probability, CZ Z/ - ,
which represents the conditional probability of going to
state j in the next step, given that the current state is i.
Further, the paper combines the different types of
classification classes using a Fuzzy composition relation.
Finally a log-odd ratio is used for deciding to which class
the given sequence belongs to.
Keywords-component; Transition Probability, Fuzzy Composition
Relation, Log-Odd ratio
I.
Introduction
A DNA sequence is a succession of the letters A, C, T and G.
The sequences are any combination of these letters. A physical
or mathematical model of a system produces a sequence of
symbols according to a certain probability associated with
them. This is known as a stochastic process, that is, it is a
mathematical model for a biological system which is governed
by a set of probability measure. The occurrence of the letters
can lead us to the further study of genetic disorder. There are
various representation schemes for a DNA sequence. This
paper uses a representation scheme in which the probability of
a symbol depends only on the occurrence of the previous
symbol and not on any other symbol before that. This type of
model is defined by two parameters, a set of states Q, which
emit symbols and a set of transitions between the states. Each
transition has on associated transition probability, CL^ , which
represents the conditional probability of going to state jfrom
state i in the next step, given that the current state is z. Each
B. M. Karan + , G. Sahoo + +
+ Dept. of Electrical and Electronics Engineering
++ Dept. of Information Technology
Birla Institute of Technology
Ranchi, India
class has a set of transition probabilities associated with it.
This transition probability is the measure of going from one
state to another. Now, each class has a set of transition
probability associated with it. We further group the similar
classes and their respective transition probability is merged
using a fuzzy composition relation. Finally a log-odd ratio is
used for deciding to which class the given sequence belongs
to.
II. DNA SEQUENCES
DNA sequence is a succession of letters representing the
primary structure of a real or hypothetical DNA molecule or
strand, with the capacity to carry information as described by
the central dogma of molecular biology. There are 4
nucleotide bases (A - Adenine, C - Cytosine, G - Guanine, T
- Thymine). DNA sequencing is the process of determining
the exact order of the bases A, T, C and G in a piece of DNA
[3]. In essence, the DNA is used as a template to generate a set
of fragments that differ in length from each other by a single
base. The fragments are then separated by size, and the bases
at the end are identified, recreating the original sequence of
the DNA[8][9]. The most commonly used method of
sequencing DNA the dideoxy or chain termination method
was developed by Fred Sanger in 1977 (for which he won his
second Nobel Prize). The key to the method is the use of
modified bases called dideoxy bases; when a piece of DNA is
being replicated and a dideoxy base is incorporated into the
new chain, it stops the replication reaction.
Most DNA sequencing is carried out using the chain
termination method [4]. This involves the synthesis of new
DNA strands on a single standard template and the random
incorporation of chain-terminating nucleotide analogues. The
chain termination method produces a set of DNA molecules
differing in length by one nucleotide. The last base in each
molecule can be identified by way of a unique label.
Separation of these DNA molecules according to size places
them in correct order to read off the sequence.
145
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8„ No. 6,2010
III. A Probabilistic approach for sequence
REPRESENTATION
A DNA sequence is essentially represented as a string of four
characters A, C, T, G and looks something like
ACCTGACCTTACG. These strings can also be represented in
terms of some probability measures and using these measures
it can be depicted graphically as well. This graphical
representation matches the Markov Hidden Model. A physical
or mathematical model of a system produces a sequence of
symbols according to a certain probability associated with
them. This is known as a stochastic process [2]. There are
different ways to use probabilities for depicting the DNA
sequences. The diagrammatical representation can be shown
as follows:
O OG
Q.02
FIG 1: [The states of A, C, G and T.]
For example, the transition probability from state G to state T
is 0.08, i,e,
P( x . = T | x M = G) = 0.08
In a given sequence x of length L, x h x 2 , x L , represent the
nucleotides. The sequence starts at the first state x h and makes
successive transitions to x 2 , x 3 and so on, till x L . Using Markov
property [6], the probability of x L , depends on the value of
only the previous state, x L . ls not on the entire previous
sequence. This characteristic is known as Markov property [5]
and can be written as:
P(x) = P(x L I x^Pix^ I x L _ 2 ) P(x 2 1 x l )P(x l )
= P(x 1 )fl miVi)
(1)
In Equation (1) we need to specify PfxJ, the probability of the
starting state. For simplicity, we would like to model this as a
transition too. This can be done by adding a begin state,
denoted by 0, so that the starting state becomes x =0.
p M=FK,
(2)
If there are n classes, then we calculate the probability of a
sequence x being in all the classes. To overcome this
drawback we use Fuzzy composition relation. That is, we
divide the n classes into different groups based on their
similarities. So, if out of n classes, m are similar then they are
treated as one group and their individual transition probability
tables are merged using the fuzzy composition relation. The
remaining (n - m) classes are similarly grouped. Lets say, if
there are two classes Rl and R2, the Fuzzy composition
relation between Rl and R2 [6] [7] can be written as follows:
R 1 °R 2 = Max{Min{R 1 {x\R 2 {y)))
(3)
O O A A ,_
O A
O O
Different class representation
Grouping of similar classes
Fig 2: Grouping of similar classes
A table is then constructed representing the entire (n - m)
similar classes. From this table we compute the probability
that a sequence x belongs to a given group using the following
equation:
log
1*,-
P(x|+)
P(x I -) — a^x,
= 2>g—
(4)
Here "+" represents transition probability of the sequence
belonging to one of the classes using fuzzy composition
relation and "-" represents the transition probability of the
same for another class [1].
If this ratio is greater than zero then we can say that the
sequence x is from the first class else from the other one.
An Example:
Let us consider an example for applying this classification
method. We have taken into consideration the Swine flu
data. [11] The different categories of the Swine flu data are
shown as R h R 2 and R 3 .
R 1} R 2 and R 3 shows the Transition Probability of Type 1, Type
2 and Type 3 varieties of Avian Flu.
Now considering O x . x . , the transition probability we can
rewrite (1) as
146
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8„ No. 6,2010
A
c
T
G
A
"0.13
0.06
0.09
0.08
Ri
c
T
0.09
0.06
0.04
0.05
0.05
0.05
0.02
0.07
G
0.08
0.04
0.05
0.06
A
c
T
G
A
"0.13
0.06
0.09
0.08"
R 2
C
~ T
0.08
0.06
0.04
0.05
0.04
0.06
0.02
0.07
G
0.08
0.04
0.04
0.06
A
0.09
0.06
0.06
0.06
T
0.07
0.04
0.05
0.05
0.06
0.07
0.07
0.07
G
0.07
0.09
0.04
0.05
A
c
T
G
A
"0.07
0.05
0.07
0.04'
x 3 = c
T
0.05
0.03
0.06
0.06
0.06
0.04
0.08
0.09
G
0.06
0.09
0.06
0.07
A
c
T
G
A
"0.08
0.05
0.08
0.08
*,
c
~ T
0.08
0.05
0.03
0.06
0.06
0.06
0.02
0.09
G
0.08
0.06
0.04
0.07
Using the Fuzzy composition relation technique on R x and R 2
and then the result of the application with relation R 3 we get
the final table for the Swine Flu class as:
Applying Fuzzy composition relation to these tables we get
the final table as
A
c
T
G
A
"0.07
0.06
0.05
0.06
c
0.08
0.09
0.07
0.06
T
0.04
0.03
0.07
0.06
G
0.05
0.08
0.04
0.02
Suppose we are given the sequence x which is to be classified
as either falling into any of the given classes and say x =
CGCG
A
c
T
G
A
"0.08
0.06
0.08
0.09
c
0.08
0.06
0.08
0.09
T
0.07
0.06
0.07
0.07
G
0.08
0.06
0.08
0.08
Similarly, we can repeat the same procedure for another class
Staphylococcus. X lf X 2 and X 3 shows the Transition
Probability of Type 1, Type 2 and Type 3 varieties of
Staphylococcus.
A
c
T
G
A
"0.08
0.05
0.07
0.06
*1
c
T
0.06
0.04
0.05
0.06
0.06
0.08
0.08
0.07
G
0.08
0.09
0.03
0.05
From the final fuzzy composition table the log odds ratio of
this sequence is:
, 0.09 , 0.06 , 0.09 „__„
log + log + log = 0.032270
0.08
0.07
0.08
Now, since this ratio is greater than 0, we can conclude that
the input sequence x belongs to the class Avian Flu. If further
classification on the data is needed we will then consult the
individual transition probabilities for all the three types.
CONCLUSION
In this paper we have used a probabilistic function for the
Markov Property. We have applied this for probabilistic
determination in the case of Avian flu virus and
Staphylococcus. The paper also presented a way for
identifying particular classes of genes or proteins. A given
input sequence can belong to either of the given classes. By
using a transition probability measure, one had to determine a
value for each class even though they were similar. The paper
presented a scheme such that the similar classes were merged
by using the fuzzy composition relation and now instead of
calculating each individual probability measure, one measure
147
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8„ No. 6,2010
is sufficient to depict all the similar classes. This measure is
further used in the log odds ratio to finally predict the class of
the input sequence.
References
[I] Anna Loanova, "Introduction to (log) odds ratio statistics and
Methodology", University of Groningen, 2008.
[2] C. E. Shanon, "A Mathematical Theory of Communciation",
The Bell System Technical Journal, Vol. 27, pp. 379 - 423, 623
-656.
[3] D. W. Mount, " Bioinformatics, Sequence and Genome
Analysis", 2 nd edition, CSHL Press, (2004). (3)
[4] Durbin, Eddy, Krogh, Mitchison, "Biological Sequence
Analysis", Cambridge University Press, 1998.
[5] L. R Rabiner, "A tutorial on Hidden Markov Models and
selected Application in speech recognition" -Proceeding of the
IEEE, Vol.77, No.2 Feb.1989.
[6] Lee, Kwang Hyung, "First course on Fuzzy Theory and
Applications", Advances in Soft Computing, Vol. 27, Springer,
2005.
[7] Michael Hanss , "Applied Fuzzy Arithmetic : An Introduction
with Engineering Applications", Springer, 2005.
[8] T. Dewey and M. Herzel, "Application of Information Theory to
Biology", Pacific Symposium on Biocomputing, 5:597 - 598
(2000).
[9] W. J. Ewens, G. R. Grant, "Statistical Methods in
Bioinformatics: An Introduction", Vol. 13, 2 nd edition, Springer.
[10] Y. Ephraim, L. R. Rabiner, "On the Relations Between
Modeling Approaches for Speech Recognition", IEEE
transactions on Information Theory, vol. 36, no. 2, March 1990.
[II] A.Priyam, B.M.Karan, G.Sahoo, "A Probabilistic Model For
Sequence Analysis", Inter national Journal of Computer Science
and Information security vol7 No. 1, (2010) 244-247.
148 http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
Data Security in Mobile Ad Hoc Networks using
Genetic Based Biometrics
B. Shanthini, Research Scholar
CSE Department
Anna University
Chennai, India
S. Swamynathan, Assistant FPofessor
CSE Department
Anna University
Chennai, India
Abstract — A mobile ad hoc network (1 AINKI) -is^^a^setf
configuring, dynamic, multi hop radio network without any fixed
infrastructure. 1 AlNKIs -fffe^coHections of wireless mobile
devices with restricted broadcast range and resources and
communication is achieved by relaying data along appropriate
routes that are dynamically discovered and maintained through
collaboration between the nodes. The main challenge in the
design of such networks is how to prevent the attacks against
data such as unauthorized data modification, impersonation etc.
En&metrics provides possible solutions for this security problem
in 1 AINKI MHceJthas the direct connection with user identity
and needs little user interruption. So, researchers have been
investigating ways to use biometric features of the user rather
than memorable password or passphrase, in an attempt to
produce tough and repeatable cryptographic keys. In this paper
such a security system based on EE&metrics and dhetic
A. Security challenges in MANET
Wireless ad hoc networks are vulnerable to various attacks
[1]. Adversaries may attempt passive and active attacks to gam
unauthori^d access to classified information, modify the
information, delete the information or disrupt the information
flow. The best way to protect data information in a most fme-
g^anular way is by providing__security at the application layer. It
is highly desirable to handle data confidentiality and integrity
in application layer, since this is the easiest way to protect data
from altering^ fabrication and compromise. With the rapid
evolution of wireless technology the reliance of ad hoc
networks to carry mission critical information is rapidly
growing^ This is especially important in a military scenario
where strategic and tactical information is sent. Therefore the
ability to achieve a highly secure authentication is becoming
algorithm which is providing data security in 1 AINL11 tiP^-E 016 cn tical.
presented.
Keywords— Mobile Ad hoc Networks, Data Security,
Biometrics, Genetic Algorithm
I. INTRODUCTION
Mobile ad hoc networks are seen as autonomous that can be
quickly formed, on demand, for specific tasks and mission
support. Communication generally happens through wireless
links, in which nodes within a radio range, communicate and
coordinate to create a virtual and temporary communication
infrastructure for data routin^and data transmission. MANET
can operate in isolation or in coordination with a wired networks
through a gateway node participatin^in both networks. This
flexibility along^ with their self-organi^ng^ capabilities, are
some of their biggest strengths, as well as their biggest security
weaknesses.
The applications of MANET include the foremost
situations such as emergency/crisis management, military,
healthcare, disaster relief operations and intelligent
transportation systems. So message security plays a vital role in
data transmission in MANET. However, because of the
absence of an established infrastructure or centralis
administration, implementation of hard-cryptographic
algorithms is a challenging prospect. So, in this paper, we
present a novel security method using_genetic based biometric
cryptography for message security in mobile ad hoc networks.
Numerous countermeasures such as strong_,authentication,
encrypting and decrypting^ the messages using_, traditional
cryptographic algorithms and redundant transmission can be
used to tackle these attacks. Even though these traditional
approaches play an important role in achieving_eonfidentiality,
integrity, authentication and non-repudiation, these are not
sufficient for more sensitive and mission-critical applications
and they can address only a subset of the threats. Moreover,
MANETs [2] cannot support complex computations or higl}
communication over head due to the limited memory and
limited computation power of mobile nodes.
B. Necessity of Biometrics Security
For mission-critical applications such as a military
application may have higher requirements regarding data or
information security. In such a scenario, we may design the
security system combining_,both biometrics and cryptography.
Biometric based security scheme overcome the limitations of
traditional security solutions. Biometrics refers to the methods
for uniquely recogni^ng^ humans based upon one or more
intrinsic physical or behavioral traits like fingerprints, iris,
retina scans, hand, face, ear geometry, hand vein, nail bed,
DNA, palm print, signature, voice, keystroke or mouse
dynamics, and gajt analysis etc.
149
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
Biometric technologies have confirmed its importance in
the fields such as security, access control and monitoring
applications. The tradeoffs among., these biometric
technologies really depend on the application and security
level involved. The best biometric technology [3] that can
easily be deployable in ad hoc networks is fingerprint
recognition. Fingerprints have been successfully used in
civilian identification for years because of their
unchanggability during_,the human life time and uniqueness of
each individual. As biometrics can't be borrowed, stolen, or
forgotten, and forging^ is practically impossible, it has been
presented as a natural identity tool that offers greater security
and convenience than traditional methods of personal
recognition.
Even though biometric has advantages, it also raises many
security and privacy concerns as giyen below:
i. Biometric is authentic but not secret,
ii. Biometric cannot be revoked or cancelled,
iii. If a biometric is lost once, it is compromised forever,
iv. Cross-matching^can be used to track^individuals without
their consent.
To overcome these disadvantages, instead of usmg^the
original biometric, a set of features are taken from it and
transformed using_, genetic alggrithm. If a biometric is
compromised, it can be simply reenrolled usmg_janother feature
set and another genetic operation, thus providing^revocability
and the privacy of the biometric is preserved.
C. Genetic Algorithms
Genetic algorithms [4] are a family of computational
models inspired by natural evolution. They belong_,to the field
of evolutionary computation and are based on three main
operators: Selection selects the fittest individuals, called
parents that contribute to the reproduction of the population at
the next generation, Crossover combines two parents to form
children for the next generation and Mutation applies random
changes to individual parents to form children. Two-point
crossover operator is used here which has the ability to generate,
promote, and juxtapose building blocks to form the optimal
strings.
This paper is organi^d into 4 sections. Section 1 introduces
the background and initiatives of the research. It also discusses
the challenges of message security, the necessity of biometric
security in MANET and Genetic alggrithms. Section 2 explains
the related research works that has been done to provide
security in MANET. Section 3 proposes a new security scheme
for MANET which combines genetic algorithm and biometrics.
Section 4 contains conclusion and suggestions for future
research.
II. RELATED WOIL
A few research works that has been done for data security
in MANET, the various approaches of biometric security and
Genetic alggrithms in security are briefly presented.
Qingljan Xiao [5] introduced a new strategy for
authentication of mobile users. Each user has a profile which
contains all the information of the ID holders. The group leader
also maintains the biometric templates of the group members.
Instead of a central authentication server, the group leaders act
as distributed authenticators. Each group has a shared
cryptographic key which is used for cryptographic
communication within the group. The proposed approach is
designed for hig^ security small group coalition operations and
may not be suitable for enterprise usage.
Jq Liu et al. [6] proposed an optimal biometric-based
continuous authentication scheme in MANET which
distinguished two classes of authentications: user-to-device and
de vice-to -networks This model focused on the user-to-device
class and it can optimally control whether or not to perform
authentication as well as which biometrics to use to minimis
the usage °f system resources.
B Ananda>Lrishna et al. [7] depicted a model which used
multiple alggrithms for encryption and decryption. Each time a
data packet is sent to the application layer it is encrypted using.,
one of these randomly selected alggrithms. When responses are
analy^d they giye a random pattern and difficult to know
neither algorithms nor keys. The proposed scheme worked
well for heavily loaded networks with hig^ mobility.
2Lr^ L et al. [8] explained the context of the study of
Genetic Algorithms as an aiding tool for generating and
optimi2ng^ecurity protocols. This paper explains how security
protocols can be represented as binary strings,, how GA tools
are used to define genome interpretation in optimi^tion
problems.
B. Shanthini et al. [9] explained Cancelable Biometric-
Based Security System (CBBSS), where cancelable biometrics
is used for data security in mobile ad hoc networks. Fingerprint
feature of the receiver is coupled with the tokened random
data by usin^ inner-product algorithm and this product is
discreti^d based on a threshold to produce a set of private
binary code which is actinias a crypto graphic key in this
system.
A. Jigadeesan et al. [10], proposed an efficient approach
based on multimodal biometrics (Iris and fingerprint) for
generating^ secure cryptographic key, where the security is
further enhanced with the difficulty of factoring large
numbers. At first, the features, minutiae points and texture
properties are extracted from the fingerprint and iris images
respectively. Then, the extracted features are fused at the
feature level to obtain the multi-biometric template. Finally, a
multi-biometric template is used for generating a 256-bit
cryptographic key.
III. FTtOroSED WOIL
In this proposed Genetic-Based Biometric Security System
(GBBSS), a genetic two-point crossover operator is applied on
biometric feature set and is used for data security in mobile ad
hoc networks. The main objective of the proposed security
scheme is to improvise the existing^data security approaches
for MANET to suit technology enhancements and to study the
networkzperformance.
150
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
A. Generation of Genetic-Based Biometric Key
In this model all the grpup members maintain the biometric
templates of the other group members. Suppose a member
wants to send a message, to any other member, the receiver's
fingerprint is divided into slices and feature set taken from the
slices is undergone a genetic two-point crossover operation and
the result is the cryptographic key in this system. Generation of
cryptographic key is shown in figure 1.
Hhgerprint
^r
1
1
Parent
^^^■^^^^^^^^^^H
Crossover
Points
■
1
Children ^^^^^^
I
Cryptographic^ ey
Figure 1 . Generation of cryptographic key from the finger print features.
The same key is generated by the receiver by using^his
biometric and the same sort of cross over operations and is
used for decryption.
Example:
Rrent
Children
01011100
00110011
01011100
00110011
1010000011111010
1111000011110000
After Crossover
1111000011110000
1010000011111010
00110101
11001100
00110101
11001100
If this biometric based key is compromised a new one can
be issued by using_a different set of features and different cross
over operation and the compromised one is rendered
completely useless. It can also be an application specific that is
different sets of fingerprint features can be used with different
cross over operations to generate respective crypto graphic key
for different applications.
B. Securing the Data
Data is secured by applying^ this cryptographic key to
encrypt the actual message, using_, a simple cryptographic
algprithm say Fiestel algorithm. The encryption and decryption
processes are specified by the formulae:
C = % R (F) andF^D* R (C)
where F*- Hain Text
C - Cipher Text
^LR^Ley created by Receiver's Biometric
E - Encryption Algprithm
D - Decryption Algprithm
In Fiestal alggrithm, a blockzof si^ N is divided into two
halves, of length N/2, the left half called XL and rig^t half
called XR. The output of the ith round is determined from the
output of the (i-l)th round. The same key is used for all
iterations without generating^ sub keys. Also the number of
iterations performed is reduced to show that security can be
achieved by using_, simple algprithm. For example if the
plaintext is of 512 bytes, then encryption is performed for
every 64 bits and the process is repeated until all 512 bytes are
encrypted. Fiestel structure is given in figure 2. [1].
Pkirdfiid(2wrbits)
(M3-
Key Gene rat ion
Algorithm
Key
<EM3-
Key
G-&-
Key
Ciprerbeirt [A<v b its)
Figyre 2. Fiestel Algorithm
Algorithm for Encryption:
1 . Divide the plaintext into two blocks of si^, 32 bytes,
XL and XR
2. ForI=lto32
Do XL = XL XOR^ey
XR = F (XL) XOR XR
Swap XL, XR
jbin XL, XR
3 . Repeat step 2 until the entire plaintext is encrypted
Algprithm for Decryption:
Do the reverse operation of Encryption process.
151
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
C. Implementation ofGBBSS in MANET
The proposed scheme can be implemented over any unicast
routing_protocols like DSR or AODV which discover routes as
and when necessary and the routes are maintained just as long^
as necessary. A typical MANET is shown in figure 3.
Suppose User A wants to send the message, to User C, after
the forward and reverse paths are set up by the route discovery
method, the data will be sent through that path to the
destination C. Before sendmg_,the data through that path, the
data will be encrypted by Fiestel algorithm using_,the genetic
based biometric key. Once the cipher text is received by the
receiver, the cipher text is decrypted by using_Jhe same key.
D.
Figure 3. MANET Structure
The security functions of the proposed system
Confidentiality: The privacy of the message, is protected
by this scheme. Suppose if the attacker wants to derive the
original message, from the encrypted text, he needs the
cryptographic key. The key can be obtained only by using^
the biometric of the receiver. Furthermore the biometric is
not used as such instead a cancelable version is used. So, it
is computationally infeasible to ge,t the key.
Authentication: In our proposed scheme, the members of
the ad hoc group can authenticate each other through their
biometric. If the receiver wants to verify whether the
message, is commg_from the genuine sender, the message,
can be encrypted by using_the sender's biometric and the
receiver can use the same biometric to decrypt the
message,. These processes can be specified by the
following_£brmulae:
C = % s ( F=) and F^ % s ( C )
where>Ls is th^Ley created by Sender's Biometric.
Integrity: In our proposed scheme, the recipient can verify
whether the received message, is the original one that was
sent by the sender. If the attacker changes the cipher text,
the original plain text can not be generated after decrypting^
with the key created by using_receiver's biometric. By the
property of one-way hash function, it is computationally
infeasible for the attacker to modify the cipher text.
Man-in-the-middle attackr An attacker sits between the
sender and the receiver and sniffs any information beings
sent between two ends is called man in the middle attackz
Even though the attacker can ge,t the cipher text he cannot
view the original message, since it is secured using_ge,netic
based biometric cryptography.
E. Security Analysis
This section reports the analysis of the security parameters
like time taken for key generation, encryption and decryption
for various algorithms like 3DES192, AES128, AES256 and
GBBSS64 in an ad hoc networks environment. The graphs
shown in figure 4 and figure 5 are generated by usin^the
values giyen in the folio wmgjable 1 :
Encryption
Algorithm
Parameters
Key
Size
Time taken
for Key
Generation
Time taken for
Encryption
Time taken for
Decryption
3DES192
192
0.08 ms
0.08 ms
0.07 ms
AES-128
128
0.13 ms
0.1 ms
0.1 ms
AES-256
256
0.13 ms
0.12 ms
0.11 ms
GBBSS-64
64
0.06 ms
0.04 ms
0.02 ms
Table L>Ley si^ and Timing_nieasurements for various algorithms
0.14 -
0.12 -
</> 0.1 -
E
c 0.08 -
| 0.06 -
•- 0.04 -
0.02 -
-
3DES-192 AES-128 AES-256 GBBSS-64
Algorithms Applied
- Key Generation — ■— Encryption Decryption
Figure 4. Timing_nieasurements for various algorithms
3DES192 AES-128 AES-256 GBBSS-64
Algorithms Applied
- Key size
- Security Level |
Figure 5.^Eey Si^ and Security Levels for various algorithms
From the above charts we can understand that our proposed
GBBSS achieves relatively higl} performance in terms of less
overhead and higl} security level. Since the key si^ is very
small compared to the other algorithms, the time taken to
generate the key, time taken to encrypt and decrypt are also
less.
152
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
IV. CONCLUSION AND FUTURE WO^
Although MANET is a very promising^ technology,
challenges are slowing its development and deployment.
Traditional security mechanisms are not sufficient for the
nodes roaming^in a hostile environment with relatively poor
physical protection. Therefore to strengthen the encryption
algorithm and key, first the advantages of biometric and
genetic algorithms are taken into our system. Secondly,
security should be achieved by using^ simple algorithms that
involve small inherent delays rather than complex algorithms
which occupy considerable memory and delay. Finally, ad hoc
networks may consist of thousands of nodes. So, security
mechanisms should be scalable to handle such a large networks
The method presented in this paper remains as a
preliminary approach to reali^ biometric security in ad hoc
networks which needs higjj security. This approach can be used
in very critical, crucial and vital applications where data
security is very important and members who have accessed that
data is limited in number like military officers at war-field,
scientists in a confidential conference, officers in the intelligent
buildings, etc. There are many security problems still persist in
these types of ad-hoc networks and as a future work^this paper
can be extended to solve those problems with different
biometrics and also with multimodal biometrics.
[5] Qin^an Xiao, "A Biometric Authentication Approach for
Higl} Security Ad hoc Networks", F?bceedmg§ of IEEE
Workshop on Information Assistance, pp. 250-256, Jme
2004.
[6] Je Liu, F. Richard Yu, Chung^Horng^Lung_,and Helen
Tang^ "Optimal Biometric-Based Continuous
Authentication in Mobile Ad hoc Networks", Third IEEE
International Conference on Wireless and Mobile
Computing^Networkmg-and Communications, pp. 76-81,
2007.
[7] B Ananda^Lrishna, S Radha an(M_ Chenna^Lesava
Reddy, "Data Security in Ad hoc Networks using_,
Randomi^tion of Cryptographic Algorithms", Jmrnal of
Applied Sciences, pp. 4007-4012, 2007.
[8] 2Lr^ L., Fe"gueroles J and Soriano M "Interpretation of
Binary Strings, as Security Frbtocols for their Evolution
by means of Genetic Algorithms", International
Conference on Database and Expert Systems
Applications, pp. 708-712, 2007.
[9] B. Shanthini and S. Swamynathan "A Cancelable
Biometric-Based Security System for Mobile Ad Hoc
Networks", International Conference on Computer
Technology (ICONCT 09), pp. 179-184, December, 2009.
[10] A. Jtgadeesan, T. Thillaikkarasi ancf^L. Duraiswamy,
" Crypto gr.aphic^Ley Generation from Multiple Biometric
Modalities: Fusing Minutiae with Iris Feature",
International Jmrnal of Computer Applications , Vol. 2,
No.6, pp. 0975-8887, .line 2010.
References
[l] Stalling^ W, "Cryptography and Networks Security-
FFmciples and Practices", 3 r Edition, F^arson Education,
2004.
[2] Animesh^L. Trivedi, Raj an Arora, Rishi^Lapoor, Sudip
Sanyal, Ajith Abraham, Sugata Sanyal, "Mobile Ad Hoc
NetworltSecurity Vulnerabilities", IGI Global, 2009.
[3] Maltoni D. Maio, Jiin A^L. and Frabhakar S, "Handbooks
of Fingerprint Recognition", Springer Verlag^ 2003 .
[4] Fessi B A, Ben Abdallah, S, Hamdi Mand Boudrig§, "A
new genetic algorithm approach for intrusion response
system in computer networks", IEEE Symposium on
Computers and Communications, pp. 342-347, 2009.
BShanthini is a research scholar in
Anna University, Chennai, India.
She received her Bachelor's degree
in C.S.E. from MsL.University,
Madurai and Master's degree in
C.S.E. from M.S. University,
Tirunelveli. Her research interests
include Networks Security, Web
Security, Wireless Communication,
Biometrics and Cloud Computing^
Dr. S. Swamynathan is an
Assistant Professor of Computer
Science and Engineering at Anna
University Chennai, India. He
received his Master's in Computer
Science and Engineering^ and
Doctorate in Reactive Web Services
from Anna University, Chennai. His
research interests include Web
Service, Security, Web Mining_,and
Automated Workflow Systems.
153
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
Effective Multi-Stage Clustering for Inter- and
Intra-Cluster Homogeneity
Sunita M. Karad
Assistant Professor of Computer Engineering,
MIT, Pune, INDIA
V.M.Wadhai
Professor and Dean of Research, MITSOT,
MAE, Pune, INDIA
M.U.Kharat
Principle of Pankaj Laddhad IT,
Yelgaon, Buldhana, INDIA
Prasad S.Halgaonkar
Faculty of Computer Engineering
MITCOE, Pune, INDIA
Dipti D. Patil
Assistant Professor of Computer Engineering,
MITCOE, Pune, INDIA
Abstract - A new algorithm for clustering high-dimensional
categorical data is proposed and implemented by us. This
algorithm is based on a two-phase iterative procedure and
is parameter-free and fully-automatic. Cluster assignments
are given in the first phase, and a new cluster is added to
the partition by identifying and splitting a low-quality
cluster. Optimization of clusters is carried out in the
second phase. This algorithm is based on quality of cluster
in terms of homogeneity. Suitable notion of cluster
homogeneity can be defined in the context of high-
dimensional categorical data, from which an effective
instance of the proposed clustering scheme immediately
follows. Experiment is carried out on real data; this
innovative approach leads to better inter- and intra-
homogeneity of the clusters obtained.
Index Terms - Clustering, high-dimensional categorical
data, information search and retrieval.
I. INTRODUCTION
Clustering is a descriptive task that seeks to identify
homogeneous groups of objects based on the values of their
attributes (dimensions) [1] [2]. Clustering techniques have
been studied extensively in statistics, pattern recognition, and
machine learning. Recent work in the database community
includes CLARANS, BIRCH, and DBSCAN. Clustering is an
unsupervised classification technique. A set of unlabeled
objects are grouped into meaningful clusters, such that the
groups formed are homogeneous and neatly separated.
Challenges for clustering categorical data are: 1) Lack of
ordering of the domains of the individual attributes.
2) Scalability to high dimensional data in terms of
effectiveness and efficiency. High-dimensional categorical
data such as market-basket has records containing large
number of attributes. 3) Dependency on parameters. Setting of
many input parameters is required for many of the clustering
techniques which lead to many critical aspects.
Parameters are useful in many ways. Parameters
support requirements such as efficiency, scalability, and
flexibility. For proper tuning of parameters a lot of effort is
required. As number of parameters increases, the problem of
parameter tuning also increases. Algorithm should have as less
parameters as possible. If the algorithm is automatic it helps to
find accurate clusters. An automatic approach technique
searches huge amounts of high-dimensional data such that it is
effective and rapid which is not possible for human expert. A
parameter free approach is based on decision tree learning,
which is implemented by top-down divide-and-conquer
strategies. The above mentioned problems have been tackled
separately, and specific approaches are proposed in the
literature, which does not fit the whole framework. The main
objective of this paper is to face the three issues in a unified
framework. We look forward to an algorithmic technique that
is capable of automatically detecting the underlying interesting
structure (when available) on high-dimensional categorical
data.
We present Two Phase Clustering (MPC), a new
approach to clustering high-dimensional categorical data that
scales to processing large volumes of such data in terms of
both effectiveness and efficiency. Given an initial data set, it
searches for a partition, which improves the overall purity.
The algorithm is not dependent on any data-specific parameter
(such as the number of clusters or occurrence thresholds for
frequent attribute values). It is intentionally left parametric to
154
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
the notion of purity, which allows for adopting the quality
criterion that best meets the goal of clustering. Section-2
reviews some of the related work carried out on transactional
data, high dimensional data and high dimensional categorical
data. Section-3 provides background information on the
clustering of high dimensional categorical data (MPC
algorithm). Section-4 describes implementation results of
MPC algorithm. Section-5 concludes the paper and draws
direction to future work.
II. RELATED WORK
In current literature, many approaches are given for clustering
categorical data. Most of these techniques suffer from two
main limitations, 1) their dependency on a set of parameters
whose proper tuning is required and 2) their lack of scalability
to high dimensional data. Most of the approaches are unable to
deal with the above features and in giving a good strategy for
tuning the parameters.
Many distance-based clustering algorithms [3] are
proposed for transactional data. But traditional clustering
techniques have the curse of dimensionality and the sparseness
issue when dealing with very high-dimensional data such as
market-basket data or Web sessions. For example, the K-
Means algorithm has been adopted by replacing the cluster
mean with the more robust notion of cluster medoid (that is,
the object within the cluster with the minimal distance from
the other points) or the attribute mode [4]. However, the
proposed extensions are inadequate for large values of m:
Gozzi et al. [5] describe such inadequacies in detail and
propose further extensions to the K-Means scheme, which fit
transactional data. Unfortunately, this approach reveals to be
parameter laden. When the number of dimensions is high,
distance-based algorithms do not perform well. Indeed, several
irrelevant attributes might distort the dissimilarity between
tuples. Although standard dimension reduction techniques [6]
can be used for detecting the relevant dimensions, these can be
different for different clusters, thus invalidating such a
preprocessing task. Several clustering techniques have been
proposed, which identify clusters in subspaces of maximum
dimensionality (see [7] for a survey). Though most of these
approaches were defined for numerical data, some recent work
[8] considers subspace clustering for categorical data.
A different point of view about (dis)similarity is
provided by the ROCK algorithm [9]. The core of the
approach is an agglomerative hierarchical clustering procedure
based on the concepts of neighbors and links. For a given
tuple x, a tuple y is a neighbor of x if the Jaccard similarity
J(x, y) between them exceeds a prespecified threshold O. The
algorithm starts by assigning each tuple to a singleton cluster
and merges clusters on the basis of the number of neighbors
(links) that they share until the desired number of clusters is
reached. ROCK is robust to high-dimensional data. However,
the dependency of the algorithm to the parameter O makes
proper tuning difficult.
Categorical data clusters are considered as dense
regions within the data set. The density is related to the
frequency of particular groups of attribute values. The higher
the frequency of such groups the stronger the clustering.
Preprocessing the data set is carried by extracting relevant
features (frequent patterns) and discovering clusters on the
basis of these features. There are several approaches
accounting for frequencies. As an example, Yang et al. [10]
propose an approach based on histograms: The goodness of a
cluster is higher if the average frequency of an item is high, as
compared to the number of items appearing within a
transaction. The algorithm is particularly suitable for large
high-dimensional databases, but it is sensitive to a user
defined parameter (the repulsion factor), which weights the
importance of the compactness/sparseness of a cluster. Other
approaches [11], [12], [13] extend the computation of
frequencies to frequent patterns in the underlying data set. In
particular, each transaction is seen as a relation over some sets
of items, and a hyper-graph model is used for representing
these relations. Hyper-graph partitioning algorithms can hence
be used for obtaining item/transaction clusters.
The CLICKS algorithm proposed in [14] encodes a
data set into a weighted graph structure G(N, E), where the
individual attribute values correspond to weighted vertices in
N, and two nodes are connected by an edge if there is a tuple
where the corresponding attribute values co-occur. The
algorithm starts from the observation that clusters correspond
to dense (that is, with frequency higher than a user-specified
threshold) maximal k-partite cliques and proceeds by
enumerating all maximal k-partite cliques and checking their
frequency. A crucial step is the computation of strongly
connected components, that is, pairs of attribute values whose
co-occurrence is above the specified threshold. For large
values of m (or, more generally, when the number of
dimensions or the cardinality of each dimension is high), this
is an expensive task, which invalidates the efficiency of the
approaches. In addition, technique depends upon a set of
parameters, whose tuning can be problematic in practical
cases.
Categorical clustering can be tackled by using
information-theoretic principles and the notion of entropy to
measure closeness between objects. The basic intuition is that
groups of similar objects have lower entropy than those of
dissimilar ones. The COOLCAT algorithm [15] proposes a
scheme where data objects are processed incrementally, and a
suitable cluster is chosen for each tuple such that at each step,
the entropy of the resulting clustering is minimized. The
scaLable InforMation Bottleneck (LIMBO) algorithm [16]
also exploits a notion of entropy to catch the similarity
between objects and defines a clustering procedure that
minimizes the information loss. The algorithm builds a
Distributional Cluster Features (DCF) tree to summarize the
data in k clusters, where each node contains statistics on a
subset of tuples. Then, given a set of k clusters and their
corresponding DCFs, a scan over the data set is performed to
assign each tuple to the cluster exhibiting the closest DCF.
The generation of the DCF tree is parametric to a user-defined
branching factor and an upper bound on the distance between
a leaf and a tuple.
Li and Ma [17] propose an iterative procedure that is
aimed at finding the optimal data partition that minimizes an
155
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
entropy-based criterion. Initially, all tuples reside within a
single cluster. Then, a Monte Carlo process is exploited to
randomly pick a tuple and assign it to another cluster as a trial
step aimed at decreasing the entropy criterion. Updates are
retained whenever entropy diminishes. The overall process is
iterated until there are no more changes in cluster assignments.
Interestingly, the entropy-based criterion proposed here can be
derived in the formal framework of probabilistic clustering
models. Indeed, appropriate probabilistic models, namely,
multinomial [18] and multivariate Bernoulli [19], have been
proposed and shown to be effective. The classical
Expectation-Maximization framework [20], equipped with any
of these models, reveals to be particularly suitable for dealing
with transactional data [21], [22], being scalable both in n and
in m. The correct estimation of an appropriate number of
mixtures, as well as a proper initialization of all the model
parameters, is problematic here.
The problem of estimating the proper number of
clusters in the data has been widely studied in the literature.
Many existing methods focus on the computation of costly
statistics based on the within-cluster dispersion [23] or on
cross-validation procedures for selecting the best model [24],
[25]. The latter requires an extra computational cost due to a
repeated estimation and evaluation of a predefined number of
models. More efficient schemes have been devised in [26],
[27]. Starting from an initial partition containing a single
cluster, the approaches iteratively apply the K-Means
algorithm (with k = 2) to each cluster so far discovered. The
decision on whether to switch the original cluster with the
newly generated sub-clusters is based on a quality criterion,
for example, the Bayesian Information Criterion [26], which
mediates between the likelihood of the data and the model
complexity, or the improvement in the rate of distortion (the
variance in the data) of the sub-clusters with respect to the
original cluster [27]. The exploitation of the K-Means scheme
makes the algorithm specific to low-dimensional numerical
data, and proper tuning to high-dimensional categorical data is
problematic.
Automatic approaches that adopt the top-down
induction of decision trees are proposed in [28], [29], [30].
The approaches differ in the quality criterion adopted, for
example reduction in entropy [28], [29] or distance among the
prototypes of the resulting clusters [29]. All of these
approaches have some of the drawbacks. The scalability on
high-dimensional data is poor. Some of the literature that
focused on high dimensional categorical data is available in
[31], [32].
III. The MPC Algorithm
The key idea of Two Phase Clustering (MPC) algorithm is to
develop a clustering procedure, which has the general sketch
of a top-down decision tree learning algorithm. First, start
from an initial partition which contains single cluster (the
whole data set) and then continuously try to split a cluster
within the partition into two sub-clusters. If the sub-clusters
have a higher homogeneity in the partition than the original
cluster, the original is removed. The sub-clusters obtained by
splitting are added to the partition. Split the clusters on the
basis of their homogeneity. A function Quality(C) measures
the degree of homogeneity of a cluster C. Clusters with high
intra-homogeneity exhibit high values of Quality.
Let M be set of Boolean attributes such that M =
{a l9 , a m } and a data set D = {x l9 x 2 , , x n } of tuples which
is defined on M. a e M is denoted as an item, and a tuple x e D
as a transaction x. Data sets containing transactions are
denoted as transactional data, which is a special case of high-
dimensional categorical data. A cluster is a set S which is a
subset of D. The size of S is denoted by n s , and the size of M s
= {a|a C x, x C S} is denoted by m s . A partitioning problem is
to divide the original collection of data D into a set P =
{Ci, ,C k } where each clusters Cj are nonempty. Each
cluster contains a group of homogeneous transactions.
Clusters where transactions have several items have higher
homogeneity than other subsets where transactions have few
items. A cluster of transactional data is a set of tuples where
few items occur with higher frequency than somewhere else.
Our approach to clustering starts from the analysis of
the analogies between a clustering problem and a
classification problem. In both cases, a model is evaluated on
a given data set, and the evaluation is positive when the
application of the model locates fragments of the data
exhibiting high homogeneity. A simple rather intuitive and
parameter-free approach to classification is based on decision
tree learning, which is often implemented through top-down
divide and conquers strategies. Here, starting from an initial
root node (representing the whole data set), iteratively, each
data set within a node is split into two or more subsets, which
define new sub-nodes of the original node. The criterion upon
which a data set is split (and, consequently, a node is
expanded) is based on a quality criterion: choosing the best
"discriminating" attribute (that is, the attribute producing
partitions with the highest homogeneity) and partitioning the
data set on the basis of such attribute. The concept of
homogeneity has found several different explanations (for
example, in terms of entropy or variance) and, in general, is
related to the different frequencies of the possible labels of a
target class.
The general schema of the MPC algorithm is
specified in Fig. 1 . The algorithm starts with a partition having
a single cluster i.e whole data set (line 1). The central part of
the algorithm is the body of the loop between lines 2 and 15.
Within the loop, an effort is made to generate a new cluster by
1) choosing a candidate node to split (line 4), 2) splitting the
candidate cluster into two sub-clusters (line 5), and (line 3)
calculating whether the splitting allows a new partition with
better quality than the original partition (lines 6-13). If this is
true, the loop can be stopped (line 10), and the partition is
updated by replacing the original cluster with the new sub-
clusters (line 8). Otherwise, the sub-clusters are discarded, and
a new cluster is taken for splitting.
The generation of a new cluster calls STABILIZE-
CLUSTERS in line 9, improves the overall quality by trying
relocations among the clusters. Clusters at line 4 are taken in
increasing order of quality,
a. Splitting a Cluster
156
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
A splitting procedure gives a major improvement in
the quality of the partition. Choose the attribute that gives the
highest improvement in the quality of the partition.
GENERATE-CLUSTERS (D)
Input: A set D ={x 1/ ...,x N } of transactions;
Output: A partition P = {C l7 ...,C k } of clusters;
1.
Let initially P = {D};
2.
repeat
3.
Generate a new cluster C initially empty;
4.
for each cluster C, E P do
5.
PARTITION-CLUSTERS(QC);
6.
P' <— PU {C};
7.
if Quality(P) < Quality(P') then
8.
P <— P';
9.
STABIUZE-CLUSTERS(P);
10.
break
11.
else
12.
Restore all x y gCintoC,;
13.
end if
14.
end for
15.
until no further cluster C can be generated
Figure 1 : Generate Clusters
PARTITION-CLUSTER(aO)
PI.
repeat
P2.
for all x ECiU C 2 do
P3.
if cluster(x) = CI then
P4.
C u ^ Ci; C v ^ C 2 ;
P5.
else
P6.
C u ^ C 2 ; C v ^ Ci;
P7.
end if
P8.
Qi <— Quality(C u ) + Quality(C v );
P9.
Q s <— Quality(C u -{x}) + Quality(C v U{x});
P10.
if Qs > Qj then
Pll.
C u .Remove(x);
P12.
C v .lnsert(x);
P13.
end if
P14.
end for
P15.
until C x and C 2 are stable
Figure 2: Partition Cluster
PARTITION-CL USTER
The PARTITION-CLUSTER algorithm is given in Fig.2. The
algorithm continuously evaluates, for each element x eQU
C 2 , to check whether a reassignment increases the
homogeneity of the two clusters.
Lines P8 and P9 compute the involvement of a: to the
local quality in two cases: either x remains in its original
cluster (C u ) or x is moved to the other cluster (C v ). If moving x
gives an improvement in the local quality, then the swapping
is done (lines P10-P13). Lines P2-P14 in the algorithm is
nested into a main loop: elements are continuously checked
for swapping until a convergence is met. The splitting process
can be sensitive to the order upon which elements are
considered: In the first stage, it could be not convenient to
reassign the generic Xi from Ci to C 2 , whereas a convenience
in performing the swap can be found after the relocation of
some other element Xj. The main loop partly smoothes this
effect by repeatedly relocating objects until convergence is
met. Better PARTITION-CLUSTER can be made strongly
insensitive to the order with which cluster elements are
considered. The basic idea is discussed next. The elements that
mostly influence the locality effect are either outlier
transactions (that is, those containing mainly items, whose
frequency within the cluster is rather low) or common
transactions (which, dually, contain very frequent items). In
the first case, C 2 is unable to attract further transactions,
whereas in the second case, C 2 is likely to attract most of the
transactions (and, consequently, Ci will contain outliers).
The key idea is to rank and sort the cluster elements
before line PI, which is on the basis of their splitting
effectiveness. To this purpose, each transaction x belonging to
cluster C can be associated with a weight w(x), which
indicates its splitting effectiveness, x is eligible for splitting C
if its items allow us to divide C into two homogeneous sub-
clusters. In this respect, the Gini index is a natural way to
quantify the splitting effectiveness G(a) of the individual
attribute value a 6 x. Precisely, G(a) = 1 - Pr(a|C) 2 -
(1 - Pr(a|C)) 2 , where Pr(a|C) denotes the probability of a
within C. G(a) is close to its maximum whenever a is present
in about half of the transactions of C and reaches its minimum
whenever a is unfrequent or common within C. The overall
splitting effectiveness of x can be defined by averaging the
splitting effectiveness of its constituting items
w(x) = avg a g x (G(a)). Once ranked, the elements x e C can be
considered in descending order of their splitting effectiveness
at line P2. This guarantees that C 2 is initialized with elements,
which do not represent outliers and still are likely to be
removed from C x . This removes the dependency on the initial
input order of the data. With decision tree learning, MPC
exhibits a preference bias, which is encoded within the notion
of homogeneity and can be viewed as the preference for
compact clustering trees. Indeed, due to the splitting
effectiveness heuristic, homogeneity is enforced by the effects
of the Gini index. At each split, this tends to isolate clusters of
transactions with mostly frequent attribute values, from which
the compactness of the overall clustering tree follows.
b. STABILIZE-CLUSTERS
PARTITION-CLUSTER improves the local quality
of a cluster. And STABILIZE-CLUSTERS try to increase
partition quality. It is carried out by finding the most suitable
clusters for each element among the ones which are there in
the partition.
Fig. 3 shows the pseudo code of the procedure. The
central part of the algorithm is a main loop which (lines S2-
157
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
SI 7) examines all the available elements. For each element x,
a pivot cluster is identified, which is the cluster containing x.
Then, the available clusters are continuously evaluated. The
insertion of x in the current cluster is done (lines S5-S6), and
the updated quality is compared with the original quality.
STABILIZE
-CLUSTERS^
SI.
repeat
S2.
for all x E0do
S3.
Cpi V ot **— cluster(x); Q *— Quality (P);
S4.
for all CEP do
S5.
C p/Vot .REMOVE(x);
S6.
CINSERT(x);
S7.
if Quality(P) > Q then
S8.
if C pivot = then
S9.
P.REMOVE(C p/Vot );
S10.
end if
Sll.
Cpivot +— Q Q +— Quality (P);
S12.
else
S13.
C pivot .INSERT(x);
S14.
C.REMOVE(x);
S15.
end if
S16.
end for
S17.
end for
S18.
until P is stable
Figure 3 : Stabilize Clusters
If an improvement is obtained, then the swap is accepted (line
Sll). The new pivot cluster is the one now containing x, and if
the removal of x makes the old pivot cluster empty, then the
old pivot cluster is removed from the partition P. If there is no
improvement in quality, x is restored into its pivot cluster, and
a new cluster is examined. The main loop is iterated until a
stability condition for clusters is achieved.
c. Cluster and Partition Qualities
AT-DC gives two different quality measures, 1) local
homogeneity within a cluster and 2) global homogeneity of the
partition. As shown in Fig. 1, it is noticed that partition quality
is used for checking whether the insertion of a new cluster is
really suitable: it is for maintaining compactness. Cluster
quality in procedure PARTITIONCLUSTER is done for good
separation.
Cluster quality is known when there is a high degree
of intracluster homogeneity and intercluster homogeneity. As
given in [35], there is strong relation between intracluster
homogeneity and the probability Pr(ai|C k ) that item ai appears
in a transaction containing in C k There is a strong relationship
between intercluster separation and Pr(x E C k , ai Ex). Cluster
homogeneity and separation is computed by relating it to the
unity of items within the transactions that it contains. Cluster
quality is equal to the combination of the above probability,
2 ff tt-VD Pi (a|£)Fr. (S|fl)Fr. (fli The last term is used
for weighting the importance of item a in the summation:
Essentially, high values from low-frequency items are less
relevant than those from high-frequency values. By the Bayes
theorem, the above formula is expressed as
MC)I^, ym Jr.(m- t 33 ]- Terms Fr(>|£?
(relative strength of a within C) and Pr(C) (relative strength of
C) work in contraposition. It is easy to compute the gain in
strength for each item with respect to the whole data set, that
is
Quality (CO = Pr(Q) Eh-s^Pr ft Cft* - Rr.tffl.fl^ ]
(1)
Where,
C k - cluster
Pr(C k ) - relative strength of C k
a C MC k - an item
M = {a 1} , a m } is set of Boolean attributes
Pr(a| Ck) - relative strength of a within C k
Pr(a|D) - relative strength of a within D
D = {x b , x n } is data set of tuples defined on M
Qualit y( C0^I y ^ t (^.(^]
(2)
where na and Na represent the frequencies of a in C and D,
respectively. The value of Quality (C k ) is updated as soon as a
new transaction is added to C.
IV. RESULTS AND ANALYSIS
Two real-life data sets were evaluated. A description of each
data set employed for testing is provided next, together with
an evaluation of the MPC performances.
UCI DATASETS [34]
Zoo: Zoo dataset contains 103 instances, each having 18
attributes (animal name, 15 Boolean attributes and 2
numerics). The "type" attribute appears to be the class
attribute. In total there are 7 classes of animals, that is, class 1
has 41 set of animals, class 2 has 20 set of animals, class 3 has
5 set of animals, class 4 has 13 set of animals, class 5 has 4 set
of animals, class 6 has 8 set of animals and class 7 has 10 set
of animals. Here is a breakdown of which animals are in
which type: (it is unusual that there are 2 instances of "frog"
and one of "girl"!). There are no missing values in this dataset.
Table 1 shows that in cluster 1, a class 2 is having high
homogeneity and in cluster 2, classes 3, 5 and 7 are having
high homogeneity.
Hepatitis: Hepatitis contains 155 instances, each having 20
attributes. It represents the observation of patients. Each
instance is one patient's record according to 20 attributes (for
example, age, steroid, antivirals, and spleen palpable). Some
attributes contains missing values. A class as "DIE" or
158
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
"LIVE" is given to each instance. Out of 155 instances, 32 are
"DIE" and 123 are "LIVE". Table 2 shows that in cluster 1
and cluster 2 are having high homogeneity. In cluster 2 and 4
there are 2 (DIE) and 1 (LIVE) instances which are
misclassified.
Table 1: Confusion matrix for zoo
Cluster No.
Classes
1
2
3
4
5
6
7
1
17
20
5
2
2
24
5
8
4
6
10
Table 2: Confusion matrix for Hepatitis
Cluster No.
Classes
DIE
LIVE
1
17
2
2
63
3
59
4
13
1
V. CONCLUDING REMARK
This innovative MPC algorithm is fully-automatic, parameter-
free approach to cluster high-dimensional categorical data.
The main advantage of our approach is its capability of
avoiding explicit prejudices, expectations, and presumptions
on the problem at hand, thus allowing the data itself to speak.
This is useful with the problem at hand, where the data is
described by several relevant attributes.
A limitation of our proposed approach is that the
underlying notion of cluster quality is not meant for catching
conceptual similarities, that is, when distinct values of an
attribute are used for denoting the same concept. Probabilities
are provided to evaluate cluster homogeneity only in terms of
the frequency of items across the underlying transactions.
Hence, the resulting notion of quality suffers from the typical
limitations of the approaches, which use exact-match
similarity measures to assess cluster homogeneity. To this
purpose, conceptual cluster homogeneity for categorical data
can be easily added to the framework of the MPC algorithm.
Another limitation of our approach is that it cannot
deal with outliers. These are transactions whose structure
strongly differs from that of the other transactions being
characterized by low-frequency items. A cluster containing
such transaction exhibits low quality. Worst, outliers could
negatively affect the PARTITION-CLUSTER procedure by
preventing the split to be accepted (because of an arbitrary
assignment of such outliers, which would lower the quality of
the partitions). Hence, a significant improvement of MPC can
be obtained by defining an outlier detection procedure that is
capable of detecting and removing outlier transactions before
partitioning the clusters. The research work can be extended
further to improve the quality of clusters by removing
outliers.
REFERENCES
[I] J. Grabmeier and A. Rudolph, "Techniques of Cluster Algorithms in
Data Mining," Data Mining and Knowledge Discovery, vol. 6, no. 4,
pp. 303-360, 2002.
[2] A. Jain and R. Dubes, Algorithms for Clustering Data. Prentice Hall,
1988.
[3] R. Ng and J. Han, "CLARANS: A Method for Clustering Objects for
Spatial Data Mining," IEEE Trans. Knowledge and Data Eng., vol. 14,
no. 5, pp. 1003-1016, Sept./Oct. 2002.
[4] Z. Huang, "Extensions to the K-Means Algorithm for Clustering Large
Data Sets with Categorical Values," Data Mining an Knowledge
Discovery, vol. 2, no. 3, pp. 283-304, 1998.
[5] C. Gozzi, F. Giannotti, and G. Manco, "Clustering Transactional Data,"
Proc. Sixth European Conf. Principles and Practice of Knowledge
Discovery in Databases (PKDD '02), pp. 175-187, 2002.
[6] S. Deerwester et al., "Indexing by Latent Semantic Analysis," J. Am.
Soc. Information Science, vol. 41, no. 6, 1990.
[7] L. Parsons, E. Haque, and H. Liu, "Subspace Clustering for High-
Dimensional Data: A Review," SIGKDD Explorations, vol. 6, no. 1, pp.
90-105,2004.
[8] G. Gan and J. Wu, "Subspace Clustering for High Dimensional
Categorical Data," SIGKDD Explorations, vol. 6, no. 2, pp. 87-94, 2004.
[9] M. Zaki and M. Peters, "CLICK: Mining Subspace Clusters in
categorical Data via k-Partite Maximal Cliques," Proc. 21st IntT Conf.
Data Eng. (ICDE '05), 2005.
[10] Y. Yang, X. Guan, and J. You, "CLOPE: A Fast and Effective
Clustering Algorithm for Transactional Data," Proc. Eighth ACM Conf.
Knowledge Discovery and Data Mining (KDD '02), pp. 682-687, 2002.
[II] E. Han, G. Karypis, V. Kumar, and B. Mobasher, "Clustering in a High
Dimensional Space Using Hypergraph Models," Proc. ACM SIGMOD
Workshops Research Issues on Data Mining and Knowledge Discovery
(DMKD '97), 1997.
[12] M. Ozdal and C. Aykanat, "Hypergraph Models and Algorithms for
Data-Pattern-Based Clustering," Data Mining and Knowledge
Discovery, vol. 9, pp. 29-57, 2004.
[13] K. Wang, C. Xu, and B. Liu, "Clustering Transactions Using Large
Items," Proc. Eighth IntT Conf. Information and Knowledge
Management (CIKM '99), pp. 483-490, 1999.
[14] D. Barbara, J. Couto, and Y. Li, "COOLCAT: An Entropy-Based
Algorithm for Categorical Clustering," Proc. 11th ACM Conf.
Information and Knowledge Management (CIKM '02), pp. 582-589,
2002.
[15] P. Andritsos, P. Tsaparas, R. Miller, and K. Sevcik, "LIMBO: Scalable
Clustering of Categorical Data," Proc. Ninth IntT Conf. Extending
Database Technology (EDBT '04), pp. 123-146, 2004.
[16] M.O.T. Li and S. Ma, "Entropy-Based Criterion in Categorical
Clustering," Proc. 21st IntT Conf. Machine Learning (ICML '04), pp.
68-75, 2004.
[17] I. Cadez, P. Smyth, and H. Mannila, "Probabilistic Modeling of
Transaction Data with Applications to Profiling, Visualization, and
Prediction," Proc. Seventh ACM SIGKDD IntT Conf. Knowledge
Discovery and Data Mining (KDD '01), pp. 37-46, 2001.
[18] M. Carreira-Perpinan and S. Renals, "Practical Identifiability of Finite
Mixture of Multivariate Distributions," Neural Computation, vol. 12, no.
l,pp. 141-152,2000.
[19] G. McLachlan and D. Peel, Finite Mixture Models. John Wiley & Sons,
2000.
[20] M. Meila and D. Heckerman, "An Experimental Comparison of Model-
Based Clustering Methods," Machine Learning, vol. 42, no. 1/2, pp. 9-
29,2001.
[21] J.G.S. Zhong, "Generative Model-Based Document Clustering: A
Comparative Study," Knowledge and Information Systems, vol. 8, no. 3,
pp. 374-384, 2005.
[22] A. Gordon, Classification. Chapman and Hall/CRC Press, 1999.
159
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
[23] C. Fraley and A. Raftery, "How Many Clusters? Which Clustering
Method? The Answer via Model-Based Cluster Analysis," The
Computer J., vol. 41, no. 8, 1998.
[24] P. Smyth, "Model Selection for Probabilistic Clustering Using Cross-
Validated Likelihood," Statistics and Computing, vol. 10, no. 1, pp. 63-
72,2000.
[25] D. Pelleg and A. Moore, "X-Means: Extending K-Means with Efficient
Estimation of the Number of Clusters," Proc. 17th Int'l Conf. Machine
Learning (ICML '00), pp. 727-734, 2000.
[26] M. Sultan et al., "Binary Tree- Structured Vector Quantization Approach
to Clustering and Visualizing Microarray Data," Bioinformatics, vol. 18,
2002.
[27] S. Guha, R. Rastogi, and K. Shim, "ROCK: A Robust Clustering
Algorithm for Categorical Attributes," Information Systems, vol. 25, no.
5, pp. 345-366,2001.
[28] J. Basak and R. Krishnapuram, "Interpretable Hierarchical Clustering by
Constructing an Unsupervised Decision Tree," IEEE Trans. Knowledge
and Data Eng., vol. 17, no. 1, Jan. 2005.
[29] H. Blockeel, L.D. Raedt, and J. Ramon, "Top-Down Induction of
Clustering Trees," Proc. 15th Int'l Conf. Machine Learning (ICML'98),
pp. 55-63, 1998.
[30] B. Liu, Y. Xia, and P. Yu, "Clustering through Decision Tree
Construction," Proc. Ninth Int'l Conf. Information and Knowledge
Management (CIKM '00), pp. 20-29, 2000.
[31] Yi-Dong Shen, Zhi-Yong Shen and Shi-Ming Zhang,"Cluster Cores -
based Clustering for High - Dimensional Data".
[32] Alexander Hinneburg and Daniel A. Keim, Markus Wawryniuk,"HD-
Eye-Visual of High-Dimensional Data: A Demonstration".
[33] http://en.wikipedia.org/wiki/Bayes'_theorem
[34] UCI Machine Learning Repository http://www.ics.uci.edu/~mlearn/
[35] D. Fisher, "Knowledge Acquisition via Incremental Conceptual
Clustering," Machine Learning, vol. 2, pp. 139-172, 1987.
AUTHORS PROFILE
Sunita M. Karad has received B.E. degree in
Computer Engineering from Marathvada University,
India in 1992, M.E. degree from Pune University in
2007. She is a registered Ph.D. student of Amravati
University. She is currently working as Assistant
Professor in Computer Engineering department in
MIT, Pune. She has more than 10 years of teaching
experience and successfully handles administrative
work in MIT, Pune. Her research interest includes
Data mining, Business Intelligence & Aeronautical space research.
I
1 1 "
Dr. Madan U. Kharat has received his B.E. from Amravati University, India
in 1992, M.S. from Devi Ahilya University (Indore), India in 1995 and Ph.D.
degree from Amravati University, India in 2006. He has
experience of 18 years in academics. He has been
working as a Principle of PLIT, Yelgaon, Budhana. His
research interest includes Deductive Databases, Data
Mining and Computer Networks.
Prasad
S.
bachelor's
degree
Amravati
Computer
Science
Engineering
y
currently a
lecturer
research
interest
and Data
Mining,
databases
and
Halgaonkar received his
in Computer Science from
University in 2006 and M.Tech in
from Walchand College of
Shivaji University in 2010. He is
in MITCOE, Pune. His current
includes Knowledge discovery
deductive databases, Web
Semi- Structured data.
Dipti D. Patil has received B.E. degree in Computer
Engineering from Mumbai University in 2002 and M.E.
degree in Computer Engineering from Mumbai
University, India in 2008. She has worked as Head &
Assistant Professor in Computer Engineering
Department in Vidyavardhini's College of Engineering
& Technology, Vasai. She is currently working as
Assistant Professor in MITCOE, Pune. Her Research
interests include Data mining, Business Intelligence and
Body Area Network.
I Dr. Vijay M.Wadhai received his B.E. from
! Nagpur University in 1986, M.E. from Gulbarga
University in 1995 and Ph.D. degree from Amravati
University in 2007. He has experience of 25 years
| which includes both academic (17 years) and
I research (8 years). He has been working as a Dean
I of Research, MITSOT, MAE, Pune (from 2009) and
simultaneously handling the post of Director -
Research and Development, Intelligent Radio
Frequency (IRF) Group, Pune (from 2009). He is
currently guiding 12 students for their PhD work in both Computers and
Electronics & Telecommunication area. His research interest includes Data
Mining, Natural Language processing, Cognitive Radio and Wireless
Network, Spectrum Management, Wireless Sensor Network, VANET, Body
Area Network, ASIC Design - VLSI. He is a member of ISTE, IETE, IEEE,
IES and GISFI (Member Convergence Group), India.
J L
160
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
A Pilot Based RLS Channel Estimation for LTE
SC-FDMA in High Doppler Spread
M. M. Rana
Department of Electronics and Communication Engineering
Khulna University of Engineering and Technology
Khunla, Bangladesh
Abstract — Main challenges for a terminal implementation are
efficient realization of the inner receiver, especially for channel
estimation (CE) and equalization. In this paper, pilot based
recursive least square (RLS) channel estimator technique is
investigate for a long term evolution (LTE) single carrier-
frequency division multiple access (SC-FDMA) system in high
Doppler spread environment. This CE scheme uses adaptive RLS
estimator which is able to update parameters of the estimator
continuously, so that knowledge of channel and noise statistics
are not required. Simulation results show that the RLS CE
scheme with 500 Hz Doppler frequency has 3 dB better
performances compared with 1.5 kHz Doppler frequency.
Keywords— Channel estimation, LTE, RLS, SC-FDMA.
I. INTRODUCTION
The 3rd generation partnership project (3 GPP) members
started a feasibility study on the enhancement of the universal
terrestrial radio access (UTRA), to improve the mobile phone
standard to cope with future requirements. This project was
called long term evolution (LTE) [1], [2]. LTE uses
orthogonal frequency division multiple access (OFDM A) for
downlink and single carrier-frequency division multiple
access (SC-FDMA) for uplink transmission [1]. A highly
efficient way to cope with the frequency selectivity of
wideband channel is OFDMA. OFDMA is an effective
technique for combating multipath fading and for high bit rate
transmission over mobile wireless channels. Channel
estimation (CE) has been successfully used to improve the
system performance. It can be employed for the purpose of
detecting received signal, improve signal-to-noise ratio
(SNR), channel equalization, cochannel interference (CCI)
rejection, and improved the system performance [3-5].
In general, CE techniques can be divided into three
categories such as pilot CE, blind CE, and semi-blind CE [10],
[11]. Pilot CE techniques offer low computational complexity
and good performance [12]. The blind CE techniques exploit
the statistical behavior of the received signals and require a
large amount of data [13]. Semi-blind CE methods are used a
combination of data aided and blind methods [11]. The pilot
CE algorithm requires probe sequences; the receiver can use
this probe sequence to reconstruct the transmitted waveform
[6-8]. Pilot symbols can be placed either at the beginning of
each burst as a preamble or regularly through the burst. Pilot
sequences are transmitted at certain positions of the SC-
FDMA frequency time pattern, in its place of data.
Adaptive CE has been, and still is, an area of active research
topics, playing imperative roles in an ever growing number of
applications such as wireless communications where the
channel is rapidly time-varying. Signal processing techniques
that use recursively estimated, time varying models are
normally called adaptive. Different adaptive CE algorithms
have been proposed over the years for the purpose of updating
the channel coefficient. The least mean square (LMS) method,
its normalized version (NLMS), the affine projection
algorithm (APA), as well as the recursive least square (RLS)
method are well known examples of such CE algorithms. The
well known LMS/NLMS CE algorithms are attractive from a
computational complexity point of view but their convergence
behavior for highly correlated input signals is poor. The RLS
CE method resolves this trouble, but at the expense of
increased complexity. A very large number of fast RLS CE
methods have been developed over the years, but regrettably,
it seems that the better a fast RLS CE method is in terms of
computational efficiency and numerical stability. In addition,
the RLS algorithm has the recursive inversion of an estimate
of the autocorrelation matrix of the input signal as its
cornerstone, problems arise, if the autocorrelation matrix is
rank deficient.
In this paper, we investigate the adaptive RLS CE method
in the LTE SC-FDMA systems in high Doppler spread
environment. This CE method uses adaptive estimator which
is able to update parameters of the estimator continuously so
that knowledge of channel and noise statistics are not
required. Simulation results show that the RLS CE scheme
with 500 Hz Doppler frequency has 3 dB better performances
compared with 1500 Hz Doppler frequency.
We use the following notations throughout this paper: bold
face lower letter is used to represent vector. Superscripts x*
and x T denote the conjugate and conjugate transpose of the
complex vector x respectively.
The remainder of the paper is organized as follows: section
II describes wireless communication systems and LTE SC-
FDMA systems model is describes in section III. The RLS CE
scheme is presented in section IV, and its performance is
161
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
analyzed in section V. Finally, some concluding remarks are
given in section VI.
II. WIRELESS COMMUNICATION SYSTEMS
III. LTE SC-FDMA SYSTEMS DESCRIPTION
In this section, we briefly explain LTE SC-FDMA system
model, fading channel statistics, and received signal model.
Nowadays, cellular mobile phones have become an
important tool and part of daily life. In the last decade, cellular
systems have experienced fast development and there are
currently about two billion users over the world. The idea of
cellular mobile communications is to divide large zones into
small cells, and it can provide radio coverage over a wider
area than the area of one cell. This concept was developed by
researchers at AT & T Bell laboratories during the 1950s and
1960s. The initial cellular system was created by nippon
telephone & telegraph (NTT) in Japan, 1979. From then on,
the cellular mobile communication has evolved.
The mobile communication systems are frequently
classified as different generations depending of the service
offered. The first generation (1G) comprises the analog
communication techniques, and it was mainly built on
frequency modulation (FM) and frequency division multiple
accesses (FDMA). Digital communication techniques
appeared in the second generation (2G) systems, and main
access schemes are time division multiple access (TDMA) and
code division multiple access (CDMA). The two most
commonly accepted 2G systems are global system for mobile
(GSM) and interim standard-95 (IS-95). These systems mostly
offer speech communication, but also data communication
limited to rather low transmission rates. The concept of the
third generation (3G) system started operations on October,
2002 in Japan. The 3GPP members started a feasibility study
on the enhancement of the universal terrestrial radio access
(UTRA) in December 2004, to improve the mobile phone
standard to cope with future requirements. This project was
called LTE. LTE uses SC-FDMA for uplink transmission and
OFDMA for downlink transmission. Fig. 1 summarizes the
evolution path of cellular mobile communications systems.
Narrowband
Circuit
switched
Circuit
Wideband
cet switched
Broadband
Fully IP based core network
Analog
FDMA
Digital
TDMA,
CDMA
WCDMA,
CDMA
HSPA,
CDMA,
OFDMA
OFDMA,
SC-FDMA
OFDMA,
SC-FDMA,
SC-CDMA,
MC-CDMA
2G V 3G /3.50V 3.9G
MMT2000 Y 3GLTE
V 14kbpS A 64kbpS 1 2M PP S I 14Mbps A^lOOMl^ W
Voice voice, c
Data with video ) I Voice and data with vide
A. Baseband system model
A baseband block diagram for the communications system
under investigation is shown in Fig. 2.
Bit stream
Bit stream
'
*
Demodulation
Modulation
A
'
i
N-point IDFT
N-point DFT
Pilots signal
generation
4
V
|
— ►
CE and
equalization
Mapping
*
Mapping
restore
'
i
Multipath
fading
channel
AWGN
noise
T
M-point IDFT
M-point
DFT
v
1 !
i i
+
Add cyclic
prefix (CP)
-*&—+&>-+
Synchronization
and remove CP
Fig. 2. Block diagram of a LTE SC-FDMA system.
At the transmitter, a baseband multiple phase shift keying
modulator takes binary sequence and produces the signaling
waveforms
m^t) = x j — cos((Dt + aj, < t < T
[cos^) cos(cot) - sin^) sin(cot)]
= a i b(t) + c i d(t), (1)
where T is the symbol duration, E is the energy of m i (t),
cd = 27rf, f is the carrier frquency, phase anagle
a- , M is the alphabate size, a '. = yj E COS OC
M l
inphasse basis , b(t) =. \— COs((Dt), C i = V £'sina i , and
Fig. 1. Evolution path in mobile communication systems.
quadrature basis, d(t) = -A — COs((Ot). CE is often achieved
by multiplexing known symbols, so called, pilot symbols into
data sequences [1]. These modulated symbols and pilots
perform N-point discrete Fourier transform (DFT) to produce
a frequency domain representation:
162
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
-j2^mt
s 1 (t)=- ? =X m i( t ) e N '
VNtD
where j is the imaginary unit. It then maps each of the N-point
DFT outputs to one of the orthogonal sub-carriers mapping that
can be transmitted. There are two principal sub-carrier mapping
modes: localized mode, and distribution mode. In distributed
sub-carrier mode, the outputs are allocated equally spaced sub-
carrier, with zeros occupying the unused sub-carrier in
between. While in localized sub-carrier mode, the outputs are
confined to a continuous spectrum of sub-carrier. Interleaved
sub-carrier mapping mode of FDMA (IFDMA) is another
special sub-carrier mapping mode [13], [14]. The difference
between DFDMA and IFDMA is that the outputs of IFDMA
are allocated over the entire bandwidth, whereas the DFDMAs
outputs are allocated every several subcarriers [15], [16].
Finally, the inverse DFT (IDFT) module output is followed
by a cyclic prefix (CP) insertion that completes the digital stage
of the signal flow. The CP is used to eliminate ISI and preserve
the orthogonality of the tones. Assume that the channel length
of CP is larger than the channel delay spread [17].
B. Channel model
Channel model is a mathematical representation of the
transfer characteristics of the physical medium. These models
are formulated by observing the characteristics of the received
signal. According to the documents from 3GPP, a radio wave
propagation can be described by multipaths which arise from
reflection and scattering [17]. The received signal at the mobile
terminal is a superposition of all paths. If there are L distinct
paths from transmitter to the receiver, the impulse response of
the multipath fading channel can be represented as [17]:
L
CD(nvc) = ^©.(rrj) 8[m- x.(m)], (3)
where CO . (m) and T . (m) are attenuations and delays for each
path at time instant m, and 8(.) is the Dirac delta function. The
fading process for the mobile radio channel is given by
(D(v) = «) jA /l-(v/f d ) , (4)
where Doppler frquency f d = s/X, s is the speed of the mobile,
and X is the wavelength of the transmitted carrier. In order to
do simulations as close to the reality as possible, it is essential
to have a good channel model. This model is used to describe
the fast variations of the received signal strength due to
changes in phases when a mobile terminal moves. In case of
wideband modeling, each path of the impulse response can be
modeled as Rayleigh distributed with uniform phase except line
of sight (LOS) component cases [17].
C. Received signal model
The transmitted symbols propagating through the radio
channel can be modeled as a circular convolution between the
CIR and the transmitted data block i.e., [s(m)*&>(m,r)] .
Since, the channel coefficient is usually unknown to the
(2) receiver, it needs to be efficiently estimated. The impulse
response of multipath fading channel can be represented by a
tap-delayed line filter with time varying coefficients and
symbol-rate spaced coefficients.
Input
signal
s(m-l)
Delay
s(m-2)
Delay
s(m-3)
Wi(m)-*(X) w 2 (m)— Mj
Delay
s(m-L)
w 3 (m)
-G>
w L (m) — >{X)
(+>-
— -<£h-
Fig. 3. L-tapped delay line filter of a fading channel.
At the receiver, the opposite set of the operation is
performed. After synchronization, CP samples are discarded
and the remaining samples are processed by the DFT to
retrieve the complex constellation symbols transmitted over
the orthogonal sub-channels. The received signals are de-
mapped and equalizer is used to compensate for the radio
channel frequency selectivity. After IDFT operation, these
received signals are demodulated and soft or hard values of
the corresponding bits are passed to the decoder. The decoder
analyzes the structure of received bit pattern and tries to
reconstruct the original signal.
IV. RLS ADAPTIVE CE METHOD
An adaptive CE technique is a process that changes its
parameters as it gain more information of its possibly
changing environment. Among many iterative techniques that
exist in the open literature, the well-liked classes of
approaches which are achieve from the minimization of the
mean square error (MSE) between the output of the adaptive
filter and desired signal to perform CE as shown in Fig. 4.
Transmitted
sequence
AWGN noise
z(m)
s(m)
Multipath channel
model w(m)
Receive signal
AM.
e(m)
_E
Mechanism for adaptive
update of weight coefficients
West(m)
T~
Filter output
y(m)=w est T (m)s(m)
Fig. 4. Scheme for adaptive CE.
The signal s(m) is transmitted via a time-varying channel
w(m), and corrupted by an additive noise estimated by using
any kind of CE method. The main aim of most channel
estimation algorithms is to minimize the MSE i.e., between
the received signal and its estimate, while utilizing as little
163
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
computational resources as possible in the estimation process.
In the Fig. 4, we have unknown multipath fading channel, that
has to be estimated with an adaptive filter whose weight are
updated based on some criterion so that coefficients of
adaptive filter should be as close as possible to the unknown
channel. The RLS CE requires all the past samples of the
input and the desired output is available at each iteration. The
objective function of a RLS CE algorithm is defined as an
exponential weighted sum of errors squares:
c(m) = J r m e H (m)e(m) + SV w H (m)w(m), (5)
m= 1
where 8 is a positive real number called regularization
parameter, e(m) is the prior estimation error, and X is the
exponential forgetting factor with < X < 1. The X is used to
ensure that data in the distant past are paid less concentration
in order to provide the filter with estimating facility when it
operates in a time varying environment. When X = 1, the
algorithm has growing memory because the values of the filter
coefficients are a function of all the precedent input. In this
case, all the values of the error signal, from the time the filter
starts its process to the present, have the same influence on the
cost function. Consequently, the adaptive filter losses its
estimating ability, which is not important if the filter is used in
a stationary environment. In contrast, when < X < 1, the
algorithm has exponentially decaying memory as the recent
values of the observations have greater influence on the
formation of the filter coefficients and tends to forget the old
ones as shown in Fig. 5.
Weighting
factor (a)
8c(m)
8w(m)
= = -2 J r m s(m)e H (m) + 28^ n w(m)
m=l
n
= -2^ r m s(m)[h(m) - w H (m)s(m)] H + 28X n w(m)
w(m)[ J r m s(m)s H (m) + Sri] = £r m s(m)h H (m)
R s (m)w(m) = R sh (m)
w(m)= Rj^nOR^m)
(8)
where R s (m) is the transmitted auto-correlation matrix
R s (m) = Jr m s(m)s H (m) + SVI = XR s (m-l) + s(m)s H (m)
m=l
and R sh (m) is the cross correlation matrix
i.e.,
R sh ( m ) =J k n ' m s(m)h H (m) = ^R sh (m-1) + s(m)h H (m) .
m= 1
According to the Woodbury identity , the above R sh (m) can
be written as
a
.0
a
a
i
i
/ /
/ /
/ /
/ /
/ '
/ 1
/ 1
' 1
^
/
/
/
\ «
, — 1 *►
(m-m) (m-1) ( m -0) (ra + 1)
Sample observation time
Fig. 5. Exponential weighting of observations at different time index.
The prior estimation error is the difference between the
desired response and estimation signal:
e(m) = h(m) - w H (m) s(m) (7)
The objective function is minimized by taking the partial
derivatives with respect to w(n) and setting the results equal to
zero.
Ri(m) = X- 1 Ri(m-l)-
^- 2 R;l(m-l)s(m)s H (m)R;l(m-l)
l+XV^RiCm-lJsCni)
(9)
For convenience of computing, let D(m) = R S h(m) and
K(m) = ^W (10)
l+}i 1 s H (m)D(m-l)s(m)
The K(m) is referred as a gain matrix. We may rewrite (9)
as:
D(m) = ^DCm-l) - r 1 K(m)s H (m)D(m-l) (11)
So simply (10) to
K(m) = D(m)s(m) = R^(m)s(m)
(12)
Substituting (11), (12) into (8), we obtain the following RLS
CE formula
w(m) = w(m-l) + K(m)[h(m) - w H (m-l)s(m)] H
= w(m-l) + K(m)s H (m), (13)
where e(m) is a prior estimation error as
e(m) = h(m) - w H (m-l)s(m) (14)
Therefore equation (13) is the recursive RLS CE algorithm to
update channel coefficient.
164
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
V. PERFORMANCE ANALYSIS
A. Complexity Analysis
The complexity of CE algorithm is of vital importance
especially for time varying wireless communication channels,
where it has to be performed periodically or even
continuously. Table I summarizes the computational
complexity of the RLS CE technique, where L is the channel
length, and real number indicates scalar operation. Here we
assume that each iteration requires the evaluation of the inner
product D(m)h(m) between two vectors of size L each. This
calculation requires L multiplications and L-l additions. Also
assumed that the evaluation of the scalar addition or
subtraction needs one real addition and multiplying the scalar
by the vector requires L multiplications.
TABLE I
COMPLEXITY PER ITERATION
versus SNR for the RLS CE method with different Doppler
frequencies 500Hz and 1.5kHz. One can observed that the
RLS CE method with 500 Hz Doppler frequency has 3 dB
better performances compared with 1.5kHz Doppler
frequency as desired. This CE scheme uses adaptive RLS
estimator which is able to update parameters of the estimator
continuously, so that knowledge of channel and noise
statistics are not required. The similar behavior can be
observed for BER performance in Fig. 7.
Operation
Complexity
Division
1
Multiplication
L 2 + 5L+1
Addition
L 2 + 3L
B. Experimental results
The error performance of the aforementioned iterative
estimation algorithm is explored by performing extensive
computer simulations. All simulation parameters of the LTE
SC-FDMA system in Doppler spread environments are
summarized in Table II.
1
1
□ For Doppler frequency lbuui-iz :
• For Doppler frequency 500Hz ~
\
= = = "^= S : =: q: : : : ;
X x j~" ~i~
-^s< +
^\ 4^
j _ _ X=X- - - - -
= = = = = 1 = = = =^Px= = =
^\^~
1 ^X \L
K
-^X
1 1
X_ ^\
- %V
\
10 15 20
SNR [dB]
25
30
Fig. 6. MSE performance comparisons of the LMS CE method.
Table II
THE SYSTEMS PARAMETERS FOR SIMULATION
Systems parameters
Assumptions
System bandwidth
5 MHz
Sampling frequency
7.68 MHz
Sub-carrier spacing
9.765 kHz
Modulation data type
BPSK
FFT size
16
Sub-carrier mapping scheme
IFDMA
IFFT size
512
Data block size
32
Cyclic prefix
4us
Channel
Rayleigh fading
Forgetting factor
0.99
Equalization
ZF
Doppler frequency
100, and 1.5 kHz
&— For Doppler frequency 1500Hz
For Doppler frequency 500Hz
In practice, the perfect channel coefficient is unavailable,
so estimated channel coefficient must be used instead. The
more correct estimated channel coefficient is, the better MSE
performance of the CE will achieve. Fig. 6 shows the MSE
10 15
SNR [dB]
Fig.7. BER performance comparisons of the LMS CE method.
165
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
VI. CONCLUSION
In this paper, we explore the performance of RLS CE
method for LTE SC-FDAM wireless communication systems
with different Doppler frequencies. The complexities, MSE
and BER performance of the RLS CE method, are analyzed
and compared with the different Doppler frequencies. We can
come to the conclusion that the RLS CE method with 500 Hz
Doppler frequency has 3 dB superior performances compared
with 1.5 kHz Doppler frequency.
REFERENCES
[I] B. Karakaya, H.Arslan, and H. A. Cirpan, "Channel estimation for
LTE uplink in high Doppler spread," Proc. Int. Con. on WCNC,
pp. 1126-1130, April 2008.
[2] J. Berkmann, C. Carbonelli, F.Dietrich, C. Drewes, and W. Xu,
"On 3G LTE terminal implementation standard, algorithms,
complexities and challenges," Proc. Int. Con. on WCMC , pp. 970-
975, August 2008.
[3] A. Ancora, C. Bona, and D.T.M. Slock, "Down-sampled impulse
response least-squares channel estimation for LTE OFDMA,"
Proc. Int. Con. ASSP, Vol. 3, pp. 293-296, April 2007.
[4] L. A. M. R. D. Temino, C. N. I Manchon, C. Rom, T. B. Sorensen,
and P. Mogensen, "Iterative channel estimation with robust wiener
filtering in LTE downlink," Proc. Int. Con. on VTC, pp. 1-5,
September 2008.
[5] S. Y. Park, Y.Gu. Kim, and C. Gu. Kang, "Iterative receiver for
joint detection and channel estimation in OFDM systems under
mobile radio channels, " IEEE Trans. On Comm., vol. 53, ilssue 2,
pp. 450-460, March 2004.
[6] S. Haykin, "Adaptive Filter Theory," Prentice-Hall International
Inc, 1996.
[7] J. J. V. D. Beek, O. E. M. Sandell, S. K. Wilsony, and P. O.
Baorjesson, "On channel estimation in OFDM systems," Proc. Int.
Con. on VTC, vol. 2, pp. 815-819, July 1995.
[8] O. Edfors, M. Sandell, J. V. D. Beek, and S. Wilson, "OFDM
channel estimation by singular value decomposition," IEEE Trans.
on Comm., vol. 46, no. 7, pp. 931-939, July 1998.
[9] M.H. Hsieh, and C.H. Wei, "Channel estimation for OFDM
systems based on comb-type pilot arrangement in frequency
selective fading channels," IEEE Trans, on Consumer Electronics,
vol. 44, issue 1, pp. 217-225, February 1998.
[10] P. Hoeher, S. Kaiser, and P. Robertson, "Two-dimensional pilot
symbol aided channel estimation by wiener filtering," Proc. Int.
Con. on ASSP, pp. 1845-1848, vol.3, April 1997.
[II] M. M. Rana, J. Kim, and W. K. Cho,"Low complexity downlink
channel estimation for LTE systems," in Proc. Int. Con. On
Advanced Commun. Technology, February 2010, pp. 1198-1202.
[12] M. M. Rana, J. Kim, and W. K. Cho," Performance Analysis of
Sub-carrier Mapping in LTE Uplink Systems," in Proc. Int. Con.
On COIN, August 2010.
[13] F. Adachi, H. Tomeba, and K. Takeda, "Frequency-domain
equalization for broadband single-carrier multiple access," IEICE
Trans, on Commun., vol. E92-B, no. 5, pp. 1441-1456, May 2009.
[14] S. Yameogo, J. Palicot, and L. Cariou, "Blind time domain
equalization of SC-FDMA signal," in Proc. Vehicular Technology
Conference, pp. 1-4, September 2009.
[15] S. H. Han and J. H. Lee, "An overview of peak to average power
ratio reduction techniques for multicarrier transmission," IEEE
Trans, on Wireless Commun., vol. 12, no. 2, pp. 56-65, 2005.
[16] H. G. Myung, J. Lim, and D. J. Goodman, "Peak-to-average power
ratio of sngle carrier FDMA signals with pulse shaping," in Proc.
Personal, Indoor and Mobile Radio Commun., September 2006.
[17] W. C. Jakes, Ed., Microwave mobile communications. New York:
Wiley-IEEE Press, Jan. 1994.
1 66 http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
Priority Based Congestion Control for
Multimedia Traffic In 3G Networks
Neetu Sharma 1 , Amit Sharma 2 , V.S Rathore 3 , Durgesh Kumar Mishra 4
1 3 Department of Computer Engineering, Rajasthan, India
l3 Rajasthan College of Engineering for women, Rajasthan, India
Shri Balagi College of Engineering & Technology, Rajasthan, India
Acropolis Institute of Technology and Research, Indore, MP, India
ABSTRACT- There is a growing demand for efficient
multimedia streaming applications over the Internet and next
generation mobile networks. Multimedia streaming services are
receiving considerable interest in the mobile network business.
As communication technology is being developed, the user
demand for multimedia services raises. The third generation
(3G) mobile systems are designed to further enhance the
communication by providing high data rates of the order of 2
Mbps. High Speed Downlink Packet Access (HSDPA) is an
enhancement to 3G networks that supports data rates of several
Mbit/s, making it suitable for applications like multimedia, in
addition to traditional services like voice call. Services like
person- to- person two way video calls or one way video calls, aim
to improve person- to- person communication. Entertainment
services like gaming, video streaming of a movie, movie trailers
or video clips are also supported in 3G. Many more of such
services are possible due to the augmented data rates supported
by the 3G networks and because of the support for Quality of
Service (QoS) differentiation in order to efficiently deliver
required quality for different types of services.
This paper present congestion control schemes that are suitable
for multimedia flows. The problem is that packet losses, during
bad radio conditions in 3G, not only degrade the multimedia
quality, but render the current congestion control algorithms as
inefficient. This paper proposed a solution that integrated the
congestion control schemes with a priority based multimedia
packets to increase the speed of multimedia data and reduce the
packet loss that is developed due to congestion in networks
Key words: UMTS, CN, BS, TFMCC, UTRAN, RNC
I.
INTRODUCTION
The emerging multimedia application requires a fresh approach for
congestion control. A widely popular congestion control schemes are
TCP friendly rate control (TFRC) and Adaptive increase
multiplicative decrease (AIMD) used in networks. These algorithms
used for multimedia traffic but not much effective in packet loss.
TCP is the dominant transport protocol in the Internet, and the
current stability of the Internet depends on its end-to-end congestion
control, which uses an Additive Increase Multiplicative Decrease
(AIMD) algorithm. End-to-end congestion control of best-effort
traffic is required to avoid the congestion collapse of the global
Internet [11]. While TCP congestion control is appropriate for
applications such as bulk data transfer, some real-time applications
(that is, where the data is being played out in real-time) find halving
the sending rate in response to a single congestion indication to be
unnecessarily severe, For providing a better congestion control with
higher data rates a new effective scheme is used. Congestion control
is an important issue in both wired and wireless streaming
applications. Multimedia applications should use some form of
congestion control, both in wired and cellular networks, in order to
adapt the sending rate to the available bandwidth. Today's Internet
stability is due to TCP and its congestion control algorithm. TCP
represents a very efficient transport protocol in general and is
suitable for data transfer. However, it has been argued [13] that TCP
is unsuitable for video streaming because strict delay and jitter
requirements of video streaming are not respected by TCP.
Moreover, some TCP retransmissions are unnecessary for video
when data may miss the arrival deadline and become obsolete. This
has led researchers to look for alternative options. Most of the work
related to congestion control for video flows has either emulated TCP
or has used the TCP model. The well-known TCP-Friendly Rate
Control (TFRC) congestion control consists in an equation based rate
control mechanism [13] [14] [15], designed to keep a relatively steady
sending rate while still being responsive to congestion. When used
over wireless links, TFRC and TCP cannot distinguish between the
wireless losses and the congestion losses. They both may suffer from
the link underutilization if the connection traverses a wireless link.
167
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
This is because they consider dropped packets as a sure sign of
congestion and reduce the ending rate significantly. The inability to
identify a wireless loss followed by unnecessary reduction in sending
rate results in link underutilization.
A. UMTS Introduction
Universal Mobile Telecommunications System (UMTS) is a third-
generation (3G), wireless cellular network that uses Wideband Code
Division Multiple Access (WCDMA) as its radio interface
technology. UMTS offers higher data rates with respect to older 2G
and 2.5G networks and, with the Release 5 version, is evolving into
an all-IP, wireless network. The increased bandwidth provided by
UMTS allows for the deployment of a wide range of services, like
voice, data and multimedia streaming services. In wireless networks,
congestion control, alone, may not be enough to ensure good quality
of multimedia streaming and efficient utilization of the network.
Packet losses due to the high bit error rate not only degrade the
multimedia quality, but render the current congestion control
algorithms as inefficient: these algorithms back-off on every packet
loss even when there is no congestion. We integrate the congestion
control schemes with an adaptive retransmission scheme in order to
selectively retransmit some lost multimedia packets. Fig.l shows the
transmission of multimedia data over a wireless channel.
■ Overview
How to improve QoS & resource utilization ?
Network
■ Video Streaming over 3G
•Header Compression
User
How can applications adapt :
improve the multimedia quality ?
• Congestion Control for Video Flows
Fig. 1 Transmission of Multimedia data
B. GENERAL UMTS NETWORK:
UMTS, the successor of GSM, is evolving toward a future wireless
all-IP network. In this paper we present how it supports real-time IP
multimedia services, as these services are expected to drive the
adoption of wireless all-IP networks.
UMTS networking architecture is organized in two domains. The
user equipment (UE) and the public land mobile network (PLMN).
The UE is used by the subscriber to access the UMTS services.
PLAN is further divided into two land-based infrastructures
(i) UTRAN (UMTS terrestrial radio- access network)
(ii) CN (core network).
The UTRAN handless all radio-related functionalities and the CN is
responsible for maintaining subscribes data and for connections.
UTRAN contain two types of nodes Radio network controller (RNC)
and Node B. Node B is the base station and provides radio coverage
to one or more cells. Node B connected to UE via Uu interface and to
the RNC via Iub interface. Uu is a radio interface based on the
wideband code division multiple access (WCDMA) technology [7].
The CN consist by two types of general packet radio service support
nodes (GSNs). That is gateway GSN (GGSN) and serving GSN
(SGSN). SGSN provide the routing functionality. It manages a group
of RNCs and interacts with the home location register (HLR). HLR
permanently store the subscriber data. SGSN connected to GGSN via
the Gn interface. RNC connect to SGSN via Iu interface. Through the
GGSN the UNTS network connect to external packet data network
like the internet.
/Tub lub jfS|
Fig.2 General UMTS Network
C. 3G/UMTS Problems
• Problems due to the use of IP
o IP doesn't support real time streaming
requirements
o Overhead due to packet header
• Problems due to radio conditions
o Scarce and time varying bandwidth
o Congestion, wireless losses & large delay
168
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
D. UMTS QoS Classes
UMTS defines four QoS classes [2] and the classified traffic gets the
treatment inside the UMTS network according to its class. The four
QoS classes are:
• Conversational class: The traffic from the applications like person-
to-person video or voice call is classified into conversational class.
The delay and jitter requirements for this type of traffic are very
strict. This is because on the both end points there is a human
expecting the delivery of the voice and/or video data in very short
time after it is sent.
• Streaming class: Video on Demand (VoD) falls under this class.
The delay requirements are there but are not as strict as the
conversational class.
• Interactive class: The interactive traffic like interactive e-mail or
web browsing falls under this category. Though there is still some
delay requirement, it is less strict than the conversational and
streaming classes. Moreover, since the traffic mostly pertains to data
applications, the bit error rate should be very low.
• Background class: This class is the most insensitive to delay. It
includes the traffic from background applications like background
email and SMS. Though, the bit error rate, like the Interactive class,
should be very low.
D Congestion Control for Multimedia data
TCP-Friendly Rate Control (TFRC) is an end-to-end congestion
control mechanism, whose goal is to provide rate control for unicast
flows in IP networks. The main feature of TFRC is its ability to
smoothly adapt the sending rate of a flow to network conditions,
while competing for bandwidth with TCP flows in a relatively fair
manner. TFRC was designed to offer a more stable sending rate than
TCP on wired, best-effort networks, making it suitable for
applications like multimedia streaming. We evaluate the performance
of TFRC, compare it with that of TCP and new TFRC for different
multimedia classes, under different scenarios of MAC -layer
scheduling, radio conditions and background traffic.
TFRC [4] [10] is an end-to-end congestion control mechanism
suitable for applications with constraints on rate stability, like voice
or streaming media. It has been designed to adapt the sending rate of
a flow in a smooth manner, while trying to fairly share the available
bandwidth with competing TCP flows. TFRC is an Internet standard
[4], and it has been adopted at the IETF as one of the congestion
control profiles that may be used with the DCCP transport protocol
[10]; TFRC may also be implemented by UDP -based applications
wishing to perform congestion control. This paper presents a
simulation study of TFRC over UMTS networks supporting
HSDPA. Since we are interested in video streaming applications, we
evaluate the performance of TFRC in terms of rate stability over
different time scales, and compare it with that of TCP. Several
scenarios of MAC-layer scheduling, radio conditions and background
traffic are considered.
This paper proposed a more reliable algorithm that provides
congestion control for different multimedia classes. Priority assigned
to each of the packet according to multimedia classes. So whenever
the congestion occurs in the network the lowest priority packets are
dropped. If overall loss rate for lower priority packets is not very
high, then we can safely assume that the congestion loss rate for the
highest priority packets will be insignificant. In such a case, the loss
of highest priority packets will be mainly due to wireless errors.
Thus, it is to be expected that, in general, there is a good correlation
between wireless packet loss rate and the total loss rate of highest
priority packets.
2. THE PROPOSED SCHEME
This paper provides a mechanism of congestion control for the
multimedia transmission over UMTS. We analyze TCP friendly
multicast congestion control (TFMCC) over UMTS and generalize it
to different multimedia classes [5] [6]. We design a novel mechanism
for congestion control that is Content Sensitive TCP Friendly
Multicast Congestion Control (CSTFMCC). We perform a little
modification in UMTS network and the packet field. At various level
of network we provide the control mechanism that prevents the
network from the congestion. Multimedia Class I traffic includes
video and audio traffic from users equipped with an adjustable rate.
Class II traffic includes non-real time data traffic such as e-mail, file
transfer and web browsing traffic. These two classes contain different
multimedia traffic that is more delay sensitive or less delay sensitive.
So class I traffic support the real time applications and more delay
sensitive. Due to congestion, if any loss of the packet or the delay
between the packets can reduce the quality of received video/audio.
Whether in class II traffic, if congestion occur it is acceptable to
buffer non-real time data at a network node or at the user station and
transmit them at a slower rate. In a large multicast group, there will
usually be at least one receiver that has experienced a recent packet
169
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
loss. If the congestion control mechanisms require that the sender
reduces its sending rate in response to each loss, as in TCP, then
there is little potential for the construction of scalable multicast
congestion control.
In wireless communication systems like UMTS, the packet loss may
not mean network congestion. The quality of wireless link may be
degrading due to signal fading. During a fading period, the bit error
rate of wireless link may become very high but after that period the
wireless link is expected to recover. TFMCC uses a feed back
scheme which allows the receiver to calculating the slowest
transmission rate to always reach the sender.
A. Sender end: Fragmentation of data packets perform at the sender
end. The sender fragment the data packet with on bit of priority.
There are two parts of the data (i) packet header and (ii) payload. The
size of header part change by one bit shows the priority of packet. So
there is only one bit modification perform in the size of data packet
and it increases the speed of multimedia packets.
If (incoming request for higher priority packets)
If (there is a free channel) then
allocate the free channel
else
If (lower priority packets)
put in a buffer
If (there is free channel again)
allocate the free channel to lower
priority packets
else
Ignore request
endif
else
Ignore request
endif
endif
End.
ignore request
endif
Fig.3 Algorithm For proposed model
Fig. 4 shows the flow chart for the proposed model. Flow chart
shows the arrival of packet and priority check by the routers at
layer2. According to this priority the packets is being processed.
B. Multimedia Packet Size: Multimedia packet size depends on the
multimedia classes. The proposed scheme redesigns the multimedia
packets. It increases the multimedia packet size by one bit. This bit
shows the priority of multimedia packets. The highest priority
packets serve first by the routers at the layer 2. So the size of the
packet is increased by one bit.
C. Routing Scenario: For fast transmission of multimedia
information the proposed scheme give the priority to all multimedia
packets. When a user want to send multimedia data the data framing
perform at the sender end. The sender constructs the frame with a
priority bit. This information stored in the header of the packet for
priority access to the router. Sender sends the packets towards its
destination. Multimedia packets reach at the network. At layer two
the router checks the destination address and priority bit of the
packet. If a higher priority packet arrives then router serves first to
the packet which contains a highest priority. This increases the speed
of multimedia packets and decreases the congestion in the networks.
For implementing this scenario the changes perform in the size of
packets and in routers. Following algorithm shows the scenario for
routing the various packets according to priority. Fig. 3 shows the
algorithm for the proposed model.
^
r
i
i
k
Channel? ^^-i£2
> 1
No
ir
assign' c tonne I* ^
pack".'" 6
Accept packet and
Fig. 4 Flow Chart for the proposed model
Receiving End: At the receiving end defragmentation perform. The
receiving data packet reaches at the destination and multimedia
information is available for the user respectively
A. Simulation Platform
The simulation that we use for this is EURANE (NS-2
Extension) [16]. Following fig. 3 shows the simulation topology to
increase the multimedia quality. This paper focuses on the problem
of evaluating the subjective video quality and presents the quality
estimation tool that we employed, a performance evaluation study
done with the well known ns-2 network simulator
170
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
I I
I I UMTS
IKTTERNET I jJOREJ4ETWORK_
§]j^X^v a;
g
Fig:3 Simulation Topology
the video packet trace file is fed to the ns-2 simulator (compiled with
the EURANE extensions). This trace file serves as a traffic generator
during the simulation. A simulation script allows defining the
particular scenario under consideration (network topology,
simulation parameters, and so on). When the simulation is run, an
output trace file is produced which contains delay- and loss-related
information for every video packet sent by the (simulated) video
server.
Conclusion
This paper represents to design a multicast group congestion control
mechanism over the UMTS networks. This mechanism is content
sensitive and optimized for multimedia traffic. This paper performs
some changes in existing UMTS architecture and the packet field of
data packet. User equipment (UE) has some additional function to
detect the multimedia class type. The additional functionality of UE
and the field of data packet has main target to remove the packet loss
and congestion.
Refrences
[1] Holma. H. and Toskala, A. "WCDMA for UMTS: Radio
Access for Third Generation Mobile Communications"
(3 rd Edition),Wiley,2004.
[2] Antonios Alexiou, Dimitrios Antonellis and Chritos
Bouras" A Study of Multicast Congestion Control for
UMTS" proc Int. J. Commun. Sys. (2009)
[3] Miiighua Clien and Avitleh Zaklior "Rate Control for
Streaming Video over wireless "proc of IEEE
INFOCOM 2004 Hong Kong China
[4] J. Widmer, R. Denda, M. Mauve "A Survey on TCP-
Friendly Congestion Control" IEEE Network, 15 "(3),
May- June 2001
[5] Cui-Quing Yang and Alapati V.S. Reddy "A Taxonomy
for congestion control algorithms in packet switching
networks"proc IEEE 1995
[6] Ljiljana Trajkovic and S. Jamaloddin Golestant"
Congestion Control for Multimedia Services" proc of
IEEE INFOCOM 1992
[7] Antonios Alexiou, Dimitrios Antonellis and Chritos Bouras
"Equation Based Congestion Control for Video
Transmission Over WCDMA Networks"proc of IEEE
AINA'06 Vienna, Austria, pp .445-450.
[8] Minghua Chen and Avideh Zakhor "Rate Control for
Streaming Video over Wireless" proc of IEEE
INFOCOM 2004
[10] Kamal Deep Singh*, Arpad Huszak., David Ros§, Cesar
Viho* and Gabor Jeney.
"Congestion Control and Adaptive Retransmission for
Multimedia Streaming over Wireless Networks" The
Second International Conference on Next Generation
Mobile Applications, Services, and Technologies
[11] Kamal Deep Singh ^, Julio Orozcof, David Ros/ and
Gerardo Rubino "Improving perceived streaming-video
quality in High Speed Downlink Packet Access" proc of
IEEE IEEE "GLOBECOM" 2008
[12] S. Floyd and K. Fall. Promoting the Use of End-to-end
Congestion Control in the Internet. IEEE/ACM
Transactions on Networking, Aug. 1999.
[13] R. Jain, K. Ramakrishnan, and D. Chiu. Congestion
Avoidance in Computer Networks with a
Connectionless Network Layer. Tech. Rep. DEC-TR-
506, Digital Equipment Corporation, August 1987.
[14] S. Floyd, M. Handley, J. Padhye, and J. Widmer .Equation
based congestion control for unicast applications., In
Proceedings of ACM SIGCOMM 2000, pages 43.56,
Stockholm, Aug. 2000.
[15] S. Floyd, E. Kohler, and J. Padhye, .Profile for Datagram
Congestion Control Protocol (DCCP) Congestion Control
ID 3: TCP-Friendly Rate Control (TFRC)., RFC 4342,
IETF, Mar. 2006.
[16] M. Handley, S. Floyd, J. Padhye, and J. Widmer. TCP
Friendly Rate Control (TFRC): Protocol Specification.
Internet Standards Track RFC 3448, IETF, Jan. 2003.
[16] Ns-2 . Network Simulator,
http:///www.isi.edu/nsnam/ns/index.html
171
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
Authors Profile
Ms. Neetu Sharma, Reader
Biography: Mrs. Neetu Sharma obtained her
Engineering degree from University of Rajasthan
and Masters Degree from Rajasthan Vidyapeeth,
Udiapur securing First division with honors in both.
Currently she is pursuing Ph.D. (CSE) in
Congestion Control in 3G from Gyanvihar
University, Jaipur, India. She has been Reader and
HOD of the department of CSE at Rajasthan
College of Engineering for Women, Jaipur, India.
She has extensively worked in various field of
Computer Engineering. She has published many
national papers in the reputed journals and
conferences. She is an author of the book 'System
Software Engineering' for B.Tech. students. She is
the member of renowned societies like IEEE, IEEE
computer society, ISTE and CSI also.
Dr. Vijay Rathore, Associate Professor
O
3
Biography: Dr. Vijay Singh Rahore obtained his
MCA and Ph.D. (CSE) from University of
Rajasthan, India. He is an Associate Professor,
Shree Kami College, Jaipur, India. He has more
than 10 years of industrial and teaching experience.
His areas of interest are Computer Organization &
Architecture, Operating System Fundamentals,
DBMS & RDBMS (Oracle, DB2, SQL Server, MS-
Access, DBASE, etc.), Data Structures,
Programming Languages (C, C++, Java (J2SE,
J2ME, J2EE), VB, COBOL), Networking
Technologies (Data Communications, Internet &
Intranet, E-Commerce, Network Security,
Cryptology etc.), Software Engineering, System
Analysis & Design, Management Information
System, Decision Support System, Artificial
Intelligence, E-Governance, Computer Center
Management, UNIX, etc.. He is the member of
renowned society like ISTE.
Dr. Durgesh Kumar Mishra
Professor (CSE) and Dean (R&D),
Acropolis Institute of Technology and Research,
Indore, MP, India,
Ph - +91 9826047547, +91-731-4730038
Email: durgeshmishra@ ieee.org
Chairman IEEE Computer Society, Bombay Chapter
Vice Chairman IEEE MP Subsection
Biography: Dr. Durgesh Kumar Mishra has
received M.Tech. degree in Computer Science from
DAW, Indore in 1994 and PhD degree in
Computer Engineering in 2008. Presently he is
working as Professor (CSE) and Dean (R&D) in
Acropolis Institute of Technology and Research,
Indore, MP, India. He is having around 21 Yrs of
teaching experience and more than 7 Yrs of
research experience. He has completed his research
work with Dr. M. Chandwani, Director, IET-DAVV
Indore, MP, India in Secure Multi- Party
Computation. He has published more than 60
papers in refereed International/National Journal
and Conference including IEEE, ACM etc. He is a
Senior Member of IEEE, Chairman of IEEE
172
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
Computer Society, Bombay Chapter, India. Dr.
Mishra has delivered his tutorials in IEEE
International conferences in India as well as other
countries also. He is also the programme committee
member of several International conferences. He
visited and delivered his invited talk in Taiwan,
Bangladesh, Nepal, Malaysia, Bali-Indonesia,
Singapore, Sri Lanka, USA and UK etc in Secure
Multi-Party Computation of Information Security.
He is an author of one book also. He is also the
reviewer of tree International Journal of
Information Security. He is a Chief Editor of
Journal of Technology and Engineering Sciences.
He has been a consultant to industries and
Government organization like Sale tax and Labor
Department of Government of Madhya Pradesh,
India.
Mr. Amit Sharma, Assistant Professor
Biography: Mr. Amit sharma obtained his MCA
from University of Rajasthan and aboout to
complete his M.Tech.(CSE) from Rajasthan
Technical University. He is an Assistant Professor
in Sri Balaji College of Engineering & Technology,
Jaipur. He has more than 5 years of industrial and
teaching experience. His areas of interest are Open
Source, Networking, Advance Computer
Architecture and Information Security. He is the
member of renowned society like ISTE.
1 73 http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
Adaptive Sub-block ARQ techniques for wireless networks.
A.N.Kemkar, Member, ISTE and Dr. T.R.Sontakke Member,ISTE
Abstract — Wireless channels are highly affected
by unpredictable factors such as co channel
interference, adjacent channel interference,
propagation path loss, and shadowing and multi
path fading. An adaptive ARQ scheme, we mean
an ARQ scheme with two or more different
transmission modes meant for different channel
conditions, which uses some channel sensing
mechanism to decide which transmission mode is
used. In this paper, we propose an adaptive
transmission of sub-blocks scheme, for wireless
networks. As the channel becomes increasingly
noisy, the data block is divided into smaller sub-
blocks for transmission. Each sub-block is
encoded for error control by a CRC code. The
received block is checked for errors sub-block by
sub-block. The propose sub-block retransmission
scheme provides improved throughput over
conventional ARQ schemes by retransmitting
only the same number of sub-blocks in the
occurrence of errors. .
Index Terms — Retransmission protocol; Hybrid
ARQ,CRC
1. INTRODUCTION:
In a mobile radio channel, burst errors frequently
occur because of Rayleigh fading and shadowing.
In particular, for a large cell- size system with the
radius of more than several km, shadowing often
becomes the predominant source of burst errors.
(Shown in Fig- 1)
Therefore, in order to provide reliable packet data
transmission in such a channel, an efficient
automatic repeat request (ARQ) protocol must be
employed, since data service can tolerate some
□ A.N.Kemkar 1 ,S.R.T.M.U,Nanded.
+91-9819150392, ankemkar@gmaiLcom
Dr.T.R.Sontakke 2
Ex.Director - S.G.G.S.I.T.E.- Nanded
Principal,Sidhant college of Engineering
Pune.+91 -98223 92766,trsontakke@gmail.com
delays but is sensitive to loss and errors. Many
researchers have devoted much effort to analyze
the throughput for various ARQ protocols in
Rayleigh fading channels. In the packet data
transmission, short packets are less likely to
encounter fades than long packets, but on the
other hand, they are more burdened by overheads.
In other words, the packet length to maximize the
instantaneous throughput is closely related to the
dynamic channel condition due to fading,
shadowing, and propagation path loss. Therefore,
if we choose the optimum packet length
adaptively by estimating the channel condition,
we can continuously achieve the maximum
attainable throughput.
Paper is organized as follows. We start by,
describes in detail about related work in section 2.
A Communication system model and proposed
method of adaptation in section 3. In section 4
we have presented system analysis 5. Simulation
parameters and results . Followed by conclusion
in section 6.
2. RELATED WORK:
A change of transmission mode can mean, for
example, a change of the packet size in the SR
scheme [1] or a change of the number of
transmitted copies of a packet in the GBN scheme
[2] or a change of the code rate in an HARQ-I
scheme [3]. In these schemes, the channel sensing
is usually done by observing the
acknowledgements sent by the receiver to the
transmitter. This can mean either estimation of
error rates, as in [4], or detection of channel state
changes, as in [5] and [6], which does not require
as long an observation interval (OBI) as reliable
error rate estimation.
In [2], an adaptive SR scheme was proposed,
where the packet size used in the current
transmission was selected from a finite set of
values based on a long-term BER estimate. This
estimate was obtained by counting the incorrectly
received packets over a time interval and
174
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
assuming that there can be at most one bit error in
an erroneous packet. Another adaptive SR
scheme with variable packet size was proposed in
[5], where the a posteriori distribution of the
BER was computed based on the number of
retransmissions during the OBI, and the packet
size was selected so that the expected efficiency
of the protocol was maximized. In [2], Yao
proposed an adaptive GBN scheme where the
transmitted number the transmitted number of
copies of a packet was variable.
Numerous adaptive HARQ schemes have been
suggested in the literature. Typically, the code
rate is varied according to the estimated channel
conditions. In [6] and [7], adaptive HARQ-I
schemes were studied with convolutional codes
used for error correction. Finite-state Markov
models were assumed for the channel. Switching
between transmission modes depended on the
number of erroneous blocks occurring during an
OBI. A similar adaptive HARQ-I scheme with
either block or convolutional codes were
proposed in [3]. In [4], sequential statistical tests
were applied on the acknowledgements to detect
channel state changes. An adaptive HARQ-II
scheme with variable packet size was proposed
for wireless ATM networks in [8]. This scheme
used rate compatible convolutional (RCC) codes
for error correction. In [9], three different
adaptive HARQ schemes are proposed using
Reed- Solomon codes for error correction.
Another adaptive HARQ scheme using Reed-
Solomon codes with variable rate for error
correction was proposed in [10]. In this scheme
short-term symbol error rate was estimated by
computing the bitwise modulo-2 sum of two
erroneous copies of a packet. This method was
originally proposed in [1 1].
3. A COMMUNICATION SYSTEM MODEL
AND PROPOSED METHOD OF
ADAPTATION:
3.1 A communication system model:
Fig. 1 shows the communication system model.
In the non cellular or large cell- size system, a
radio base station continuously transmits data
packets to a single mobile terminal with no
packet collision after the link connection is
established. Table I summarizes the digital
mobile communication system characteristics,
where we choose a binary frequency shift keying
non coherent detection (non coherent FSK)
scheme in terms of easy implementation, because
it requires no complicated carrier recovery ci rcuit .
v
► Buffer
Transmitter
Electronics
Adaptive Logic
Buffer
v
Receiver
Electronics
Fig.l. Communication System model.
Table 1 -Communication system parameters
Parameters
Description
Channel type
Contention Free, half duplex
Modulation
/Demodulation
Binary FSK,Non coherent
detection.
Packet structure
Information packet length
256,128,64,32,16 bits
3.2. Adaptation Policy
According to the variations of SNR, the receiver
channel may be consider to be in one of the states
at each instant 't\ We assume that the sender
knows the state at the transmission time for
receiver. Let's define the transmission status at
time t as the set of all channel state. Before
transmitting, the adaptive algorithm in the sender
must estimate the efficiency and packet loss rate
of the ARQ/FEC protocol using all the available
coding schemes as well as the ARQ protocol as a
function of the transmission status. It then tries to
find the protocol satisfying the desired packet loss
rate. The time is divided into transmission rounds.
Each transmission round corresponds to the
transmission c k' packets in case of ARQ. A
transmission round ends when the sender is
informed about the reception state of the receiver.
The adaptive algorithm is repeated at the end of
175
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
each transmission round. Basically, the algorithm
goes through the following steps:
1 . At the beginning of the algorithm, the sender
determines the desired packet loss rate (SNR) of
the session. It also determines the transmission
status.
2. The sender estimates the packet loss rate of the
ARQ protocol as well as the ARQ/FEC protocol
using all the available coding schemes, based on
the transmission status. It then adjusts its
parameters and starts the transmission of the
blocks.
3. At the end of a transmission round, the sender
again determines the transmission status. It then
repeats the step.
4. PERFORMANCE ANALYSIS OF THE
PROPOSED SCHEME:
The performance analysis of the scheme is
measured in terms of throughput of the proposed
scheme. Further we show the comparison of
throughput with sub block and without sub block
transmission schemes with Adaptive scheme.
Expression of throughput for ARQ for present
scheme:
K
V =
E[T]
(1)
Where ^information bits in a block. E[]
=Expectation of number of transmitted bits in a
given block.
oo
T = Mn + Yji (2)
Where M =number of sub blocks, « =number of
bits in a sub block, Ti =number of transmitted bits
for i th transmission.
E[T]=Mn + f d E[T l ] (3)
1=1
Where E[T\ =Average number of transmitted
bits.
Out of M sub blocks if L sub blocks are
transmitted at the f 1 retransmission, then random
variable, Ti takes the value Ln ,if L out of M
sub-blocks are retransmitted at the f 1
retransmission. Our algorithm compute the value
of (3) to get the result from equation (l).In our
analysis we have consider the variable packet size
as a retransmission units with fixed rate of data
transmission. Hence we can send the packet of
specific size based on the estimated signal to
noise ratio (SNR). As shown in the Fig.l the
system configuration of ARQ techniques
combines with adaptive packet size modulation.
With an exact bit error rate equation for FSK at
certain signal to noise ratio '7\ we can decide
the value of packet size satisfying the required
BER (bit error rate) Assuming that we have ' M
'different block sizes { L1L2L3L4 Lm\ .
Let Ai',ie{0 9 M) with ^,-as the threshold
value of signal to noise ratio, being between the
i th level and i + 1 level.
Ao is the lowest possible signal to noise ratio and
Am is the highest possible signal to noise ratio.
5. SIMULATION RESULTS:
We evaluate the performance of the proposed
scheme implemented with Matlab. We run the
simulation for three schemes i.e. with sub block
transmission and without sub block transmission
and adaptive. The simulation parameters are
shown in the table 2.
Table 2- Simulation Parameters
Parameters
Notation
Values
Signal to Noise Ratio
7
Varied
Threshold values of
SNR
snr
8,7,6,5
Max. number of
Retransmissions
3
Number of sub blocks
retransmitted
L
Varied
Cyclic Redundancy
Check
CRC
Varied
Bit error rate
BER
Varied
Packet error rate
PER
Varied
Throughput efficiency
V
Varied
Data Rate
R
9.6kbps
Simulation runs for 5000 total blocks. Result
is the average of independent experiments where
each experiment uses different randomly
generated uniform parameters. We use mean
values which are obtained independent
176
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
experiments as a basic data to get the result.
Simulation results are shown Table 3.
For the packet error rate 0.3,0.5,0.7 the
throughput of the system is say VvVi^Vs-
Adaptive Scheme
parameters
Throughput efficiency
SNR
Sub
block
Packet
size
Vi
v 2
v 2
8
2
256
0.92
0.84
0.82
7
4
128
0.93
0.85
0.83
6
8
64
0.94
0.92
0.88
5
16
32
0.97
0.95
0.93
Below
this
32
16
0.98
0.94
0.96
Table 3.Simulation Results of adaptive
scheme
I h f o i) o \\ p i) t V s . P a c k
Li.
k
i
I
i
9
8 5
1 o jc r
-0-2 S
u 1
4-
-0-t S
u 1 ____B_
1 o jc k
-0-8 S
u 1
1 o |c k
■ N. "X
-0-1 6
S b
B l!o c
k I X\
-0-3 2
S b
B Ijo c
i
k j
U^
i
i
i
10 . 20 . 30 . 40 . 50 . 60 . 70 . 80 . 9 1
P a c k e t E r r o r R a
T h ro u g h p u t V s . P a c k e t E rro r R a
W it h Micro
W ill) o i) t
Adaptive A
R Q
j i i i i i i_
lo c k
o b Ic
Fig. 2. Comparison as per Table -2
.20.30.40.50.60.70.80.9 1
P a c k e t E rro r R ate
Fig. 3. Comparison of three scheme
6.conclusion: Table -2 and Fig.2 shows the
result of five modes of sub block
transmission .Proposed Adaptive scheme
will choose proper mode in accordance with
the channel parameters, here threshold
values of SNR (considered as channel
sensing mechanism).From the Fig. 3 the
simulation results showed that the simulated
throughput points for adaptive scheme
follows the ideal throughput curve(with
micro block transmission) very closely. The
proposed Adaptive sub-block retransmission
scheme improved the throughput and the
reliability by using dynamically adapting the
number of Sub block transmission according
177
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
to length to the varying channel packet error
rate..
REFERENCES :
[1] J.A.C. Martins and J.C. Alves. ARQ
protocols with adaptive block size
perform better over a wide range of
bit error rates. IEEE Transactions on
Communications, 38:73 7-73 9, June
1990.
[2] Y.-D. Yao. An effective go-back-N
ARQ scheme for variable-error-rate
channels. IEEE Transactions on
Communications, 43:20-23, January
1995.
[3] M. Rice and S.B. Wicker. Adaptive
error control for slowly varying
channels. IEEE Transactions on
Communications, 42:917-926,
February/March/April 1994.
[4] M. Rice and S.B. Wicker. A
sequential scheme for adaptive error
control over slowly varying
channels. IEEE Transactions on
Communications, 42: 1 533-1 543,
February/ March/April 1994.
[5] E. Modiano. An adaptive algorithm
for optimizing the packet size used in
wireless ARQ protocols. Wireless
Networks, 5:279-286, July 1999
[6] B. Vucetic. An adaptive coding
scheme for time- varying channels.
IEEE Transactions on
Communications, 39:653-663, May
1991.
[7] B. Vucetic, D. Drajic, and D. Perisic.
Algorithm for adaptive error control
synthesis
[8] I. Joe. A novel adaptive hybrid ARQ
scheme for wireless ATM networks.
Wireless Networks, 6:211-219, July
2000.
[9] S. Choi and K.G. Shin. A class of
adaptive hybrid ARQ schemes for
wireless links. IEEE Transactions on
Vehicular Technology, 50:777-790,
May 2001.
[10] H.Minn,M. Zeng, and V.K.
Bhargava. On ARQ scheme with
adaptive error control. IEEE
Transactions on Vehicular
Technology, 50:1426-1436,
November 2001. IEE Proceedings,
PartF, 135:85-94, February 1988
[11] S.S. Chakraborty, M.
Liinaharja, and E. Yli-Juuti. An
adaptive ARQ scheme with packet
combining for time varying
channels. IEEE Communications
Letters, 3:52-54, February 1999.
178
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol 8, No. 6, September 2010
Trigon Based Authentication Service Creation with
Globus Middleware
Ruckmani V l Anitha Kumari K 2 Sudha Sadasivam G 3
Senior Lecturer , MCA, Lecturer, IT, , Professor ,CSE,
Ramakrishna Engineering College, PSG College of Technology, PSG College of Technology,
Coimbatore, India Coimbatore,India Coimbatore,India
Dhaarini M P 4
Lecturer , IT,
PSG College of Technology
Coimbatore,India .
Abstract — A Grid is built from multi-purpose protocols and
interfaces that address fundamental issues as authentication,
authorization, resource discovery, and resource access.
Security is of utmost importance in grid computing
applications as grid resources are heterogeneous, dynamic, and
multi- domain. Authentication remains as the significant
security challenge in grid environment. The proposed
approach uses a dual authentication protocol in order to
improve the authentication service in grid environment. The
protocol utilizes the fundamental concepts of trigon and based
on the parameters of the trigon the user authentication will be
performed. In the proposed protocol, the password is
interpreted and alienated into more than one unit and these
units are stored in two different servers, namely,
Authentication Server and Backend Server. Only when the
combined authentication scheme from both the servers
authenticates the user, the privilege of accessing the requested
resources is obtained by the user. The main advantage of
utilizing the dual authentication protocol in grid computing is
that an adversary user cannot attain the access privilege by
compromising a single consolidated server because of the fact
that the split password is stored in different servers. Grid
service is stateful and transient web service, which can be
invoked by clients, and is considered to be the mainstream of
future internet. The creation of Web Services standards is an
industry-led initiative, with some of the emerging standards in
various states of progress through the World Wide Web
Consortium (W3C). To achieve reuse of behaviors of this
authentication concept, operations are often grouped together
to form a trigon based authentication service.
Keywords — Trigonbasedauthentication, web services,
globus.
virtual organizations boundaries. A VO is a dynamic group
of individuals, groups, or organizations that have common
rules for resource sharing [8]. Confidentiality of information
in a VO Should also be ensured [28]. The necessity for
secure communication between grid entities has motivated
the development of the Grid Security Infrastructure (GSI).
GSI provides integrity, protection, confidentiality and
authentication for sensitive information transferred over the
network in addition to the facilities to securely traverse the
distinct organizations that are part of collaboration.
Authentication is done by exchanging proxy credentials and
authorization by mapping to a grid map file. Grid
technologies have adopted the use of X.509 identity
certificates to support user authentication. SOAP protocol
[12] is used by the emerging OGSA. This necessitates for
support message layer security using XML digital signature
standard and the XML encryption standard [11]. Globus
Toolkit [24] provides security services for authentication,
authorization, management of user credentials and user
information. Laccetti and G. Schmid [14] have introduced a
unified approach for access control of grid resources. PKI
(Public Key Infrastructure) and PMI (Privilege Management
Infrastructure) infrastructures were utilized at the grid layer
after authentication and authorization procedures.
Czajkowski [5] have explained about agreement based grid
management. Nagaratnam [18] have introduced security
architecture for open grid services. H.-L. Truong[26] define a
framework for monitoring and analyzing qos metrics of grid
Services. The proposed work aims at authenticating the users
by using trigon concept and to host this operation as a web
service.
I. INTRODUCTION
A. Globus Middleware
Grid computing has emerged as a significant new field,
distinguished from conventional distributed computing by its
concentration on large-scale resource sharing, innovative
applications, and, in some cases, high-performance
orientation . Grid computing is concentrating on large-scale
resource sharing and collaboration over enterprises and
Globus [25] provides a software infrastructure that
enables applications to handle distributed heterogeneous
computing resources as a single virtual machine. Globus is
constructed as a layered architecture in which high-level
global services are built upon essential low-level core local
services. Middleware is generally considered to be the layer
179
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
of software sandwiched between the operating system and
applications, providing a variety of services required by an
application to function correctly. Recently, middleware has
re-emerged as a means of integrating software applications
running in distributed heterogeneous environments. In a
Grid, the middleware is used to hide the heterogeneous
nature and provide users and applications with a
homogeneous and seamless environment by providing a set
of standardized interfaces to a variety of services.
B. Web Services
The term Web Services describes an important emerging
distributed computing paradigm that differs from other
approaches such as DCE, CORBA, and Java RMI in its focus
on simple, Internet-based standards to address heterogeneous
distributed computing. Web services define a technique for
describing software components to be accessed, methods for
accessing these components, and discovery methods that
enable the identification of relevant service providers. Once a
web service is created, it is advertised in a registry called
UDDI (Universal Description, Discovery and Integration)
[27], where it can be searched. The UDDI will provide the
location to the service provider's WSDL (Web Services
Description Language) [29] file that describes the methods
that can be invoked and the parameters required. Messages
are exchanged through the protocol SOAP (Simple Object
Access Protocol) [30].
The established standards include:
SOAP (XML Protocol). SOAP provides an envelope
which encapsulates XML data for transfer through the Web
infrastructure (e.g. over HTTP, through caches and proxies),
with a convention for Remote Procedure Calls (RPCs) and a
serialization mechanism based on XML Schema data types.
SOAP is being developed by W3C in cooperation with the
Internet Engineering Task Force (IETF).
Web Services Description Language (WSDL). Describes
a service in XML, using an XML Schema; there is also a
mapping to the Resource Description Framework (RDF). In
some ways WSDL is similar to an interface definition
language IDL. WSDL is available as a W3C note [WSDL].
Universal Description Discovery and Integration (UDDI).
This is a specification for distributed registries of web
services, similar to yellow and white pages services. UDDI
supports 'publish, find and bind': a service provider
describes and publishes the service details to the directory;
service requestors make requests to the registry to find the
providers of a service; the services 'bind' using the technical
details provided by UDDI. It also builds on XML and SOAP
[UDDI].
(IJCSIS) International Journal of Computer Science and Information Security,
Vol 8, No. 6, September 2010
Web Services have certain advantages over other
technologies:
Web Services are platform-independent and language-
independent, since they use standard XML languages. This
means that my client program can be programmed in C++
and running under Windows, while the Web Service is
programmed in Java and running under Linux.
Service Processes: This part of the architecture generally
involves more than one Web service. For example, discovery
belongs in this part of the architecture, since it allows us to
locate one particular service from among a collection of Web
services.
Service Description: One of the most interesting features
of Web Services is that they are self-describing. This means
that, once you've located a Web Service, you can ask it to
'describe itself and tell you what operations it supports and
how to invoke it. This is handled by the Web Services
Description Language (WSDL).
Service Invocation: Invoking a Web Service (and, in
general, any kind of distributed service such as a CORBA
object or an Enterprise Java Bean) involves passing
messages between the client and the server. SOAP (Simple
Object Access Protocol) specifies how we should format
requests to the server, and how the server should format its
responses. In theory, we could use other service invocation
languages (such as XML-RPC, or even some ad hoc XML
language). However, SOAP is by far the most popular choice
for Web Services.
Transport: Finally, all these messages must be
transmitted somehow between the server and the client. The
protocol of choice for this part of the architecture is HTTP
(Hypertext Transfer Protocol), the same protocol used to
access conventional web pages on the Internet. Again, in
theory we could be able to use other protocols, but HTTP is
currently the most used one.
C. Web Service Definition Language(WSDL)
Web Services programmers usually only have to
concentrate on writing code in their favorite programming
language and, in some cases, in writing WSDL. SOAP code,
on the other hand, is always generated and interpreted
automatically for us. Once we've reached a point where our
client application needs to invoke a Web Service, we
delegate that task on a piece of software called a stub. Using
stubs simplifies our applications considerably. We don't have
to write a complex client program that dynamically generates
SOAP requests and interprets SOAP responses (and similarly
for the server side of our application). We can simply
concentrate on writing the client and/or server code, and
leave all the dirty work to the stubs (which, again, we don't
even have to write ourselves... they can be generated
automatically from the WSDL description of a web
service). The stubs are generally generated only once. In
general, we only go through the discovery step once, then
generate the stubs once (based on the WSDL of the service
we've discovered) and then reuse the stubs as many times as
we want (unless the maintainers of the Web service decide to
180
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
change the service's interface and, thus, its WSDL
description).
II. TRIGON BASED AUTHENTICATION
ARCHITECTURE
When legitimate entities (users) login, the trigon based
authentication server splits the password into its components
and stores the authentication information in two servers -
namely authentication and backend server. Users have to
register with the Authentication server, so that it can hold a
part of the interpreted password with itself and another part
in the Backend server. The block diagram illustrating the
registration process of the users is depicted in the Figure
5. As illustrated in Figure5, the users who require services
from the VO have to register initially with the Authentication
server using their username and password. The
Authentication server calculates the Pi as given in (1). Along
with this authentication server generates two large prime
numbers, namely, a and a', which are considered as the two
sides of a trigon. It is difficult to hack the values of a and a'
as they are large prime numbers (as per RSA Factoring
Challenge). Here, Pi is taken as the angle between the two the
two sides of the trigon a and a'. Now, the Authentication
server can easily determine the opposite side of the angle P i?
termed as a".
With these trigon parameters, a, Vaa' and Paa' are
found as
(IJCSIS) International Journal of Computer Science and Information Security,
Vol 8, No. 6, September 2010
Vaa'=a-a'
(i)
User registers with username
and password in Authenticati on
1
f -\
Authentication file calculates
a. a" .a"
I
Determines variance and pro duct
c£ the sides V ££ ' and P^ ~~ and a
Saves fj. : username in authentication
file and forwards username, "VW and
P^ 'to Backend file
I
B acken d f il e saves VW an d P ££ Tor
the corresrjondine username
Fig 1 Flow Diagram
In (4), PAI is the ASCII-interpreted value of the given
password pwi , n is the total number of digits in PAI and PAI
( j) represents the first j digits of PAI . The PAI can be
calculated by the following steps.
Change the pwi into its corresponding ASCII value.
Paa' = a * a' (2)
a = 2Paa'-a" 2 (3)
where, a, a' and a' 'are the three sides of trigon. a is a
strengthening parameter used as the index . Vaa' and Paa' are
the variance and the product of the sides a and a'
respectively. With the parameters a , a' and a' ' as the
sides of trigon and Pi be the angle between the sides a
and a' the generated trigon will be assumed . After the
calculation of a , Vaa' and Paa' , the authentication server
stores the a value and its corresponding username in a
database and forwards Vaa' and Paa' to the Backend server
along with the username. Hence, the password is interpreted
and alienated into two units and stored in two separate
server. The authentication procedure is based on the
fundamental concepts of a trigon. Initially, the user who
wants the services of VO has to login to the Authentication
server using the username and password. Here, ui and
pwi refers to username and password of i th user. The
Authentication server calculates the Password index ( Pi )
from the password as
Pi =
P AI(i) / 10 pow n-2 ; if P AI (i) > 180
Calculate the three-fourth of total digits of the
ASCII value modulo 1 80, which results the first three digits
of PAI .
Append the remaining one- fourth of the ASCII digits to
PAI,
Then, from Pi the Authentication Server determines
the Authentication index ( AI ) for ui as
AI(i) = Pi/2
(5)
PAI(i)/10pown-3 ; else
(4)
Then, the Authentication Server searches for the
username index a i for the corresponding ui which has
already been stored in the server database during
the process of the registration. Subsequently, a i is
sent to the backend server along withui . When the
Backend server receives the index a i and the
username from the Authentication server, it
searches for Vaa' and Paa' the Variance and the product
of the sides a and a' respectively, which have been
saved in the backend server database during the
process of registration. From these values,
181
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol 8, No. 6, September 2010
the Backend server calculates an Authentication Token
ATi and sends it to the Authentication server to xmlns:tns= M http://www.globus.org/namespaces/add/hello_in
authenticate the ui . The ATi can be calculated as stance"
AT(i) = ai + Vaa'I * 2Paa'I (6)
In (6), Vaa' and Paa' are pre-calculated
values computed during individual user registration. After
retrieval of ATi from the Backend server, the
Authentication server authenticates the user based on the
token from the Backend server and the index calculated at
the Authentication server. The authentication code (or)
condition which authenticates the ui is given by
SinAI(i) =(l-ATi /2) 1 / 2 (7)
The authentication process is performed by the
authentication condition given in (7). When the condition is
satisfied, the user is decided to be valid and the Server sends
a token called Token for VO access to the user.
III. IMPLEMENTATION - AUTHENTICATION AS
SERVICE
A service is an entity that provides some capability to its
clients by exchanging messages. A service is defined by
identifying sequences of specific message exchanges that
cause the service to perform some operation. By thus
defining these operations only in terms of message exchange,
we achieve great flexibility in how services are implemented
and where they may be located. A service-oriented
architecture is one in which all entities are services, and thus
any operation visible to the architecture is the result of
message exchange.
Prerequisites are:
build.xml
globus-build-service. sh
1 . Creation of auth. wsdl File
<?xml version= M 1.0" encoding="UTF-8"?>
<defmitions name-'auth"
targetNamespace= "hrtp ://www. globus . org/namespaces/add/h
elloinstance"
xmlns= "hrtp ://schemas .xmlsoap . org/wsdl/"
xmlns:xsd="http://www.w3 .org/200 l/XMLSchema">
<types>
<xsd: schema
targetNamespace= "hrtp ://www. globus . org/namespaces/add/h
ello instance"
xmlns:tns="http://www.globus.org/namespaces/add/hello_in
stance"
xmlns:xsd="hrtp://www.w3 .org/200 l/XMLSchema">
<xsd:element name="addition">
<xsd:complexType>
<xsd:sequence>
<xsd:element name="inputl" type="xsd:int"/>
<xsd:element name="input2" type="xsd:int"/>
</xsd:sequence>
</xsd: complexType>
</xsd:element>
<xsd:element name="response" type="xsd: string "/>
<xsd: element
type="xsd: string "/>
name-'additionrequest"
</xsd:schema>
</types>
<message name="AddInputMessage">
<part name-'parameters"
element="tns:additionrequest"/>
</message>
<message name="AddOutputMessage">
<part name="resp" element="tns:response"/>
</message>
<portType name="authPortType" >
<operation name="addition">
<inputmessage="tns:AddInputMessage"/>
<outputmessage="tns:AddOutputMessage"/>
</operation>
</portType>
</definitions>
2. Create namespace2package.mappings for mapping
instances ,bindings and services.
182
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
3. Write Implementation program .
4.Create deploy-server.wsdd
[globus@g20 service]$ vi deploy-server.wsdd
5. Create deploy-jndi-config.xml
[globus@g20 service] $ vi deploy-jndi-config.xml
6.Build the service
[globus@g20 -example] $ sh globus-build-service.sh -d
org/add/service/ -s schema/add/hello.wsdl
7. After the successful building Grid Archive(GAR) file
has been created. Now we have to deploy the GAR file using
globus-deploy-gar command.
[globus@g20~example] $ globus-deploy-gar
orgaddser vice, gar
8. After successful deployment of the GAR file start the
globus container.
[globus@g20 ~]$ globus-start-container
(IJCSIS) International Journal of Computer Science and Information Security,
Vol 8, No. 6, September 2010
[37]:
hrtp://192. 168. 100. 3 :8443/wsrtf services/Trigger Service
[38]:
hrtp://192. 168. 100.3 :8443/wsrf / services/gsi/AuthenticationS
ervice
i BJE3SSEH
0§§
File Edit View Terminal Help
Anitha
Authentication check
Anitha
Authentication check
Anitha
Authentication check
Anitha
Authentication check
Anitha
Authentication check
Anitha t
alpha value
Alpha : -5.1360B02760354224E11
Vp : -764080.0
Pp : 1.35901310201E11
Token is : 0.2583132520673121
Pi : 104.97
Authentication token <At> 0.2583132520673121
myToken : 0.3708433739663
UserToken : 0.3708433739663
RegToken : 0.3708433739663
VALID USER
Authentication call
1
[39]:
hrtp://192. 168. 100.3 :8443/wsrf7services/TestRPCService
[40]:
hrtp://192. 168. 100.3:8443/wsrFservices/ManagedMultiJobSe
rvice
9. Write Client Program for authentication.
10. Before running the compiler, make sure to run the
following:
source SGLOBUSLOC ATION/etc/globus-devel-env. sh
The globus-devel-env.sh script takes care of putting all
the Globus libraries into your CLASSPATH.
[globus@gcluster example] $ source /usr/local/globus-
4.0.7/etc/globus-devel-env.sh
1 1 . [globus@gcluster example] $ j avac
org/add/client/Client.java
12. [globus@gcluster example] $ Java
org/add/client/Client
iObus@anitt:~/TrigonBased Authenticate
glii
File Edit View Terminal Help
[globus@anitt TrigonBasedAuthentication] $ source /usr/local/globus-4. Q . 7/etc/glo
bus-devel-env.sh
[globus@anitt TrigonBasedAuthentication] $ javac org/add/client/Client.java
[globus@anitt TrigonBasedAuthentication]$ Java org/add/client/Client
https : //192 . 168 . 4 . 114 : 8443/wsrf /services/TrigonBasedAuthenticationService
Inputs to TrigonBasedAuthentication Service : Anitha p Anitha
Username and Password Authentication: INVALID USER
[globus@anitt TrigonBasedAuthentication] $ Java org/add/client/Client
https : //192 . 168 . 4 . 114 : 8443/wsrf /services/TrigonBasedAuthenticationService
Inputs to TrigonBasedAuthentication Service : Anitha p Anitha
Username and Password Authentication: VALID USER
[globus@anitt TrigonBasedAuthentication] $ \\
n
Fig 2. GUI - Server side authentication
Fig 3. GUI -Valid User
[35]:
hrtp://192. 168. 100.3:8443/wsrf7services/DefaultTriggerServi
ce
[36]:
http://192.168.100.3:8443/wsrf/services/TrigonBasedAuth
enticationService
183
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
The user is invalid since the username and password
haven't stored in authentication and backend file. First time
when user login his/her trigon value of the password gets
stored in the respective files. Next time when they login they
gets authenticated and token gets issued .
IV. CONCLUSION
The Internet is a reasonable model for the Grid, providing
both an early version of its services and a platform from
which to evolve. The authentication protocol, proposed here,
enhanced the grid security as the authentication mechanism
utilized two servers for authentication. This simple trigon
concept utilization in the authentication protocol introduced a
novel and revolutionary idea in the authentication
mechanism as well as in grid environment. The
implementation of our dual authentication protocol showed
its effective performance in pinpointing the adversaries and
paving the way to valid users for access with the VO for
resource sharing. So by hosting this authentication as a
service it make the grid environment more secure. In future
these services will be located by type instead by names.
ACKNOWLEDGMENT
Our thanks to Dr.R.Rudramoorthy,Principal,PSG
College of Technology and Mr. K.Chidambaram, Director,
Grid and Cloud systems group, Yahoo software
development, India Private Limited for their support. This
project is carried out in Grid and Cloud lab, PSG College of
Technology.
REFERENCES
[1] S. Albayrak, S. Kaiser, and J. Stender. Advanced grid management
software for seamless services. Multiagent Grid Syst, 1(4):263.270,
2005.
[2] Andrew Johnson, Carl Kesselman, Jason Leigh, and Steven Tuecke,
Application Experiences with the Globus Toolkit, Seventh IEEE
International Symposium on High Performance Distributed
Computing (HPDC-7 ... T).
[3] Antonioletti, M, and Jackson, M, OGSA-DAI Product Overview,
2003. Available at www.ogsa-dai.org.uk/downloads/docs/OGSA-
DAI-USER-M3 -PRODUCTOVERVIEW.pdf.
[4] Christensen, E., Curbera, F., Meredith, G., and Weerawarana, S., Web
Services Description Language (WSDL) 1.1, W3C, Note 15, 2001.
Available at www.w3.org/TR/wsdl.
[5] K. Czajkowski, A. Dan, J. Rofrano, S. Tuecke, and M. Xu.
Agreement-based grid service management (ogsiagreement),
Proceedings of 6th IEEE/ACM International Workshop on Grid
Computing (Grid2005).
[6] Della-Libera, G. et al. (2002) Web Services, Secure Conversation
Languages (WS-Secure Conversation). Version 1.0, available at
http://msdn.microsoft.com/library/default.asp7urW
(IJCSIS) International Journal of Computer Science and Information Security,
Vol 8, No. 6, September 2010
lib/enus/dnglobspec/html/wssecureconversation.asp (accessed on
2002).
[7] Eastlake, D. and Reagle, J. (Eds.) (2002) XML Encryption Syntax
and Processing. W3C Recommendation, available at
http://www.w3.org/TR/xmlenc-core/ (accessed on December 2002).
[8] Foster. I., Kesselman. C. and Tuecke. S, "The Anatomy of the Grid:
Enabling Scalable Virtual Organizations", International Journal of
High Performance Computing Applications", vol. 15, no. 3, pp. 200-
222,2001.
[9] I. Foster, C. Kesselman, J. Nick, and S. Tuecke. The physiology of
the grid: An open grid services architecture for distributed systems
integration, 2002.
[10] I. foster, Argonne National Laboratory and the University of Chicago
,carl kesselman, Information Sciences Institute, University of
Southern California "THE GRID 2 blueprint for a new computing
infrastructure",2004
[11] Graham, S., Simeonov, S., Boubez, T. etc, "Building Web Service
with Java: Making Sense of XML, SOAP, WSDL and UDDI",
Indianapolis, IN: Sams Publishing, 2002.
[12] Gudgin, M., Hadley, M., Mendelsohn, N., Moreau, J-J. and Nielsen,
H.F. (2003) SOAP Version 1.2 Part 1: Messaging Framework. W3C
Recommendation, Available at http://www.w3.org/TR/soapl2-partl/
(accessed on June 2003).
[13] Grid Archive Creation: http://gdp.globus.org/gt4-
tutorial/multiplehtml/ch08s02.html.
[14] G. Laccetti and G. Schmid, "A framework model for grid security",
Future Generation Computer Systems, vol. 23, no. 5, pp.702-713,
June 2007.
[15] Li, Y., Jin, H., Zou, D., Chen, J. and Han, Z. (2007) 'A scalable
service scheme for secure group communication in grid', 31st Annual
International Computer Software and Applications Conference
(COMPSAC 2007).
[16] Li, Y., Jin, H., Zou, D., Liu, S. and Han, Z. (2008) 'An authenticated
encryption mechanism for secure group communication in grid', 2008
International Conference on Internet Computing in Science and
Engineering.
[17] Li Hongweia, Sun Shixina and Yang Haomiaoa, "Identity-based
authentication protocol for grid", Journal of Systems Engineering and
Electronics, Vol. 19, no. 4, pp.860-865, August 2008.
[18] Nagaratnam, N., Janson, P., Dayka, J., Nadalin, A., Siebenlist, F.,
Welch, V., Foster, I., and Tuecke, S., Security architecture for Open
Grid Services, 2002. Available at
www.globus.org/ogsa/Security/draft-ggf-ogsa-sec-arch-0 1 .pdf.
[19] Open Grid Services Architecture Data Access and Integration
(OGSA-DAI) Project: www.ogsa-dai.org.uk.
[20] K. Rochford, B. A. Coghlan, and J. Walsh. An agent-based approach
to grid service monitoring. In Proc. International Symposium on
Parallel and Distributed Computing (ISPDC 2006), July July, 2006.
[21] Shirasuna, S., Nakada, H., Matsuoka, S., and Sekiguchi, S.,
Evaluating Web services based implementations of GridRPC, in 11th
IEEE International Symposium on High Performance Distributed
Computing, Edinburgh, Scotland. IEEE Computer Society Press, Los
Alamitos, CA, 2001.
[22] Siebenlist, F., Welch, V., Tuecke, S., Foster, I., Nagaratnam, N.,
Janson, P., Dayka, J., and Nadalin, A., Roadmap towards a secure
OGSA. Global Grid Forum, draft, 2002.
[23] Shengxian Luo, Xiaochuan Peng, Shengbo Fan Peiyu Zhang, Study on
Computing Grid Distributed Middleware and Its
Application,International Forum on Information Technology and
Application,2009.
[24] Thatte, S., XLANG, Web services for business process design Web
site, Microsoft Corporation.
www.gotdotnet.com/team/xml_wsspecs/xlang-c/default.htm..
[25] Ian Foster ,Carl Kesselman,The Globus Project: A Status Report, 5th
IEEE Symp. on High. Performance Distributed Computing.
[26] H.-L. Truong, R. Samborski, and T. Fahringer. Towards a framework
for monitoring and analyzing qos metrics of grid services. In E-
SCIENCE '06: Proceedings of the Second IEEE International
184
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
[27]
[28]
[29]
[30]
Conference on e-Science and Grid Computing, page 65, Washington,
DC, USA, 2006. IEEE Computer Society.
UDDI. The UDDI technical white paper, http://www.uddi.org/, 2000.
Von Welch, Frank Siebenlist, Ian Foster, John Bresnahan, Karl
Czajkowski, Jarek Gawor, Carl Kesselman, Sam Meder, Laura
Pearlman and Steven Tuecke, "Security for Grid Services", in
proceedings of the 12th IEEE International Symposium on High
Performance Distributed Computing, pp.48- 57, June 2003.
W3C Note "Web Services Definition Language (WSDL) 1.1",
http://www.w3.org/TR/WSDL.
W3C Note "Simple Object Access Protocol(SOAP)
1 . 1",http://swww.w3 .org/TR/WSDL.
(IJCSIS) International Journal of Computer Science and Information Security,
Vol 8, No. 6, September 2010
Cloud Computing. She has presented 1 paper in National Conference. She
is the Best Outgoing Student in MTech 2010-2011 in PSG College of
Technology. You may contact her at dhaarinimp@gmail.com
AUTHORS PROFILE
Dr G Sudha Sadasivam is working as a
Professor in Department of Computer Science and
Engineering in PSG College of Technology,
India. Her areas of interest include, Distributed
Systems, Distributed Object Technology, Grid
and Cloud Computing. She has published 20
papers in referred journals and 32 papers in
National and International Conferences. She has
coordinated two AICTE - RPS projects in
Distributed and Grid Computing areas. She is also the coordinator for PSG-
Yahoo Research on Grid and Cloud computing. You may contact her at
sudhasadhasivam@yahoo . com
in Sri Ramakrishna En
Grid Computing special
ruckmaniv@yahoo.com
V Ruckmani received B. Sc, MCA and M. Phil
degrees from the department of computer science,
Bharathiar University, India in 1994, 1997 and
2003 respectively. She is currently pursuing the
Ph. D degree, working closely with Prof. G.
Sudha Sadasivam. From 1997 to 2000 she worked
at PSG College of Arts and Science in the
department of Computer Science. Since
December 2000 she is working as a senior
lecturer in Department of Computer Applications
ineering College, India. She works in the field of
izing in the area of security. You may contact her at
EK Anitha Kumari received BE(CSE) from
Department of Computer Science ,Avinashilingam
Deemed University and ME(SE) from Department
of Computer Science ,Anna University. She is
working as a Lecturer in Department of Information
Technology in PSG College of Technology, India.
Her areas of interest include Grid and Cloud
Computing. She has published 1 paper in referred
international journal and 5 papers in National and
International Conferences. She is the Best Outgoing Student in ME 2010-
201 1 in PSG College of Technology. She awarded Gold Medal in BE(CSE)
in Avinashilingam Deemed University . You may contact her at
anitha.psgsoft@gmail.com
M P Dhaarini received BTech(IT) from
Department of Information Technology ,Anna
University and MTech(IT) from Department of
Information Technology, Anna University. She is
working as a Lecturer in Department of
Information Technology in PSG College of
Technology, India. Her areas of interest include
185
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
Dr. H. B. Kekre et. al.(IJCSIS) International Journal of Computer Science and Information Security, Vol. 8, No. 6, 2010
Performance Evaluation of Speaker Identification for
Partial Coefficients of Transformed Full, Block and
Row Mean of Speech Spectrogram using DCT,
WALSH and HAAR
Dr. H. B. Kekre
Senior Professor,
MPSTME, SVKM's NMIMS
University
Mumbai, 400-056, India
Dr. Tanuja K. Sarode
Assistant Professor,
Thadomal Shahani Engg.
College, Bandra (W),
Mumbai, 400-050, India
Shachi J. Natu
Lecturer,
Thadomal Shahani Engg.
College, Bandra (W),
Mumbai, 400-050, India
Prachi J. Natu
Assistant Professor,
GVAIET, Shelu
Karjat 410201,
India
Abstract- In this paper an attempt has been made to provide
simple techniques for speaker identification using transforms
such as DCT, WALSH and HAAR alongwith the use of
spectrograms instead of raw speech waves. Spectrograms form a
image database here. This image database is then subjected to
different transformation techniques applied in different ways
such as on full image, on image blocks and on Row Mean of an
image and image blocks. In each method, results have been
observed for partial feature vectors of image. From the results it
has been observed that, transform on image block is better than
transform on full image in terms of identification rate and
computational complexity. Further, increase in identification rate
and decrease in computations has been observed when
transforms are applied on Row Mean of an image and image
blocks. Use of partial feature vector further reduces the number
of comparisons needed for finding the most appropriate match.
Keywords- Speaker Identification, DCT, WALSH, HAAR,
Image blocks, Row Mean, Partial feature vector.
I. INTRODUCTION
To provide security in a multiuser environment, it has
become crucial to identify users and to grant access only to
those users who are authorized. Apart from the traditional login
and password method, use of biometric technology for the
authentication of users is becoming more and more popular
nowadays. Biometrics comprises methods for uniquely
recognizing humans based upon one or more intrinsic physical
or behavioral traits. Biometric characteristics can be divided in
two main classes: Physiological which are related to the shape
of the body. Examples include fingerprint, face recognition,
DNA, hand and palm geometry, iris recognition etc.
Behavioral, which are related to the behavior of a person.
Examples include typing rhythm, gait and voice. Techniques
like face recognition, fingerprint recognition and retinal blood
vessel patterns have their own drawbacks. To identify an
individual by these methods, he/she should be willing to
undergo the tests and should not get upset by these procedures.
Speaker recognition allows non-intrusive monitoring and also
achieves high accuracy rates which conform to most security
requirements. Speaker recognition is the process of
automatically recognizing who is speaking based on some
unique characteristics present in speaker's voice [2]. There are
two major applications of speaker recognition technologies and
methodologies: speaker identification and speaker verification.
In the speaker identification task, a speech utterance from
an unknown speaker is analyzed and compared with speech
models of known speakers. The unknown speaker is identified
as the speaker whose model best matches the input utterance.
In speaker verification, an identity is claimed by an unknown
speaker, and an utterance of this unknown speaker is compared
with a model for the speaker whose identity is being claimed. If
the match is good enough, that is, above a threshold, the
identity claim is accepted. The fundamental difference between
identification and verification is the number of decision
alternatives [3]. In identification, the number of decision
alternatives is equal to the size of the population, whereas in
verification there are only two choices, acceptance or rejection,
regardless of the population size. Therefore, speaker
identification performance decreases as the size of the
population increases, whereas speaker verification performance
approaches a constant, independent of the size of the
population, unless the distribution of physical characteristics of
speakers is extremely biased.
Speaker identification can be further categorized into text-
dependent and text independent speaker identification based on
the relevance to speech contents [2, 4].
Text Dependent Speaker Identification requires the speaker
saying exactly the enrolled or given password/speech. Text
Independent Speaker Identification is a process of verifying the
identity without constraint on the speech content. It has no
186
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
Dr. H. B. Kekre et. al.(IJCSIS) International Journal of Computer Science and Information Security, Vol. 8, No. 6, 2010
advance knowledge of the speaker's utterance and is more
flexible in situation where the individuals submitting the
sample may be unaware of the collection or unwilling to
cooperate, which presents more difficult challenge.
Compared to Text Dependent Speaker Identification, Text
Independent Speaker Identification is more convenient because
the user can speak freely to the system. However, it requires
longer training and testing utterances to achieve good
performance. Text Independent Speaker Identification is more
difficult problem as compared to Text Dependent Speaker
Identification because the recognition system must be prepared
for an arbitrary input text.
Speaker Identification task can be further classified into
closed set and open set identification.
In closed set problem, from N known speakers, the speaker
whose reference template has the maximum degree of
similarity with the template of input speech sample of unknown
speaker is obtained. This unknown speaker is assumed to be
one of the given set of speakers. Thus in closed set problem,
system makes a forced decision by choosing the best matching
speaker from the speaker database.
In the open set text dependent speaker identification,
matching reference template for an unknown speaker's speech
sample may not exist. So the system must have a predefined
tolerance level such that the similarity degree between the
unknown speaker and the best matching speaker is within this
tolerance.
In the proposed method, speaker identification is carried out
with spectrograms and transformation techniques such as DCT,
WALSH and HAAR [15-18]. Thus an attempt is made to
formulate a digital signal processing problem into pattern
recognition of images.
The rest of the paper is organized as follows: in section II
we present related work carried out in the field of speaker
identification. In section III our proposed approach is
presented. Section IV elaborates the experiment conducted and
results obtained. Analysis of computational complexity is
presented in section V. Conclusion has been outlined in section
VI.
II. RELATED WORK
All speaker recognition systems at the highest level contain
two modules, feature extraction and feature matching.
Feature extraction is the process of extracting subset of
features from voice data that can later be used to identify the
speaker. The basic idea behind the feature extraction is that the
entire feature set is not always necessary for the identification
process. Feature matching is the actual procedure of identifying
the speaker by comparing the extracted voice data with a
database of known speakers and based on this suitable decision
is made.
There are many techniques used to parametrically represent
a voice signal for speaker recognition task. One of the most
popular among them is Mel-Frequency Cepstrum Coefficients
(MFCC)[1].
The MFCC parameter as proposed by Davis and
Mermelstein [5] describes the energy distribution of speech
signal in a frequency field. Wang Yutai et. al. [6] has proposed
a speaker recognition system based on dynamic MFCC
parameters. This technique combines the speaker information
obtained by MFCC with the pitch to dynamically construct a
set of the Mel-filters. These Mel-filters are further used to
extract the dynamic MFCC parameters which represent
characteristics of speaker's identity.
Sleit, Serhan and Nemir [7] have proposed a histogram
based speaker identification technique which uses a reduced set
of features generated using MFCC method. For these features,
histograms are created using predefined interval length. These
histograms are generated first for all data in feature set for
every speaker. In second approach, histograms are generated
for each feature column in feature set of each speaker.
Another widely used method for feature extraction is use of
linear Prediction Coefficients (LPC). LPCs capture the
information about short time spectral envelope of speech. LPCs
represent important speech characteristics such as formant
speech frequency and bandwidth [8].
Vector Quantization (VQ) is yet another approach of
feature extraction [19-22]. In Vector Quantization based
speaker recognition systems; each speaker is characterized with
several prototypes known as code vectors [9]. Speaker
recognition based on non-parametric vector quantization was
proposed by Pati and Prasanna [10]. Speech is produced due to
excitation of vocal tract. Therefore in this approach, excitation
information can be captured using LP analysis of speech signal
and is called as LP residual. This LP residual is further
subjected to non-parametric Vector Quantization to generate
codebooks of sufficiently large size. Combining nonparametric
Vector Quantization on excitation information with vocal tract
information obtained by MFCC was also introduced by them.
III.PROPOSED METHODS
In the proposed methods, first we converted the speech
samples collected from various speakers into spectrograms
[11]. Spectrograms were created using Short Time Fourier
Transfer method as discussed below:
In the approach using STFT, digitally sampled data are
divided into chunks of specific size say 128, 256 etc. which
usually overlap. Fourier transform is then obtained to calculate
the magnitude of the frequency spectrum for each chunk. Each
chunk then corresponds to a vertical line in the image, which is
a measurement of magnitude versus frequency for a specific
moment in time.
Thus we converted the speech database into image
database. Different transformation techniques such as Discrete
Cosine Transform [12], WALSH transform and HAAR
transform are then applied to these images in three different
ways to obtain their feature vectors.
1. Transform on full image
2. Transform on image blocks obtained by dividing an
image into four equal and non-overlapping blocks
187
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
Dr. H. B. Kekre et. al.(IJCSIS) International Journal of Computer Science and Information Security, Vol. 8, No. 6, 2010
3. Transform on Row Mean of an image and on Row
Mean of image blocks.
From these feature vectors, again identification rate is
obtained for various portions selected from the feature vector
i.e. for partial feature vector [15, 23, 24]. Two different sets of
database were generated. First set, containing 60% of the total
images as trainee images and 40% of the total images as test
images. Second set, containing 80% of the images as trainee
images and 20% of the total images as test images. Euclidean
distance between test image and trainee image is used as a
measure of similarity. Euclidean distance between the points
X(X1, X2, etc.) and point Y (Yl, Y2, etc.) is calculated using
the formula shown in equation. (1).
D:
S(X i -Y i ) z
i=l
(i)
Smallest Euclidean distance between test image and trainee
image means the most probable match of speaker. Algorithms
for transformation technique on full image and transformation
techniques on image blocks are given below.
A Transformation techniques on full image[27, 28]:
In the first method 2-D DCT / WALSH / HAAR is applied
on the full image resized to 256*256. Further, instead of full
feature vector of an image only some portion of feature vector
i.e. partial feature vector is selected for identification purpose.
This selection of feature vector is illustrated in Fig. 1 and it is
based on the number of rows and columns that have been
selected from the feature vector of an image. For example,
initially first full feature vector (i.e. 256*256) has been selected
and then partial feature vectors of size 192*192, 128*128,
64*64, 32*32, 20*20 and 16*16 were selected from the feature
vector. For these different sizes, identification rate was
obtained.
+
t
*
1
32
135
192
Fig. 1: Selection of partial feature vector
Algorithm for this method is as follows:
Step 1. For each trainee image in the database, resize an
image to size 256*256.
Step 2. Apply the transformation technique (DCT /
WALSH / HAAR) on resized image to obtain its
feature vector.
Step 3. Save these feature vectors for further comparison.
Step 4. For each test image in the database, resize an
image to size 256*256.
Step 5. Apply the transformation technique (DCT /
WALSH / HAAR) on resized image to obtain its
feature vector.
Step 6. Save these feature vectors for further comparison.
Step 7. Calculate the Euclidean distance between feature
vectors of each test image with each trainee
image corresponding to the same sentence.
Step 8. Select the trainee image which has smallest
Euclidean distance with the test image and
declare the speaker corresponding to this trainee
image as the identified speaker.
Repeat Step 7 and Step 8 for partial feature vector
obtained from the full feature vector.
B. Transformation technique on image blocks[27, 29]:
In this second method, resized image of size 256*256 is
divided into four equal parts as shown in Fig. 2 and then 2-D
DCT / WALSH / HAAR is applied to each part.
Fig. 2: Image divided into four equal non-overlapping blocks
Thus when N*N image is divided into four equal and non-
overlapping blocks, blocks of size N/2*N/2 are obtained.
Feature vector of each block when appended as columns forms
a feature vector of an image. Thus size of feature vector of an
image in this case is of 128*512. Again Euclidean distance is
used as a measure of similarity. Here also using partial feature
vectors, identification rate has been obtained. Partial feature
vectors of size 96*384, 64*256, 32*128, 16*64 and 8*32 have
been selected to find identification rate. Detailed steps are
explained in algorithm given below:
Step 1. For each trainee image in the database, resize an
image to size 256*256.
Step 2. Divide the image into four equal and non-
overlapping blocks as explained in Fig. 2.
Step 3. Apply transformation technique (DCT/ WALSH
/HAAR) on each block obtained in Step 2.
Step 4. Append the feature vectors of each block one
after the other to get feature vector of an image.
Step 5. For each test image in the database, resize an
image to size 256*256.
Step 6. Divide the image into four equal and non-
overlapping blocks as shown in Fig. 2.
Step 7. Apply transformation technique (DCT /WALSH
/HAAR) on each block obtained in Step 6.
Step 8. Append the feature vectors of each block one
after the other to get feature vector of an image.
188
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
Dr. H. B. Kekre et. al.(IJCSIS) International Journal of Computer Science and Information Security, Vol. 8, No. 6, 2010
Step 9. Calculate the Euclidean distance of each test
image with each trainee image corresponding to
the same sentence.
Step 10. Select the trainee image which has smallest
Euclidean distance with the test image and
declare the speaker corresponding to this trainee
image as the identified speaker.
Repeat Step 9 and Step 10 for partial feature vectors
selected from feature vector obtained in Step 4 and Step 8.
Selection of partial feature vector is similar to the one shown in
Fig. 1. But in this method, size of feature vector is 128*512,
96*384, 64*256, 32*128, 16*64 and 8*32.
C. Transformation techniques on Row Mean [16-18] of an
image and on Row Mean of image blocks [27, 29]:
In this approach, Row Mean of an image is calculated. Row
mean is nothing but an average of pixel values of an image
along each row. Fig. 3 shows how the Row Mean of an image
is obtained.
IV.
EXPERIMENTS AND RESULTS
Fig. 3: Row Mean of an image
1-D DCT / WALSH / HAAR is then applied on this Row
mean of an image to its feature vector and Euclidean distance is
used as measure of similarity to identify speaker. Detail
algorithm is given below:
Step 1: For each trainee image in the database, resize an
image to size 256*256.
Step 2: Calculate Row Mean of an image as shown in
Fig. 3.
Step 3: Apply 1-D transformation technique (DCT /
WALSH / HAAR) on Row Mean obtained in
Step 2. This gives the feature vector of an image.
Step 4: For each test image in the database, resize an
image to size 256*256.
Step 5: Apply 1-D transformation technique (DCT /
WALSH / HAAR) on Row Mean obtained in
Step 4. This gives the feature vector of an image.
Step 6: Calculate the Euclidean distance of each test
image with each trainee image corresponding to
the same sentence.
Step 7: Select the trainee image which has smallest
Euclidean distance with the test image and
declare the speaker corresponding to this trainee
image as the identified speaker.
For Row Mean of image blocks, first divide the image into
equal and non-overlapping blocks (of size 128*128, 64*64,
32*32, 16*16 and 8*8). Obtain Row Mean of each block as
shown in Fig. 3. Transformation technique is then applied on
Row Mean of each block and then combined into columns to
get feature vector of an image.
Implementation for the proposed approach was done on
Intel Core 2 Duo Processor, 2.0 GHz, and 3 GB of RAM.
Operating System used is Windows XP and softwares used are
MATLAB 7.0 and Sound forge 8.0. To study the proposed
approach we recorded six distinct sentences from 30 speakers:
11 males and 19 females. These sentences are taken from
VidTIMIT database [13] and ELSDSR database [14]. For every
speaker 10 occurrences of each sentence were recorded.
Recording was done at varying times. This forms the closed set
for our experiment. From these speech samples spectrograms
were created with window size 256 and overlap of 128. Before
creation of spectrograms, DC offset present in speech samples
was removed so that signals are vertically centered at 0. After
removal of DC offset, speech samples were normalized with
respect to amplitude to -3 dB and also with respect to time.
Spectrograms generated from these speech samples form the
image database for our experiment. In all we had 1800
spectrograms in our database.
From these spectrograms, two sets were created.
Set A: Contains six spectrograms as trainee images per
speaker and four spectrograms as test images per speaker. So in
all it contains 1080 trainee images and 720 test images.
Set B: Contains eight spectrograms as trainee images per
speaker and two spectrograms as test images per speaker. So in
all it contains 1440 trainee images and 360 test images.
Since our work is restricted to text dependent approach,
Euclidean distance for a test image of speaker say 'x' for a
particular sentence say 'si' is obtained by comparing the
feature vector of that test image with the feature vectors of all
the trainee images corresponding to sentence 'si'. Results are
calculated for set of test images corresponding to each
sentence.
A Results for DCT/WALSH/HAAR on Full image:
1) Results for DCT on full image
Table I shows the identification rate for six sentences si to
s6 when DCT is applied on full image in set A and partial
feature vectors are selected to find the matching spectrogram.
Table I. Identification rate for sentences si to s6 for full and
PARTIAL FEATURE VECTOR WHEN DCT IS APPLIED TO FULL IMAGE IN SET A
Portion of feature
vector selected
Sentence
SI
S2
S3
S4
S5
S6
256*256
54.16
59.16
56.66
56.66
68.33
62.50
192*192
58.33
65
67.5
65
73.33
69.16
128*128
65.83
64.16
71.66
67.5
74.16
72.5
64*64
70.83
70.83
71.66
72.50
77.50
75.83
32*32
75
73.33
74.16
75
80
77.5
20*20
78.33
75.33
78.33
71.66
81.66
80
16*16
72.5
76.66
74.16
74.16
76.66
79.16
Table II shows the identification rate for six sentences si to
s6 when DCT is applied on full image in set B and partial
feature vectors are selected to find the matching spectrogram.
189
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
Dr. H. B. Kekre et. al.(IJCSIS) International Journal of Computer Science and Information Security, Vol. 8, No. 6, 2010
table ii. Identification rate for sentences sIto s6for varying
portion of feature vector when dctls applied to full image in set b
Portion of feature
vector selected
Sentence
SI
S2
S3
S4
S5
S6
256*256
63.33
66.67
75
66.67
76.67
76.67
192*192
73.33
70
76.67
75
78.33
78.33
128*128
78.33
73.33
80
78.33
81.67
81.67
64*64
80
80
78.33
86.67
83.33
88.33
32*32
90
86.67
86.67
86.67
86.67
90
20*20
86.67
86.67
86.67
88.33
90
90
16*16
85
85
86.67
86.67
91.67
90
Table III shows the comparison of overall identification
rate considering all sentences, for partial feature vectors of
different sizes when set A and set B is used. It also shows the
number of DCT coefficients used for identifying speaker for
corresponding selected portion of feature vector.
table in. Comparison of Overall Identification rate for
DIFFERENT NUMBER OF DCT COEFFICIENTS WHEN DCT IS APPLIED TO FULL
IMAGE IN SET A AND SET B
Portion of feature
vector selected
Number of DCT
coefficients
% Identification rate
Set A
SetB
256*256
65536
60
70.83
192*192
36864
66.38
75.27
128*128
16384
69.30
78.88
64*64
4096
73.19
82.77
32*32
1024
75.83
87.77
20*20
400
77.63
88.05
16*16
256
76.66
87.5
Table v. Identification rate for sentences s 1 to s 6 for full and
partial feature vector when walsh transform is applied to full
image from set b
Portion of feature
vector selected
Sentence
SI
S2
S3
S4
S5
S6
256*256
63.33
66.67
75
66.67
76.67
76.67
192*192
75
71.67
76.67
73.33
78.33
81.67
128*128
80
75
78.33
83.33
81.67
81.67
64*64
86.67
83.33
81.67
85
83.33
85
32*32
86.67
81.67
81.67
88.33
83.33
91.67
20*20
91.67
78.33
83.33
85
86.67
83.33
16*16
86.67
85
83.33
85
83.33
86.67
Table VI shows the overall identification rate considering
all sentences, for partial feature vectors. For set A, highest
identification rate is obtained for partial feature vector of size
64*64 i.e. 4096 WALSH coefficients. For set B, it requires
32*32 partial feature vector i.e. 1024 WALSH coefficients.
Table vi. Comparison of overall identification rate for varying
number of coefficients when walsh transform is applied to full
image from set a and set b
Portion of feature
vector selected
Number of Walsh
coefficients
% Identification rate
Set A
SetB
256*256
65536
60
70.83
192*192
36864
66.66
76.11
128*128
16384
70.69
80
64*64
4096
75
84.16
32*32
1024
73.33
85.55
20*20
400
72.91
84.72
16*16
256
71.94
85
2) Results for Walsh on full image
Results of Walsh transform on Spectrograms are tabulated
below. Table IV shows the identification rate for sentences si
to s6 for full and partial feature vectors when WALSH
transform is applied on full image and set A is used.
Table IV Identification rate for sentences sIto s6for varying
PORTION OF FEATURE VECTOR WHEN WALSH TRANSFORM IS APPLIED TO
FULL IMAGE FROM SET A
Portion of feature
vector selected
Sentence
SI
S2
S3
S4
S5
S6
256*256
54.16
59.16
57.5
57.5
68.33
63.33
192*192
59.16
66.66
65.83
63.33
73.33
71.66
128*128
65.83
66.66
70.83
73.33
75
72.5
64*64
74.16
73.33
76.66
75
75.83
75
32*32
70.83
71.66
71.66
70.83
78.33
76.66
20*20
70.83
69.67
71.67
70.83
76.67
78.33
16*16
70
70.83
71.67
66.67
75
77.5
Table V shows sentencewise identification rate for
WALSH transform on full image from set B. It can be
observed from Table IV and Table V that, identification rate
for each sentence is increased as more training is provided to
the system. From both the tables, it can be seen that as size of
the partial feature vector is decreased, the identification rate
also decreases, achieves its peak value and then again decrease.
3) Results for HAAR on full image
Table VII shows sentencewise identification rate when 2-D
HAAR transform is applied to full image with size 256*256
and partial feature vectors are selected from these feature
vectors. These results are for set A.
Table vii. Identification rate for sentences s1tos6for varying
portion of feature vector when haar transform is applied to full
image from seta
Portion of feature
vector selected
Sentence
SI
S2
S3
S4
S5
S6
256*256
54.16
59.16
57.5
57.5
68.33
63.33
192*192
59.16
66.66
65.83
63.33
73.33
71.66
128*128
65.83
66.66
70.83
73.33
75
72.5
64*64
74.16
73.33
76.66
75
75.83
75
32*32
70.83
71.66
71.66
70.83
78.33
76.66
20*20
65.83
73.33
71.67
70
75
76.67
16*16
70
70.83
71.67
66.67
75
77.5
Table VIII shows identification rate for HAAR transform
on full image when set B is used. From both the tables, it can
be seen that as the number of coefficients selected from the
feature vector is decreased, the identification rate also
decreases, achieves its peak value and then again decrease.
From the Table VII and Table VIII, it can also be noted that
when more training is provided to the system, identification
rate per sentence is increased.
190
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
Dr. H. B. Kekre et. al.(IJCSIS) International Journal of Computer Science and Information Security, Vol. 8, No. 6, 2010
Table viii. Identification rate for sentences sIto s6for varying
portion of feature vector when haar transform is applied to full
image with training set of eight images for each speaker
Portion of feature
vector selected
Sentence
SI
S2
S3
S4
S5
S6
256*256
63.33
66.67
75
66.67
76.67
76.67
192*192
80
73.33
78.33
76.67
78.33
78.33
128*128
80
75
78.33
83.33
81.67
81.67
64*64
86.67
83.33
81.67
85
83.33
85
32*32
86.67
81.67
81.67
88.33
83.33
91.67
20*20
86.67
88.33
86.67
85
85
86.67
16*16
86.67
85
83.33
85
83.33
86.67
Table IX shows identification rate obtained by considering
all six sentences, for set A and set B, with different sized partial
feature vectors. Maximum identification rate is observed for
4096 and 400 HAAR coefficients with set A and set B
respectively.
Table ix. Comparison of overall identification rate for varying
number of coefficients when haar transform is applied to full
image from set a and set b
Portion of feature
vector selected
Number of HAAR
coefficients
Identification rate (%)
Set A
SetB
256*256
65536
60
70.83
192*192
36864
67.91
77.5
128*128
16384
70.69
80
64*64
4096
75
84.16
32*32
1024
73.33
85.55
20*20
400
72.08
86.39
16*16
256
71.94
85
Table X shows the comparison of identification rates for all
three transformation techniques on full image when set A and
set B are used per speaker.
increases, reaches its maximum value and then again decreases
or remains constant.
Table xi. Identification rate for sentences sIto s6for full and
partial feature vector using dcton image blocks for images from
SETA
Portion of feature
vector selected
Sentence
SI
S2
S3
S4
S5
S6
128*512
54.16
59.16
57.5
57.5
68.33
63.33
96*384
60
63.33
65.33
65
73.33
68.33
64*256
65
65
70.83
66.66
74.16
71.16
32*128
70.83
70.83
70.83
71.66
76.66
75
16*64
75.83
74.16
75
75.83
81.66
77.5
8*32
69.16
76.66
75
75.83
75
75.83
Table xii. Identification rate for sentences si to s6for full and
PARTIAL FEATURE VECTOR USING DCTON IMAGE BLOCKS FOR IMAGES FROM
SET B
Portion of feature
vector selected
Sentence
SI
S2
S3
S4
S5
S6
128*512
63.33
66.67
75
66.67
76.67
76.67
96*384
71.67
70
76.67
75
78.33
78.33
64*256
78.33
73.33
80
76.67
81.67
81.67
32*128
78.33
80
78.33
86.67
83.33
86.67
16*64
90
88.33
86.67
90
86.67
88.33
8*32
88.33
88.33
85
86.67
90
86.67
Table XIII shows the comparison of overall identification
rate considering all sentences, for partial feature vectors using
DCT on image blocks. For both the training sets, maximum
identification rate is achieved for partial feature of size 16*64
i.e. for 1024 DCT coefficients.
Table xiii. Comparison of overall identification rate for full and
partial feature vector portion using dcton image blocks for
images from set a and set b
Table x.
Comparison of identification rates when DCT, WALSH
AND HAARON FULL IMAGE FROM SET A AND SET B
Portion of
feature
vector
selected
Identification rate (%)
when set A is used
Identification rate (%)
when set B is used
DCT
WALSH
HAAR
DCT
WALSH
HAAR
256*256
60
60
60
70.83
70.83
70.83
192*192
66.38
66.66
67.91
75.27
76.11
77.5
128*128
69.30
70.69
70.69
78.88
80
80
64*64
73.19
75
75
82.77
84.16
84.16
32*32
75.83
73.33
73.33
87.77
85.55
85.55
20*20
77.63
72.91
72.08
88.05
84.72
86.39
16*16
76.66
71.94
71.94
87.5
85
85
B. Results for DCT/WALSH/HAAR on image block:
1) Results for DCT on image blocks:
Table XI shows the identification rate for sentences si to s6
when full and partial feature vectors are selected to identify
speaker using DCT on image blocks using set A. Table XII
shows the sentence wise identification rate for full and partial
feature vectors when DCT is applied on image blocks for set B.
It can be seen from the table that identification rate is improved
when more training is provided to the system. From both the
tables, it can be seen that, as the number of coefficients used
for identification purpose decreases, the identification rate
Portion of feature
vector selected
Number of DCT
coefficients
Identification rate (%)
Set A
SetB
128*512
65536
60
70.83
96*384
36864
65.97
75
64*256
16384
68.88
78.61
32*128
4096
72.63
82.22
16*64
1024
76.66
88.33
8*32
256
74.58
86.67
2) Results for WALSH on image blocks:
Table IVX on the next page shows the sentencewise
identification rate when WALSH transform is applied to image
blocks using images in Set A.
Table IVX. Identification rate for sentences sIto s6for varying
PORTION OF FEATURE VECTOR USING WALSHON IMAGE BLOCKS WITH
IMAGES FROM SET A
Portion of feature
vector selected
Sentence
SI
S2
S3
S4
S5
S6
128*512
54.16
59.17
57.5
57.5
68.33
63.33
96*384
59.17
66.67
65.83
63.33
73.33
71.67
64*256
65.83
66.67
70.83
73.33
75
72.5
32*128
74.17
73.33
76.67
75
75.83
75
16*64
70.83
71.67
71.67
70.83
78.33
76.67
8*32
70
70.83
71.67
66.67
75
77.5
191
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
Dr. H. B. Kekre et. al.(IJCSIS) International Journal of Computer Science and Information Security, Vol. 8, No. 6, 2010
Table XV show the sentencewise identification rate when
WALSH transform is applied to image blocks using Set B.
From Table IVX and Table XV, it can be seen that, as the
number of coefficients used for identification purpose
decreases, the identification rate increases, reaches its
maximum value and then again decreases or remains constant.
Table XVI summarizes overall identification rate for both
training sets for various partial feature vectors.
Table xv. Identification rate for sentences sIto s6for varying
PORTION OF FEATURE VECTOR USING WALSHON IMAGE BLOCKS WITH
IMAGES FROM SET B
Portion of feature
vector selected
Sentence
SI
S2
S3
S4
S5
S6
128*512
63.33
66.67
75
66.67
76.67
76.67
96*384
75
71.67
76.67
73.33
78.33
81.67
64*256
80
75
78.33
83.33
81.67
81.67
32*128
86.67
83.33
81.67
85
83.33
85
16*64
86.67
81.67
81.67
88.33
83.33
91.67
8*32
86.67
85
83.33
85
83.33
86.67
Table XVI. Comparison of overall identification rate for varying
SIZE OF FEATURE VECTOR PORTION USING WALSHON IMAGE BLOCKS FOR
IMAGES IN SET A AND SET B
Portion of
feature vector
selected
Number of
WALSH
coefficients
Identification rate (%)
Set A
SetB
128*512
65536
60
70.83
96*384
36864
66.67
76.11
64*256
16384
70.69
80
32*128
4096
75
84.16
16*64
1024
73.33
85.55
8*32
256
71.94
85
It can be observed from Table XVI that the maximum
identification rate in case of Set A is obtained for 4096
WALSH coefficients i.e. for partial feature vector of size
32*128. The maximum identification rate in case of Set B is
obtained for 1024 WALSH coefficients i.e. for partial feature
vector of size 16*64.
used for identification purpose decreases, the identification rate
increases, reaches some peak value and then again decreases or
remains constant.
Table xviii. Identification rate for sentences s 1 to s 6 for full and
partial feature vector using haar on image blocks using set b
Portion of
feature vector
selected
Sentence
SI
S2
S3
S4
S5
S6
128*512
63.33
66.67
75
66.67
76.67
76.67
96*384
73.33
70
78.33
76.67
78.33
81.67
64*256
80
75
78.33
83.33
81.67
81.67
32*128
86.67
83.33
81.67
85
83.33
85
16*64
86.67
81.67
81.67
88.33
83.33
91.67
8*32
86.67
85
83.33
85
83.33
86.67
Table XIX shows overall identification rate for both
training sets obtained by considering the identification rate for
each sentence for various partial feature vectors. For Set A, the
maximum identification rate of 75.27% is obtained for 32*128
feature vector. Whereas, for Set B, the maximum identification
rate of 85.55% is obtained for 16*64 feature vector. Table XX
shows comparison of overall identification rates for all three
transformation techniques when applied on image blocks for
Set A and SetB.
Table XIX. Comparison of overall identification rate for varying
SIZE OF FEATURE VECTOR PORTION USING HAARON IMAGE BLOCKS USING
SETA AND SET B
Portion of feature
vector selected
Number of HAAR
coefficients
Identification rate (%)
Set A
SetB
128*512
65536
59.86
70.83
96*384
36864
65.97
76.39
64*256
16384
70.69
80
32*128
4096
75.27
84.44
16*64
1024
73.33
85.55
8*32
256
71.94
85.27
Table xx. Comparison of identification rates w hen dct, walsh
and haar are applied on image blocks for images in set a and set b
3) Results for HAAR on image blocks:
Table XVII shows identification rate for each sentence
when 2-D HAAR transform is applied on image blocks
obtained by dividing an image into four equal and non-
overlapping blocks as shown in Fig. 2. These results are for
training Set A.
Table xvii. Identification rate for sentences sIto s6for varying
portion of feature vector using haaron image blocks using set a
Portion of feature
vector selected
Sentence
SI
S2
S3
S4
S5
S6
128*512
53.33
59.17
57.5
57.5
68.33
63.33
96*384
58.33
61.67
65.83
68.33
72.5
69.17
64*256
65.83
66.67
70.83
73.33
75
72.5
32*128
74.17
73.33
76.67
76.67
75
75.83
16*64
70.83
71.67
71.67
70.83
78.33
76.67
8*32
70
70.83
71.67
66.67
75
77.5
Table XVIII shows results when Set B is used and 2-D
HAAR transform is applied on image blocks. For both the
training sets, it is observed that, as the number of coefficients
Portion of
feature
vector
selected
Identification rate (%)
When Set A is used
Identification rate (%)
When Set B is used
DCT
WALSH
HAAR
DCT
WALSH
HAAR
128*512
60
60
59.86
70.83
70.83
70.83
96*384
65.97
66.67
65.97
75
76.11
76.39
64*256
68.88
70.69
70.69
78.61
80
80
32*128
72.63
75
75.27
82.22
84.16
84.44
16*64
76.66
73.33
73.33
88.33
85.55
85.55
8*32
74.58
71.94
71.94
86.67
85
85.27
C. Results for DCT/ WALSH/ HAAR on Row Mean of an
image and Row Mean of image blocks :
1) Results for DCT on Row Mean of an image :
Table XXI shows sentence wise results obtained for Set A
when DCT of Row Mean is taken by dividing an image into
different number of non-overlapping blocks.
192
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
Dr. H. B. Kekre et. al.(IJCSIS) International Journal of Computer Science and Information Security, Vol. 8, No. 6, 2010
Table xxi. Identification rate for sentences si to s6for DCT on
ROW MEAN OF AN IMAGE WHEN IMAGE IS DIVIDED INTO DIFFERENT NUMBER
OF NON-OVERLAPPING BLOCKS USING SET A
No. of blocks for
image split
Sentence
SI
S2
S3
S4
S5
S6
Full image
(256*256)
57.5
66.66
64.16
60.83
60.83
62.5
4 Blocks
(128*128)
60.83
70.83
63.33
65.83
70
65.83
16 Blocks
(64*64)
69.16
75.83
70.83
65.83
73.33
71.66
64 Blocks
(32*32)
75
76.66
75.83
70
78.83
75.83
256 Blocks
(16*16)
76.66
75
75.83
72.5
80
82.5
1024 Blocks
(8*8)
74.16
72.5
75
72.5
80.83
78.33
It can be seen from the Table XXI that, as the block size
chosen for calculating Row Mean reduces, better identification
rate is achieved. For block size 16*16, maximum identification
rate is obtained and then it decreases again.
Table XXII shows the identification rate for sentence si to
s6 and Set B of images. It can be seen from the Table XXII
that, as the block size chosen for calculating Row Mean
reduces, better identification rate is achieved. For block size
16*16, maximum identification rate is obtained and then it
decreases again.
Table xxii. Identification rate for sentences sIto s6for DCT on
ROW MEAN OF AN IMAGE WHEN IMAGE IS DIVIDED INTO DIFFERENT NUMBER
OF NON-OVERLAPPING BLOCKS USING SET B
No. of blocks for image
split
Sentence
SI
S2
S3
S4
S5
S6
Full image (256*256)
73.33
76.67
78.33
76.67
75
80
4 Blocks (128*128)
80
80
78.33
81.67
81.67
80
16 Blocks (64*64)
91.67
81.67
83.33
83.33
81.67
83.33
64 Blocks (32*32)
91.67
85
86.67
86.67
86.67
88.33
256 Blocks (16*16)
91.67
88.33
88.33
85
91.67
90
1024 Blocks (8*8)
88.33
83.33
85
86.67
85
88.33
The overall identification rates for both the sets, when DCT
of Row Mean is taken by dividing an image into different
number of non-overlapping blocks are tabulated in Table
XXIII.
Table xxiii. Comparison of overall identification rate for DCT on
ROW MEAN OF AN IMAGE WHEN IMAGE IS DIVIDED INTO DIFFERENT NUMBER
OF NON-OVERLAPPING BLOCKS WITH SET A AND SET B
No. of blocks for image
split
Number of
DCT
coefficients
Identification rate
(%)
For Set A
For Set B
Full image (256*256)
256
62.08
76.67
4 Blocks 128*128)
512
66.11
80.27
16 Blocks (64*64)
1024
71.11
84.17
64 Blocks (32*32)
2048
75.27
87.5
256 Blocks 16*16)
4096
77.08
89.17
1024 Blocks (8*8)
8192
75.55
86.11
2) Results forWALSH on Row Mean of an image :
Table IVXX shows the sentence wise identification rate
when Walsh transform is applied to Row Mean of an image
when it is divided into different number of non-overlapping
and Set A is used. Table XXV shows the sentence wise
identification rate when Walsh transform is applied to Row
Mean of an image when it is divided into different number of
non-overlapping and Set B.
Table ivxx. Identification rate for sentences si to s6 for Walsh
TRANSFORM ON ROW MEAN OF AN IMAGE WHEN IMAGE IS DIVIDED INTO
DIFFERENT NUMBER OF NON-OVERLAPPING BLOCKS WITH SET A
No. of blocks for image
split
Sentence
SI
S2
S3
S4
S5
S6
Full image (256*256)
57.5
66.66
64.16
60.83
60.83
62.5
4 Blocks (128*128)
60.83
70.83
63.33
65.83
70
65.83
16 Blocks (64*64)
69.16
75.83
70.83
65.83
73.33
71.66
64 Blocks (32*32)
75
76.66
75.83
70
78.83
75.83
256 Blocks (16*16)
76.66
75
75.83
72.5
80
82.5
1024 Blocks (8*8)
74.16
72.5
75
72.5
80.83
78.33
Table xxv. Identification rate for sentences si to s6 for Walsh
TRANSFORM ON ROW MEAN OF AN IMAGE WHEN IMAGE IS DIVIDED INTO
DIFFERENT NUMBER OF NON-OVERLAPPING BLOCKS WITH SET B
No. of blocks for image
split
Sentence
SI
S2
S3
S4
S5
S6
Full image (256*256)
73.33
76.66
78.33
76.66
75
80
4 Blocks (128*128)
80
80
78.33
81.67
81.67
80
16 Blocks (64*64)
91.67
81.67
83.33
83.33
81.67
83.33
64 Blocks (32*32)
91.67
85
86.66
86.66
86.66
88.33
256 Blocks (16*16)
91.67
88.33
88.33
85
91.67
90
1024 Blocks (8*8)
88.33
83.33
85
86.66
85
88.33
It can be seen from the Table IVXX and Table XXV that,
as the block size chosen for calculating Row Mean reduces,
better identification rate is achieved.
Table XXVI summarizes overall identification rate for both
training sets by considering all six sentences. For block size
16*16, maximum identification rate is obtained and then it
decreases again.
Table xxvi. Comparison of overall identification rate for Walsh
transform on row mean of an image when image is divided into
different number of non-overlapping blocks for set a and set b
No. of blocks for
image split
Number of Walsh
coefficients
Identification rate
(%)
Set A
SetB
Full image (256*256)
256
62.08
76.67
4 Blocks (128*128)
512
66.11
80.27
16 Blocks (64*64)
1024
71.11
84.17
64 Blocks (32*32)
2048
75.27
87.5
256 Blocks (16*16)
4096
77.08
89.17
1024 Blocks (8*8)
8192
75.55
86.11
3) Results for HAAR on Row Mean of an image :
Table XXVII shows identification rate for each sentence
when 1-D HAAR transform is applied to Row Mean of an
256*256 image and when image is divided into different
number of non-overlapping blocks for Set A. Similarly, Table
XXVIII shows identification rate for each sentence when 1-D
HAAR transform is applied to Row Mean of an 256*256 image
193
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
Dr. H. B. Kekre et. al.(IJCSIS) International Journal of Computer Science and Information Security, Vol. 8, No. 6, 2010
and when image is divided into different number of non-
overlapping blocks for Set B.
Table xxvii. Identification rate for sentences s 1 to s6 for
haar transform on row mean of an image when image is divided
into different number of non-overlapping blocks using set a
No. of blocks for image
split
Sentence
SI
S2
S3
S4
S5
S6
Full image (256*256)
57.5
66.66
64.16
60.83
60.83
62.5
4 Blocks (128*128)
60.83
70.83
63.33
65.83
70
65.83
16 Blocks (64*64)
69.16
75.83
70.83
65.83
73.33
71.66
64 Blocks (32*32)
75
76.66
75.83
70
78.83
75.83
256 Blocks (16*16)
76.66
75
75.83
72.5
80
82.5
1024 Blocks (8*8)
74.16
72.5
75
72.5
80.83
78.33
Table xxviii. Identification rate for sentences sIto s6for
haar transform on row mean of an image when image is divided
into different number of non-overlapping blocks using set b
No. of blocks for
image split
Sentence
SI
S2
S3
S4
S5
S6
Full image
(256*256)
73.33
76.67
78.33
76.67
75
80
4 Blocks
(128*128)
80
80
78.33
81.67
81.67
80
16 Blocks
(64*64)
91.67
81.67
83.33
83.33
81.67
83.33
64 Blocks
(32*32)
91.67
85
86.67
86.67
86.67
88.33
256 Blocks
(16*16)
91.67
88.33
88.33
85
91.67
90
1024 Blocks
(8*8)
88.33
83.33
85
86.67
85
88.33
Table XXIX shows overall identification rate for the two
training sets when 1-D HAAR transform is applied to an image
divided into different number of equal and non-overlapping
blocks.
Table xxix. Comparison of overall identification rate for
haar transform on row mean of an image when image is divided
into different number of non-overlapping blocks for set a and b.
No. of blocks for
image split
Number of
HAAR
coefficients
Identification rate (%)
For Set A
For Set B
Full image
(256*256)
256
62.08
76.67
4 Blocks
(128*128)
512
66.11
80.27
16 Blocks
(64*64)
1024
71.11
84.17
64 Blocks
(32*32)
2048
75.27
87.5
256 Blocks
(16*16)
4096
77.08
89.17
1024 Blocks
(8*8)
8192
75.55
86.11
Overall identification rate for DCT, WALSH and HAAR
on Row Mean of an image and image blocks are summarized in
the Table XXX.
Table xxx. Comparison of DCT, WALSH and HAAR on Row
Mean of image and image blocks
No. of
blocks for
image
split
Identification rate (%)
when Set A is used
Identification rate (%)
when Set B is used
DCT
WALSH
HAAR
DCT
WALSH
HAAR
Full image
(256*256)
62.08
62.08
62.08
76.67
76.67
76.67
4 Blocks
(128*128)
66.11
66.11
66.11
80.27
80.27
80.27
16 Blocks
(64*64)
71.11
71.11
71.11
84.17
84.17
84.17
64 Blocks
(32*32)
75.27
75.27
75.27
87.5
87.5
87.5
256
Blocks
(16*16)
77.08
77.08
77.08
89.17
89.17
89.17
1024
Blocks
(8*8)
75.55
75.55
75.55
86.11
86.11
86.11
V. COMPLEXITY ANALYSIS
A
Complexity analysis of DCT, WALSH and HAAR on full
image:
For 2-D DCT on N*N image, 2N 3 multiplications are
required and 2N 2 (N-1) additions are required. For 2-D WALSH
on N*N image, 2N (N-l) additions are required. For 2-D
HAAR transform on N*N image where N=2m, number of
multiplications required are 2(m+l)N 2 and number of additions
required are 2mN 2 . Table XXXI summarizes these details
along with actual values of mathematical computations needed
for processing of 256*256 images.
Table xxxi. Comparison between DCT, WALSHand HAAR
WITH RESPECT TO MATHEMATICAL COMPUTATIONS AND IDENTIFICATION RATE
WHEN APPLIED ON FULL IMAGE
Parameter
Algorithm
DCT on full
image(N*N)
WALSH on full
image(N*N)
HAAR on
full image
(N*N)
Number of
Multiplications
2N 3
2(m+l)N 2
N=256
33554432
1179648
Number of
Additions
2N 2 (N-1)
2N 2 (N-1)
2mN 2
N=256
33423360
33423360
1048576
Identification rate
(%) for Set A
77.63
75
75
Identification rate
(%) for Set B
88.05
85.55
86.39
From the above table, it can be seen that DCT on full image
gives the highest identification rate for both the training sets as
compared to WALSH and HAAR on full image. However this
outstanding performance is achieved at the expense of higher
computations.
Number of multiplications required by DCT on full image
is approximately 28 times more than the number of
multiplications required by HAAR on full image. Whereas
number of additions required by DCT on full image is
approximately 31 times more than the number of additions
required by HAAR on full image.
194
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
Dr. H. B. Kekre et. al.(IJCSIS) International Journal of Computer Science and Information Security, Vol. 8, No. 6, 2010
Though WALSH on full image does not require any
multiplications, overall CPU time taken by it is more than that
of HAAR on full image. This is because the number of
additions taken by WALSH on full image is approximately 31
times more than the number of additions required by HAAR on
full image.
B. Complexity analysis ofDCT, WALSH and HAAR on
image blocks:
The number of multiplications required in case of 2-D DCT
on image blocks is N 3 and the number of additions required are
N 2 (N-2).
For 2-D WALSH on four image blocks of size N/2*N/2,
number of additions required are N 2 (N-2).
The number of multiplications required for 2-D HAAR on
image blocks is 2mN 2 . Similarly number of additions required
for 2-D HAAR on image blocks is 2(m-l)N 2 . Table XXXII
summarizes these details along with actual values of
mathematical computations needed for processing of 256*256
images.
Table xxxii Comparison between DCT, WALSH and HAAR with
RESPECT TO MATHEMATICAL COMPUTATIONS AND IDENTIFICATION RATE
WHEN APPLIED ON IMAGE BLOCKS
Parameter
Algorithm
DCTon
image blocks
(N/2*N/2)
Walsh on
image
blocks
(N/2*N/2)
HAAR on
image
blocks
(N/2*N/2)
Number of
Multiplications
N 3
2(m+l)N 2
N=256, four blocks
16777216
1048576
Number of Additions
N 2 (N-2)
N 2 (N-2)
2(m-l)N 2
N=256, four blocks
16646144
16646144
917504
Identification rate (%)
for Set A
76.66
75
75.27
Identification rate (%)
for Set B
88.33
85.55
85.55
From the Table XXXII, it can be seen that, in all the three
transformation techniques on image blocks, DCT on image
blocks gives best identification rate for both the training sets.
But this performance is achieved at the expense of higher
number of computations. DCT on image blocks takes 16 times
more multiplications and 18 times more additions than HAAR
on image blocks. Though WALSH transform does not need
any multiplications, still it takes more number of computations
than HAAR. This is because WALSH on image blocks requires
approximately 18 times more additions than HAAR on image
blocks.
C. Complexity Analysis of DCT, Walsh and HAAR on Row
Mean of an image and on Row Mean of image blocks:
Since Row Mean of an image is a one dimensional vector,
only 1-D DCT, WALSH and HAAR need to be applied on
Row Mean. This itself reduces the number of multiplications
and additions required for feature vector calculation. Row
Mean of an image of size N*N is a vector of size N*l. For 1-D
DCT on this N*l vector, N 2 multiplications and N(N-l)
additions are required. One dimensional Walsh on Row Mean
of an image takes N(N-l) additions and no multiplications.
Whereas, 1-D HAAR on Row Mean of an image of size N*N
requires (m+l)N multiplications and mN additions where
N=2 m . Following Table XXXIII summarizes this statistics in
case of each transformation technique applied for the Row
Mean of block size 16*16 which gives highest identification
rate.
Table xxxiii. Comparison between DCT, WALSH and HAAR
WITH RESPECT TO MATHEMATICAL COMPUTATIONS AND IDENTIFICATION RATE
WHEN APPLIED ON ROW MEAN OF AN IMAGE
Parameter
Algorithm
DCTon
Row Mean
of image
(N*l)
Walsh on
Row
Mean of
image
(N*l)
HAAR on
Row Mean
of image
(N*l)
Number of
Multiplications
N 2
(m+l)N
N=16, 256 blocks
65536
20480
Number of Additions
N(N-l)
N(N-l)
mN
N=16, 256 blocks
61440
61440
16384
Identification rate (%)
for Set A
77.08
77.08
77.08
Identification rate (%)
for Set B
89.17
89.17
89.17
From the Table XXXIII, we can see that all three
transformation techniques result in same identification rate
when applied on the Row Mean of an image and on Row Mean
of an image blocks. For both the training sets, highest
identification rate is obtained when image is divided into 16*16
size blocks. However, in terms of computations, HAAR
transform is proved to be better one. Number of multiplications
required by HAAR is approximately three times less than the
number of multiplications required in case of DCT on Row
Mean. Also the number of additions required by HAAR is 3.5
times less than the number of additions required by DCT and
WALSH on Row Mean of an image.
Along with the different approaches of applying
transformation techniques on spectrograms, comparative study
of computational complexity of three transformation
techniques for each approach has been done and is presented
below.
D. Complexity Analysis of DCT transform on Full, Block and
Row Mean of Spectrograms:
For 2-D DCT on N*N image, 2N 3 multiplications are
required and 2N 2 (N-1) additions are required. For 2-D DCT on
four blocks of size N/2*N/2, N 3 multiplications are required
and N 2 (N-2) additions are required. For 1-D DCT on N*l
image, N 2 multiplications are needed and N(N-l) additions are
needed. These computational details are summarized in Table
XXXIV along with the actual number of computations for
256*256 image using three methods of applying DCT.
195
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
Dr. H. B. Kekre et. al.(IJCSIS) International Journal of Computer Science and Information Security, Vol. 8, No. 6, 2010
Table xxxiv. Computational details for 2-DDCTon N*N
IMAGE, 2-D DCT ON N/2*N/2 IMAGE AND 1-DCT ON N*l IMAGE
RESPECTIVELY
Parameter — >
No. of
Multiplications
No. of Additions
Algorithm j
2-D DCT on N*N image
2N 3
2N 2 (N-1)
2-D DCT on 256*256
image
33554832
33424159
2-D DCT on four blocks of
size N/2*N/2 each
N 3
N 2 (N-2)
2-D DCT on four blocks of
size 256/2*256/2 each
16778240
16648191
ID DCT on N*l Row
Mean of N*N image
N 2
N(N-l)
ID DCT on N*l Row
Mean of 256*1 image
69632
69631
When all three methods of applying DCT are compared, it
has been observed that though number of coefficients used in
Row Mean method is higher, number of multiplications and
additions reduce drastically as compared to other two methods.
Number of multiplications in DCT on full image method is 480
times more than the number of multiplications in Row Mean
method whereas for DCT on image blocks it is 241 times more.
Number of additions needed in DCT on full image and DCT on
image blocks are also 480 times and 239 times more than the
additions required in Row mean method respectively. For the
Set A, the identification rate for DCT on Row Mean is almost
same as identification rate for DCT on full image. In case of
Set B, DCT on Row Mean gives better identification rate as
compared to DCT on full image and DCT on image blocks and
that too with reduced number of mathematical computations.
E. Complexity Analysis of WALSH transform on Full, Block
and Row Mean of Spectrograms:
For 2-D WALSH on N*N image, 2N 2 (N-1) additions are
required. For 2-D WALSH on four blocks of size N/2*N/2,
N 2 (N-1) additions are required. Whereas for 1-D WALSH on
N*l image, N(N-l) additions are needed as shown in table
XXXV. In all three cases number of multiplications required is
zero.
Table xxxv. Computational details for 2-D WALSHon N*N
IMAGE, 2-D WALSH ON N/2*N/2 IMAGE AND 1-D WALSH ON N*l IMAGE
RESPECTIVELY
Parameter — >
No. of
Multiplications
No. of
Additions
Algorithm J,
2-D WALSH on N*N image
2N 2 (N-1)
2-D WALSH on 256*256
image
33423360
2-D WALSH on four blocks of
size N/2*N/2 each
N 2 (N-2)
2-D WALSH on four blocks of
size 256/2*256/2 each
16646144
1-D WALSH on N*l size Row
Mean vector of image N*N
N(N-l)
1-D WALSH on 256*1 size
Row Mean vector of image
256*1
65280
WALSH transform is applied to Row Mean of an image. Also
the number of additions required when WALSH transform is
applied on image blocks is 255 times more than the number of
additions required when WALSH transform is applied to Row
Mean of an image. Thus number of additions is drastically
reduced for Walsh transform on Row Mean of an image.
F. Complexity Analysis ofHAAR transform on Full, Block
and Row Mean of Spectrograms:
For 2-D HAAR transform on N*N image where N=2 m ,
number of multiplications required are 2(m+l)N 2 and number
of additions required are 2mN 2 . For 2-D HAAR transform on
four blocks of size N/2*N/2 each, 2mN 2 multiplications and
2(m-l)N 2 additions are needed. Whereas for 1-D HAAR
transform on N*l image, number of multiplications required
are (m+l)N and number of additions are mN as shown in table
XXXVI.
Table xxxvi. Computational details for 2-DHAARon N*N
IMAGE, 2-D HAAR ON N/2*N/2 IMAGE AND 1-D HAAR ON N*l IMAGE
RESPECTIVELY
Parameter — >
No. of
Multiplications
No. of Additions
Algorithm [
2-D HAAR on N*N image
2(m+l)N 2
2mN 2
2-D HAAR on 256*256
image
1179648
1048576
2-D HAAR on four blocks
of size N/2 *N/2 each
2mN 2
2(m-l)N 2
2-D HAAR on four blocks
of size 256/2*256/2 each
1048576
917504
1-D HAAR on N*l image
(m+l)N
mN
1-D HAAR on 256*1 size
Row Mean vector of
image 256*1
2304
2048
From table 7.5 it can be seen that number of additions
required when WALSH transform is applied on full image is
512 times more than the number of additions required when
From Table XXXVI, it can be seen that number of
multiplications required when HAAR transform is applied on
full image is 512 times more than the number of
multiplications required when HAAR transform is applied to
Row Mean of an image. Also the number of multiplications
required when HAAR transform is applied on image blocks is
455 times more than the number of multiplications required
when HAAR transform is applied to Row Mean of an image.
Thus number of multiplications is drastically reduced for
HAAR transform on Row Mean of an image. Number of
additions required is also reduced to a greater extent when
transformation technique is applied on Row Mean of an image.
Number of additions required when HAAR transform is
applied on full image is 512 times more than the number of
additions required when HAAR transform is applied to Row
Mean of an image. Also the number of additions required when
HAAR transform is applied on image blocks is 448 times more
than the number of additions required when HAAR transform
is applied to Row Mean of an image.
VI.COnclusion
In this paper, closed set text dependent speaker
identification has been considered using three different
transformation techniques namely DCT, WALSH and HAAR.
Each transformation technique is applied in three ways:
a) On full image
196
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
Dr. H. B. Kekre et. al.(IJCSIS) International Journal of Computer Science and Information Security, Vol. 8, No. 6, 2010
b) On image blocks and
c) On Row Mean of an image.
For each method, two training sets were used as mentioned
earlier.
It can be clearly concluded from the results that as more
training is provided to the system, more accuracy is obtained in
the results in terms of identification rate.
Further for each method, Identification rates are obtained
for various numbers of coefficients from feature vectors of
images. It has been observed that as the number of coefficients
chosen is smaller up to a certain limit; better identification rate
is achieved in all three methods.
DCT on full image gives its best identification rate for only
20*20 portion of feature vector i.e. by using only 400 DCT
coefficients. DCT on image blocks gives highest identification
rate when 16*64 portion of its feature vector is considered
which has 1024 DCT coefficients. Finally DCT on Row Mean
gives highest identification rate for Row Mean of 16*16 size
image blocks i.e. for 4096 DCT coefficients. When these
highest identification rates in all three methods in DCT are
compared, it has been observed that DCT on image blocks
gives slightly improved results for training set of eight images
per speaker. Whereas, DCT on Row Mean, further improves
these results with drastically reduced computations. Though the
number of coefficients used in Row Mean method is higher,
overhead caused for its comparison is negligible as compared
to number of mathematical computations needed in other two
approaches.
Similarly, WALSH on Row Mean of image gives better
identification rates as compared to WALSH on full image and
WALSH on image blocks for both the training sets. These
better identification rates are obtained with the advantage of
reduced mathematical computations. For HAAR transform
also, identification rate for HAAR on Row Mean is better than
HAAR on full image and HAAR on image blocks.
From the results of DCT, WALSH and HAAR on full
image, it can be concluded that DCT on full image gives better
identification rate than WALSH and HAAR on full image but
at the expense of large number of mathematical computations.
In WALSH transform on full image, numbers of mathematical
computations required are greatly reduced as compared to DCT
since no multiplications are required in WALSH. These
computations are further reduced by use of HAAR transform
but at the slight expense of identification rate. Similar
conclusions can be drawn for DCT, WALSH and HAAR on
image blocks. So there is a trade off between better
identification rate and less CPU time for mathematical
computations.
However, in case of Row Mean approach of applying
transform, performances of all three transformation techniques
are same for a specific block size chosen for Row Mean. In that
HAAR transform proves to be better because it requires
minimum number of computations.
The overall conclusion is that Row Mean technique
requires less number of mathematical computations and hence
less CPU time for all three transformation techniques as
compared to transformation techniques on full image and on
image blocks. HAAR transform on Row Mean of an image
gives the best result with respect to identification rate as well as
number of computations required.
REFERENCES
[I] Evgeniy Gabrilovich, Alberto D. Berstin: "Speaker recognition: using a
vector quantization approach for robust text-independent speaker
identification", Technical report DSPG-95-9-001', September 1995.
[2] Tridibesh Dutta, "Text dependent speaker identification based on
spectrograms", Proceedings of Image and vision computing, pp. 238-
243, New Zealand 2007,.
[3] J.P.Campbell, "Speaker recognition: a tutorial", Proc. IEEE, vol. 85, no.
9, pp. 1437-1462, 1997.
[4] D. O'Shaughnessy, "Speech communications- Man and Machine", New
York, IEEE Press, 2nd Ed., pp. 199, pp. 437-458, 2000.
[5] S. Davis and P. Mermelstein, "Comparison of parametric representations
for monosyllabic word recognition in continuously spoken sentences,"
IEEE Transaction Acoustics Speech and Signal Processing, vol. 4, pp.
375-366, 1980.
[6] Wang Yutai, Li Bo, Jiang Xiaoqing, Liu Feng, Wang Lihao, "Speaker
Recognition Based on Dynamic MFCC Parameters", International
Conference on Image Analysis and Signal Processing, pp. 406-409, 2009
[7] Azzam Sleit, Sami Serhan, and Loai Nemir, "A histogram based speaker
identification technique", International Conference on ICADIWT, pp.
384-388, May 2008.
[8] B. S. Atal, "Automatic Recognition of speakers from their voices", Proc.
IEEE, vol. 64, pp. 460-475, 1976.
[9] Jialong He, Li Liu, and G"unther Palm, "A discriminative training
algorithm for VQ-based speaker Identification", IEEE Transactions on
speech and audio processing, vol. 7, No. 3, pp. 353-356, May 1999.
[10] Debadatta Pati, S. R. Mahadeva Prasanna, "Non-Parametric Vector
Quantization of Excitation Source Information for Speaker
Recognition", IEEE Region 10 Conference, pp. 1-4, Nov. 2008.
[II] Tridibesh Dutta and Gopal K. Basak, "Text dependent speaker
identification using similar patterns in spectrograms", PRIP'2007
Proceedings, Volume 1, pp. 87-92, Minsk, 2007.
[12] Andrew B. Watson, "Image compression using the Discrete Cosine
Transform", Mathematica journal, 4(1), pp. 81-88, 1994,.
[13] http://www.itee.uq.edu.au/~conrad/vidtimit/
[14] http://www2.imm.dtu.dkMf/elsdsr/
[15] H.B.Kekre, Sudeep D. Thepade, "Improving the Performance of Image
Retrieval using Partial Coefficients of Transformed Image",
International Journal of Information Retrieval (IJIR), Serials
Publications, Volume 2, Issue 1, pp. 72-79 (ISSN: 0974-6285), 2009.
[16] H.B.Kekre, Tanuja Sarode, Sudeep D. Thepade, "DCT Applied to Row
Mean and Column Vectors in Fingerprint Identification", In Proceedings
of International Conference on Computer Networks and Security
(ICCNS), 27-28 Sept. 2008, VIT, Pune.
[17] H.B.Kekre, Sudeep D. Thepade, Archana Athawale, Anant Shah,
Prathmesh Verlekar, Suraj Shirke,"Energy Compaction and Image
Splitting for Image Retrieval using Kekre Transform over Row and
Column Feature Vectors", International Journal of Computer Science
and Network Security (IJCSNS),Volume:10, Number 1, January 2010,
(ISSN: 1738-7906) Available at www.IJCSNS.org.
[18] H.B.Kekre, Sudeep D. Thepade, Archana Athawale, Anant Shah,
Prathmesh Verlekar, Suraj Shirke, "Performance Evaluation of Image
Retrieval using Energy Compaction and Image Tiling over DCT Row
Mean and DCT Column Mean", Springer-International Conference on
Contours of Computing Technology (Thinkquest-2010), Babasaheb
Gawde Institute of Technology, Mumbai, 13-14 March 2010, The paper
will be uploaded on online Springerlink.
[19] H.B.Kekre, Tanuja K. Sarode, Sudeep D. Thepade, Vaishali
Suryavanshi, "Improved Texture Feature Based Image Retrieval using
Kekre's Fast Codebook Generation Algorithm", Springer-International
Conference on Contours of Computing Technology (Thinkquest-2010),
197
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
Dr. H. B. Kekre et. al.(IJCSIS) International Journal of Computer Science and Information Security, Vol. 8, No. 6, 2010
Babasaheb Gawde Institute of Technology, Mumbai, 13-14 March 2010,
The paper will be uploaded on online Springerlink.
[20] H. B. Kekre, Tanuja K. Sarode, Sudeep D. Thepade, "Image Retrieval
by Kekre 's Transform Applied on Each Row of Walsh Transformed VQ
Codebook", (Invited), ACM-International Conference and Workshop on
Emerging Trends in Technology (ICWET 2010),Thakur College of
Engg. And Tech., Mumbai, 26-27 Feb 2010, The paper is invited at
ICWET 2010. Also will be uploaded on online ACM Portal.
[21] H. B. Kekre, Tanuja Sarode, Sudeep D. Thepade, "Color-Texture
Feature based Image Retrieval using DCT applied on Kekre 's Median
Codebook", International Journal on Imaging (IJI), Volume 2, Number
A09, Autumn 2009,pp. 55-65. Available online at
www.ceser.res.in/iji.html (ISSN: 0974-0627).
[22] H. B. Kekre, Ms. Tanuja K. Sarode, Sudeep D. Thepade, "Image
Retrieval using Color-Texture Features from DCT on VQ Codevectors
obtained by Kekre's Fast Codebook Generation", ICGST-International
Journal on Graphics, Vision and Image Processing (GVIP), Volume 9,
Issue 5, pp.: 1-8, September 2009. Available online at http:
//www.icgst.com /gvip /Volume9 /Issue5 /P1150921752.html.
[23] H. B. Kekre, Sudeep Thepade, Akshay Maloo, "Image Retrieval using
Fractional Coefficients of Transformed Image using DCT and Walsh
Transform", International Journal of Engineering Science and
Technology, Vol.. 2, No. 4, 2010, 362-371
[24] H. B. Kekre, Sudeep Thepade, Akshay Maloo, "Performance
Comparison of Image Retrieval Using Fractional Coefficients of
Transformed Image Using DCT, Walsh, Haar and Kekre's Transform",
CSC-International Journal of Image processing (DIP), Vol.. 4, No. 2,
pp.:142-155, May 2010.
[25] H. B. Kekre, Tanuja Sarode "Two Level Vector Quantization Method
for Codebook Generation using Kekre's Proportionate Error Algorithm"
, CSC-International Journal of Image Processing, Vol.4, Issue 1, pp.l-
10, January-February 2010
[26] H. B. Kekre, Sudeep Thepade, Akshay Maloo, "Eigenvectors of
Covariance Matrix using Row Mean and Column Mean Sequences for
Face Recognition", CSC-International Journal of Biometrics and
Bioinformatics (IJBB), Volume (4): Issue (2), pp. 42-50, May 2010.
[27] H. B. Kekre, Tanuja Sarode, Shachi Natu, Prachi Natu, "Performance
Comparison Of 2-D DCT On Full/Block Spectrogram And 1-D DCT On
Row Mean Of Spectrogram For Speaker Identification", (Selected),
CSC-International Journal of Biometrics and Bioinformatics (IJBB),
Volume (4): Issue (3), pp. 100-112, August 2010, Malaysia..
[28] H. B. Kekre, Tanuja Sarode, Shachi Natu, Prachi Natu, "Performance
Comparison of Speaker Identification Using DCT, Walsh, Haar On Full
And Row Mean Of Spectrogram" , (Selected), International Journal of
Computer Applications, pp. 30-37, August 2010, USA.
[29] H. B. Kekre, Tanuja Sarode, Shachi Natu, Prachi Natu, "Speaker
Identification using 2-d DCT, Walsh and Haar on Full and Block
Spectrograms", International Journal of Computer Science and
Engineering, Volume 2, Issue 5, pp. 1733-1740, August 2010.
AUTHORS PROFILE
Dr. H. B. Kekre has received B.E. (Hons.) in Telecomm.
FEngg. from Jabalpur University in 1958,
M.Tech (Industrial Electronics) from IIT
Bombay in 1960, M.S.Engg. (Electrical Engg.)
from University of Ottawa in 1965 and Ph.D.
(System Identification) from IIT Bombay in
1970. He has worked Over 35 years as Faculty of Electrical
Engineering and then HOD Computer Science and Engg. at
IIT Bombay. For last 13 years worked as a Professor in
Department of Computer Engg. at Thadomal Shahani
Engineering College, Mumbai. He is currently Senior
Professor working with Mukesh Patel School of Technology
Management and Engineering, SVKM's NMIMS University,
Vile Parle (w), Mumbai, INDIA. He ha guided 17 Ph.D.s, 150
M.E./M.Tech Projects and several B.E./B.Tech Projects. His
areas of interest are Digital Signal processing, Image
Processing and Computer Networks. He has more than 250
papers in National / International Conferences / Journals to his
credit. Recently six students working under his guidance have
received best paper awards. Currently he is guiding ten Ph.D.
students.
Dr. Tanuja K. Sarode has received M.E. (Computer
Engineering) degree from Mumbai University
in 2004 and Ph.D. from Mukesh Patel School
of Technology, Management and Engg.
SVKM's NMIMS University, Vile-Parle (W),
Mumbai, INDIA, in 2010. She has more than
10 years of experience in teaching. Currently
working as Assistant Professor in Dept. of Computer
Engineering at Thadomal Shahani Engineering College,
Mumbai. She is member of International Association of
Engineers (IAENG) and International Association of
Computer Science and Information Technology (IACSIT).
Her areas of interest are Image Processing, Signal Processing
and Computer Graphics. She has 70 papers in National
/International Conferences/journal to her credit.
Shachi Natu has received B.E. (Computer) degree from
Mumbai University with first class in 2004.
Currently Purusing M.E. in Computer
Engineering from University of Mumbai. She
has 05 years of experience in teaching.
Currently working as Lecturer in department
' of Information Technology at Thadomal
Shahani Engineering College, Bandra (w), Mumbai. Her areas
of interest are Image Processing, Data Structure, Database
Management Systems and operating systems. She has 3 papers
in National / International Conferences /journal to her credit.
Prachi Natu has received B.E. (Electronics and
Telecommunication) degree from Mumbai
University with first class in 2004. Currently
Purusing M.E. in Computer Engineering from
University of Mumbai. She has 04 years of
experience in teaching. Currently working as
Lecturer in Computer Engineering department
at G. V. Acharya Institute of Engineering and Technology,
Shelu. Mumbai. Her areas of interest are Image Processing,
Database Management Systems and operating systems. She
has 3 papers in National / International Conferences /journal to
her credit.
198
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
A Research Proposal for Mitigating DoS Attacks
in IP-based Networks
Sakharam Lokhande
Assistant Professor
School of Computational Science,
Swami Ramanand Teerth Marathwada University, Nanded,
MS, India, 431606.
Dr. Santosh Khamitkar
Associate Professor
School of Computational Science,
Swami Ramanand Teerth Marathwada University, Nanded,
MS, India, 431606.
Parag Bhalchandra
Assistant Professor
School of Computational Science,
Swami Ramanand Teerth Marathwada University, Nanded,
MS, India, 431606.
Nilesh Deshmukh
Assistant Professor
School of Computational Science,
Swami Ramanand Teerth Marathwada University, Nanded,
MS, India, 431606.
Santosh Phulari
Assistant Professor
School of Computational Science, Swami Ramanand Teerth
Marathwada University, Nanded,
MS, India, 431606.
Ravindra Rathod
Assistant Professor
School of Computational Science, Swami Ramanand Teerth
Marathwada University, Nanded,
MS, India, 431606.
Abstract : This paper studies denial of service (DoS) attacks in
computer networks. These attacks are known for preventing
availability of network services from their legitimate users. After
careful review of literature, we wish to presents a structured view
on possible attack and defense mechanisms. An outline to
describe some new defense mechanisms is also presented in terms
of a research proposal .
Keywords- Denial of Service Attacks, Intrusion, Security
I. PROBLEM DEFINATION
Defending against DoS attacks is a task from network and
computer security. As scientific disciplines, network and
computer security are relatively primitive. An indication of
this fact is to be aware that the computer security terminology
is not yet stabilized [4]. Computer and network security
aspects were first studied in the early 1970s. As in some of the
earliest security papers listed and available in, the Denial of
Service attacks are timely and extremely important research
topic. According to the C SI/FBI computer crime and security
survey in the United States [1] for the year 2004, DoS attacks
are the second most widely detected outsider attack type in
computer networks, immediately after virus infections. A
computer crime and security survey in Australia[l] for the
year 2004, gives similar results. It is currently not possible to
prevent DoS attacks because many of these attacks are based
on using ordinary protocols and services in an overwhelming
manner. Specific security holes in the victim hosts or networks
are thus not necessarily needed. For this reason we can only
mitigate these attacks.
II. Overview of Denial of Service Attacks
Denials of Service (DoS) attacks have proved to be a
serious and permanent threat to users, organizations, and
infrastructures of the Internet [1]. The primary goal of these
attacks is to prevent access to a particular resource like a web
server [2]. A large number of defenses against DoS attacks
have been proposed in the literature, but none of them gives
reliable protection. There will always be vulnerable hosts in
the Internet to be used as sources of attack traffic. It is simply
not feasible to expect all existing hosts in the Internet to be
protected well enough. In addition, it is very difficult to
reliably recognize and filter only attack traffic without causing
any collateral damage to legitimate traffic.
A DoS attack can be carried out either as a flooding or a
logic attack. A Flooding DoS attack is based on brute force.
Real-looking but unnecessary data is sent as much as possible
to a victim. As a result, network bandwidth is wasted, disk
space is filled with unnecessary data (such as spam e-mail,
junk files, and intentional error messages), fixed size data
structures inside host software are filled with bogus
information, or processing power is spent for un useful
purposes. To amplify the effects, DoS attacks can be run in a
coordinated fashion from several sources at the same time
199
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
(Distributed DoS, DDoS).A logic DoS attack is based on an
intelligent exploitation of vulnerabilities in the target. For
example, a skillfully constructed fragmented Internet Protocol
(IP) datagram may crash a system due to a serious fault in
the operating system (OS) software. Another example of a
logic attack is to exploit missing authentication requirements
by injecting bogus routing information to prevent traffic from
reaching a victim's network. [5, 6]
There are two major reasons that make DoS attacks
attractive for attackers. The first reason is that there are
effective automatic tools available for attacking any victim, so
expertise is not necessarily required. The second reason is that
it is usually impossible to locate an attacker without extensive
human interaction or without new features in most routers of
the Internet. DoS attacks make use of vulnerabilities in end-
hosts, routers, and other systems connected to a computer
network. The size of a population having the same
vulnerability can be large. In July 2003 a vulnerability was
found from the whole population of Cisco routers and
switches running any version of the Cisco 10 S software and
configured to process Internet Protocol version 4 (IPv4)
packets. This vulnerability made it possible to block an
interface, which resulted in a DoS condition without any
alarms being triggered. Another example of a large population
is the Microsoft Windows Metafile (WMF) vulnerability
which was found in December 2005 from all versions of
Windows 98, 98SE, ME, 2000, and XP. This vulnerability
made it possible to install any malicious software on these
hosts, for example, to send DoS attack traffic. User interaction
was, however, required to exploit this vulnerability.
III. Research Problem
Mitigating DoS attacks is difficult especially due to the
following problems:
1) Very little has been done to compare, contrast, and
categorize the different ideas related to DoS attacks and
defenses. As a result it is difficult to understand what a
computer network user needs to do and why to mitigate the
threat from DoS attacks.
2) There are no effective defense mechanisms against
many important DoS attack types. There is no guidance on
how to select defense mechanisms.
3) Existing defense mechanisms have been evaluated
according to very limited criteria.
4) Often relevant risks have been ignored (such as in
[3]) or evaluations have been carried out under ideal
conditions.
5) No research publications exist for giving a
systematic list of issues related to defense evaluation
IV. Objective of the Research
The objective of this research proposal is to help any user
in any network for mitigating DoS attacks in IP-based
networks. This study concentrates especially on the following
areas:
1) One should understand existing attack mechanisms
and available defense mechanisms, and have a rough idea
about the benefits (best-case performance) of each defense
mechanism.
2) One should acknowledge possible situation
dependency of defense mechanisms, and be able to choose the
most suitable defense when more than one defense
mechanisms are available against a specific attack type.
3) One should evaluate defense mechanisms in a
comprehensive way, including both benefits and
disadvantages (worst-case performance), as an attacker can
exploit any weakness in a defense mechanism.
Knowledge of all of these issues is necessary in successful
mitigation of DoS attacks. Without knowing how a specific
defense mechanism works under different possible conditions
and what the real benefits and weaknesses are, it is not
possible to assure the suitability of a defense mechanism
against a certain type of a DoS attack.
V. Research Methodology
Research methodologies aimed to be used in this proposal,
are primarily based on simulating different attack scenarios,
but measurements, mathematical modeling based on game
theory, and requirement specification are also planned to be
used .
VI. Scope of the Research
Since this proposal studies DoS attacks in computer
networks using the Internet Protocol (IP), namely the Internet
and mobile ad hoc networks, is extremely useful for the
security concern. DoS attacks in the physical world will not be
studied here. Major work concentrate on the fixed (wired)
Internet, but most of the considered attack and defense
mechanisms will be applicable to wireless networks, too. The
emphasis of this research proposal is on DoS attacks in
general, and DDoS attacks are treated as a subset of DoS
attacks. DDoS attacks are based on the same mechanisms as
basic DoS attacks, but there is one exception during the
deployment phase .A DDoS tool needs to be installed on many
vulnerable hosts. The installation of DoS software on a single
vulnerable host is, however, a common prerequisite for most
DoS attacks. Thus attack and defense mechanisms described
in this dissertation are applicable to both DoS and DDoS
attacks.
VII. Possible Outcome
The main contributions of this proposed work include,
1) A comprehensive and well-structured description can
be given about what DoS attacks really are? How DoS attacks
can be carried out in IP networks? And how one can defend
against DoS attacks in IP networks. A good understanding of
existing attack mechanisms and available defense mechanisms
is a prerequisite for succeeding in mitigating these attacks cost
effectively.
2) An overview of an organized approach for selecting a
comprehensive set of defense mechanisms against DoS attacks
is given. This emphasizes the importance of basic security
mechanisms at every host in the Internet, the importance of
risk management in choosing additional defenses when basic
defenses are not enough, and the necessity of implementing
new defenses against such important DoS attacks for which
there are no existing defenses.
200
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
3) A new defense mechanism for protecting
organization-specific name servers will be described and
simulated.
4) Since knowledge about DoS and DDoS is in
primitive stage, we are hopeful to extend above objectives to
study DoS attack in mobile ad hoc networks. An earlier
attempt is found successful in some similar work [6].
CONCLUSION
This proposal aim to evaluate the DoS problems and
availability of defence mechanism. It is understood that the
existing defence mechanisms are mainly passive, in the sense
that the target host or network is impaired before the attack
source(s) can be found and controlled. We wish to propose a
novel concept of active defence against DoS attacks by
mitigating them in the Internet. This proposed style has
sufficient advantages over conventional passive defence
mechanisms. However, this is only the first step toward
realizing the secure Internet paradigm. The proposed work can
also be extended for designing of robust active defence
architecture, developing a sensitive and accurate surveillance
system, or for a powerful active trace back system and
deployment of such system in real Internet environment.
Authors
Dr. S.D.Khamitkar: He is PhD in computer science and has 15+ research
papers in International Conferences and journals. His interest area includes
ICT, Green computing and Network Security.
P.U.Bhalchandra , N.K. Deshmukh , S.N.Lokhande : These are SET-
NET qualified faculties and have 8+ years teaching experience. They have 5+
research papers in international conferences and journals. At present they are
also working on research related to ICT and Green computing. The present
paper is research topic of Mr. S.N.Lokhande
S.S.Phulari , R.P.Rathod : These are also faculties and have qualified
M.Phil in computer science . They have 2+ papers in international conferences
and journals.
REFERENCES
[1] L. Zhou and Z. Haas. Securing ad hoc networks. IEEE Network,
13(6):24~30, November/December 1999.
[2] Y. Zhang and W. Lee, "Intrusion detection in wireless ad hoc
networks," ACM MOBICOM, 2000.
[3] P.Papadimitratos and Z.J. Haas, "Secure Routing for Mobile Ad
Hoc Networks," SCS Communication Networks and Distributed
Systems Modeling and Simulation Conference (CNDS 2002), San
Antonio, TX, January 27-31, 2002.
[4] S.Marti, T.Giuli, K.Lai and M.Baker, "Mitigating Routing
Behavior in Mobile Ad Hoc Networks", Proceedings ofMobicom
2001, Rome,2001.
[5] X Zeng, R. Bagrodia, and M. Gerla. GloMoSim: a library for
parallel simulation of large-scale wireless networks. In Proceedings
of the 12th Workshop on Parallel and Distributed Simulations, May
1998. 11.
[6] Jean-Pierre Hubaux, Levente Butty an, Srdjan Capkun,
"The Quest for Security in Mobile Ad Hoc Networks",
In Proceedings of the ACM Symposium on Mobile Ad Hoc
Networking and Computing (MobiHOC), Long Beach, CA,
USA, October 2001.
201
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
An Efficient and Minimum Cost Topology Construction for Rural Wireless Mesh
Networks
Prof. V. Anuratha & Dr. P. Sivaprakasam
Abstract
Many research efforts as well as
deployments have chosen IEEE802.il as a
low-cost, long-distance access technology to
bridge the digital divide. IEEE 802.11 Wi-Fi
equipment based wireless mesh networks
have recently been proposed as an
inexpensive approach to connect far-flung
rural areas. To establish such network high-
gain directional antennas are used to achieve
long-distance wireless point-to-point links.
Some nodes in the network are called
gateway nodes and are directly connected to
the wired internet, and the remaining nodes
connect to the gateway(s) using one or more
hops.
In this paper the cost of constructing the
antenna towers required is investigated. The
problem is NP hard is shown and that a
better than 0(log n) approximation cannot
be expected, where n is the number of
vertices in the graph. To minimize the
construction cost a new algorithm is
proposed called constant time approximation
algorithm.
The results of proposed approximation
algorithm are compared with both the
optimal solution, and a naive heuristic.
INTRODUCTION
There has been a huge proliferation of
Internet and other communication based
services in the last two decades. However,
this spread is confined to developed
countries, and metropolitan pockets of
developing countries. This is really
unfortunate for developing countries like
India, where around 74% of the population
is rural and are on the wrong side of the
digital divide.
Bridging this divide necessitates, providing
internet connectivity to each and every
village. Providing the same by expanding
the current telephone network to rural areas
is infeasible because of the huge initial
infrastructure costs. Also, deployment of
cellular wireless would not be sustainable
because of its business model, which
demands more high-paying consumer
density.
Emerging technologies like 802.16
WMAN[12],[13], have not yet reached the
scale of competitive mass production, hence
the equipments are expensive. In this regard,
the 802.11 Wi-Fi has shown tremendous
growth and acceptance as a last hop access
solution, because of their low price.
Although 802.11 was primarily designed for
indoor operation, but [3] has established the
possibility of using 802.11 in long-distance
networking.
The diverse requirements are in provisions
of 1) Communication pattern which deals
with the mode of communication one-to-
one, one to- many, many-to-one, and many-
to-many, 2) Delay (real-time, non-real-time,
and delay-tolerant), 3) Service availability
(centralized, distributed, and location-aware)
that deals with the awareness of the
availability of different services, such as
Internet access, real-time communications,
content distribution, interactive gaming,
medical applications, and vehicular safety
applications. 4) Security and 5) Reliability.
An essential requirement to establish long-
distance links is that line-of-sight is
maintained between the radio antennas at the
end-points.
202
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
To ensure line-of-sight across such long
distances would require the antennas to be
mounted on tall towers. The required height
of the towers depends both on the length of
the link, and the height of the obstructions
along the link.
The cost of the tower depends on its
height and the type of material used. For
relatively short heights (10- 20 meters)
antenna masts are sufficient. For greater
heights, sturdier and much more expensive
antenna towers are required. In this paper,
several important contributions are made
towards developing efficient algorithms [14]
to solve this problem. First, the requirements
to establish a point-to-point 802.11 link
between two nodes of a given network graph
is described. Then the formal definition of
the Topology Construction problem (denote
by TC) is given. It's proved that the problem
to be NP hard by a reduction from the set-
cover problem. The approximation algorithm
is presented for this NP hard problem and
the establishment cost of the tower in rural
areas using constant time approximation
algorithm is presented. The rest of this paper
is organized as the following. Section II
gives the Related Works of this technique is
presented. In Section III, the Methodology
of proposed approach is given. Section IV
has the Experiment Results and this paper is
concluded in Section V.
RELATED WORKS
802.11 -based long-distance networks have
been proposed as a cost-effective option to
provide Internet connectivity to rural areas
in developing regions, to enable Information
and Communication Technology (ICT)
services [5].
Rural areas (especially in developing
regions) have populations with very low
paying capacities. Hence, a major factor in
network deployment is the cost of the
infrastructure and the network equipment. In
this context, efficient algorithms are
investigated for the minimum cost topology
construction problem in rural wireless mesh
networks. R. Ramanathan et.al,[4] discussed
on the most critical design issues in
multihop wireless networks. Topology
control has been investigated extensively in
the literature. Nevertheless, it is noted that
most existing studies do not consider the
requirements on upper layer applications or
services. In this article the author address the
topology control issues on service-oriented
wireless mesh networks. In particular, the
author provides a comprehensive survey of
existing works on topology control from a
service-oriented perspective. A general
framework for topology control in service-
oriented WMNs is proposed. To
demonstrate the effectiveness of the
framework, a case study is conducted in
which the main objective is to maximize the
overall throughput in a network with random
unicast traffic. The performance of this
topology control scheme is evaluated by
numerical results. In addition, it is illustrated
that the generated topology can support
advanced technologies, including network
coding and physical-layer network coding,
which can significantly improve the
throughput capacity of a network. The cost
of laying wire to rural areas is prohibitively
expensive.
Also, traditional wireless technologies such
as cellular data networks (e.g., EV-DO) and
upcoming technologies like IEEE 802.16
WiMAX have prohibitively expensive
equipment costs. As a result, there has been
considerable recent interest [6], [7], [8] in
the design of rural mesh networks using
IEEE 802.11 (Wi-Fi) equipment. The cost of
an 802.11 radio (»$50/PCMCIA card) is
orders of magnitude less than that of
cellular/WiMAX base stations. Thus, this
approach is an attractive option for building
low cost networks. D. S. Lun et.al,[9]
presented a distributed random linear
network coding approach for transmission
and compression of information in general
multisource multicast networks.
Network nodes independently and randomly
select linear mappings from inputs onto
output links over some field. The author
shows that this achieves capacity with
203
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
probability exponentially approaching 1
with the code length. Random linear coding
is demonstrated which performs
compression when necessary in a network,
generalizing error exponents for linear
Slepian-Wolf coding in a natural way.
Benefits of this approach are decentralized
operation and robustness to network changes
or link failures. The author shows that this
approach can take advantage of redundant
network capacity for improved success
probability and robustness. Some potential
advantages of random linear network coding
are illustrated over routing in two examples
of practical scenarios: distributed network
operation and networks with dynamically
varying connections. The derivation result
also yields a new bound on required field
size for centralized network coding on
general multicast networks.
METHODOLOGY
In this paper a Novel Topology
Control Scheme is used to identify a set of
semi-permanent highways, such that the best
throughput capacity of the network can be
obtained. Particularly, the wireless highways
are predicted to be rather similar to the
highway system in public transportation
system, which can efficiently provide
connectivity in real application.
A. Computing tower heights at the end-
points of a link
Consider two nodes, u and v that are
separated by a distance luv. The edge (u; v)
is considered to be covered if an 802.11
based point-to-point communication link can
be established between u and v. Assume that
the transmit powers [15] and the gains of the
antennas at both ends are sufficient to over
come the free-space path loss between the
two points. The first basic requirement to
cover the edge between u and v is that there
is a clear visual line-of-sight between the
antennas at the end-points (as shown in Fig.
4a). In other words, the line joining the
antennas mounted on the towers should clear
any obstructions along the path. Secondly, it
is also required that RF line-of-sight is
maintained between the two points. This is
determined by an elliptical area between u
and v termed the first Fresnel zone. To
establish RF line-of-sight, a significant area
of the Fresnel zone (> 60% of the radius of
the Fresnel zone at the location of the
obstruction [1]) should also clear all
obstructions between u and v. However, this
can be simply modeled by extending the
height of the obstruction to include the
radius of the Fresnel zone that has to be in
the clear.
(a)
h^n
■:!>:■
fe* o
tatu)
b(v}
Ol Oa
Figure 1 : Computing the height of towers
at the end-points of a link
In reality, there can be multiple
obstructions between u and v. As in Figure
4b, consider multiple obstructions, Ol, 02,
...,Ok between u and v. Now, let h(u) and
h(v) represent the tower heights at the nodes
of u and v. Covering edge (u,v) requires a
visual and RF line-of-sight connection
between the towers at its two terminal
nodes. This would imply that the straight
line fuv joining the top of the two towers (of
heights h(u) at u and h(v) at v) should clear
every obstruction in (u,v). Hence, it is noted
that given a particular pair of tower heights
at u and v, deciding whether these heights
covers edge (u,v) can be done in time linear
in the number of obstructions on that edge.2
B. Modeling tower costs
An important component in this problem is
the nature of the cost function that maps
tower heights to the cost of building the
tower.
204
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
There are two types of antenna towers that
are used. For heights less than 20 meters,
one can use the cheaper masts. For greater
heights, the more expensive steel towers is
used.
Further, there is an order of magnitude
difference between the cost of the cheaper
masts and that of the steel towers. Thus,
roughly speaking, the cost function is
constant as long as the cheaper masts can be
used and becomes linear in height once the
steel towers are needed, with a jump in cost
when we switch from masts to steel towers.
Let us denote the height at which the
material of the tower has to be switched as
hmin.
Further, there is a physical restriction on the
maximum possible height of a tower,
denoted by hmax. Thus, the cost function c
can be formally defined as
:m=l
h + B ifh mtn <k<h„
Where A, B and K are constants and
Ahmin + B » K. Although, in practice, the
cost function can be modeled as discussed
above, our algorithm works with a much
more general cost functions. Specifically,
we only require the cost function c to satisfy
the following two natural properties CI and
C2. CI
Given the tower costs at two neighboring
nodes u and v, it can be determined (in
polynomial time) whether the corresponding
tower heights cover the edge (u,v). This
simply requires that the corresponding tower
height can be computed (in polynomial
time) given the tower cost. As mentioned
earlier, determining whether the height of
the towers is sufficient to cover an edge can
be done in polynomial time.C2 the cost
function is monotonically increasing with
height, i.e., hl> h2 and c(hl) > c(h2) for
any values of hi and h2.
It is easy to verify that the cost function
c defined earlier in this section satisfies both
of the above properties. In the remainder of
this paper, when the height function h is
unambiguous, often the cost of the tower is
denoted at a node v as c(v) rather than
c(h(v)).
C. The Topology Control Scheme
In this framework, it is considered that
the highways can be partitioned into two
groups, horizontal and vertical. Highways in
each group can operate simultaneously
because they are mutually parallel and can
be placed away enough to reduce
interference below a certain threshold.
Consequently, horizontal and vertical
highways will partition the whole
geographical area into grids, in which nodes
will try to forward their traffic to the nodes
on neighboring highways. The combination
of the following parameters can be
considered for the Topology Control
Scheme
1) Transmission range: Transmission
range of each node in the network
is traditionally an important
design parameter in topology
control. In general, a smaller
transmission range will improve
the channel reuse but may
compromise the connectivity. A
larger transmission range will
improve the connectivity but
reduce the channel reuse.
Therefore, an appropriate range is
chosen as a trade-off between
connectivity and channel reuse.
2) Type of antenna: When directional
antenna or beam forming is used it
may improve the capacity of the
network by reducing the
interference and improve the
transmission quality.
3) Traffic pattern: Traffic pattern is
very important parameter to the
topology. In most studies
previously done suggest that the
traffic is broadcast. With such an
assumption, the problem is
formulated in a way such that the
overall transmission for each
205
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
message is minimized. However,
broadcast traffic may only be a
special case in the future service-
oriented WMN, in which a variety
of patterns may appear, from one-
to one to many-to-many [10].
4) Quality of service (QoS): To
achieve an efficient network a
crucial issue is to enable services
with certain QoS requirements,
such as bandwidth, delay, security
and reliability.
For the first step of this study, only
omnidirectional antenna and purely
random unicast traffic pattern are
considered. Moreover, the random
wireless network elaborated to gain
insights for the future investigation.
Figure 2: Approximation algorithm for NP-hard problem
Constant Time Approximation Aigori thrr.
1. F t- 8 Comment: Implicitly sett growth variables y s t- and cumulative growth variables
«- for all s c V.-w(jS~} denotes the total potential expended in component S including any ju&
components that were merged in creating S.Also implicitly setw v
(1 - a}fc
for all node v
■ r.
2. c *- [i? : v E V,v ^ r} .
For each v == V set d(v') *-
Q. Cement: d(v) denotes the distance vfnade v to the boundary of ths component containing v.
3.
4.
For each veV if v = r them l({v}] *- else A(M] «-
1 , Comment :lis') equal 1 if s active and Q otherwise.
5. while there are active components
6. Comment: Find the next event
7. Find edge e = {i,j)with i E C p E C,j E C fl E C, € p ^ C fl that minimizes
_ Co - d(iO - ri(/]
e= 2(c p ) + l(C fl )
8. Find C E C writh X(£) = 1 that mmimiiss e 2 = | tf"|.— — — — mjE)
9. c = min(e L .. fj)
10. cjCC] <— lcj(CD +■ e J(C] for all C e c
11. /or aft 1? E C r E c
12. d(v0 <— rift?] +■ e.A(C L J
13. ;'/ e = e 2 (i. e. , (T z"s deactiratsd before C^ and C fl mffrfj
14. A.((!") «-
15. Mark all unlabeled vertices of € with label tT, else (i.e., C and C q meet before C is deactivated"}
16. F^F\j[e]
17. C e- CU{c p U Cj- {C p }- {cj
18. ^(C p U C q ) ^- w(C p ) + w(C fl )
19. i/ r E C p UC fl tftm l(c p UC fl ) e- else A{c p UC^g- 1.
20. For ever\- unlabeled vertex v £ CV
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
EXPERIMENTAL RESULTS
In this section, extensive numerical
simulations are carried out to evaluate our
approximation algorithm with the optimal
solution and also a naive heuristic. For this
simulations, synthetic topologies are
generated that aim to match the geographical
structure of village clusters. Simulation
setup is described in more detail as follows.
A. Generating
topologies.
synthetic graph
A circular plane is considered with a
radius of 25Kms. Nodes are placed at
random locations on this plane. A link (u,v)
is considered between any two nodes, u and
v, and for these simulations, assume just one
obstacle, ouv, located on the middle of this
link.
The height of the obstruction (ho) is
selected randomly with a maximum value of
20 meters - the typical height of trees and
small houses in a rural setting. A weight
wuv is assigned to the link equal to twice the
effective height of the obstruction on this
link. As described earlier, the effective
height of an obstruction, is the sum of the
physical height (ho) and 60% of rf , the
radius of the fresnel zone.
B. Naive heuristic
In order to compare with the proposed
approximation algorithms, a naive heuristic
is described for selecting connected
subgraphs and assigning heights to the
nodes. As a first step, to select a connected
subgraph of an input graph G, the heuristic
computes the minimum spanning tree
(MST), T of G (using the link weights
computed as described above). Next, the
heuristic has to assign heights to the nodes
in G, so as to cover all the edges in T while
minimizing the total cost.
Given a set of links to be covered, the
height assignment problem is formulated as
a simple LP, and compute the heights
required on every node.
C. Comparing with the naive heuristic
Now the naive heuristic described above is
compared with the proposed approximation
algorithm. Graphs considered with number
of nodes n =10, 15, 20,..., 50. For each
value of n we generate 50 graph instances.
For each graph, Cnaive the cost of the
solution produced by the naive heuristic, and
Capprox cost of the approximation
algorithm is computed. With this value it is
observed that the proposed approximation
algorithm performs substantially better than
the naive heuristic.
On average, the solutions returned by the
naive heuristic range from 60% (for n = 10)
to as much as 225% more expensive (for n =
50) compared to the solution returned by the
naive algorithm.
D. Comparing with the optimal solution
The optimal solution is computed by
solving an ILP that models the topology
construction problem. The CPLEX LP-
solver [11] to solve this ILP. This approach
is, however, computationally very
expensive, and the LP-solver could return
solutions for graphs with at most 1 1 nodes.
The solution returned by our
approximation algorithm is compared with
the optimal solution for graphs with number
of nodes n = 8, 9, 10, 11. For
each value of n we generate 50 graphs. For
each graph Capprox, the cost of the solution
returned by proposed approximation
algorithm, and Copt, the cost of the optimal
solution.
207
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
Then the mean and standard deviation is
computed over all graphs for different
values of n R
*- n t-.-t.-t n v t-n
K o j pt
L opt
n
Mean (std. dev.) of
Ropt
8
0.58 (0.30)
9
0.57 (0.25)
10
0.55 (0.23)
11
0.52 (0.25)
Table 1 : New Approximation algorithm
vs. Optimal Solution.
The results presented in Table 7 show
that the incorporated approximation
algorithm gives solutions that are 50 - 60%
more expensive than the optimal solution
(for small values of n). Thus, the new
approximation algorithm performs much
better than the worst case guarantee of 0(log
n) on the approximation factor.
While this gap between constant time
approximation algorithm and the optimal
solution is not small, its expected computing
the optimal solution (even if it has to be
done only once) is practically infeasible for
real-life networks. Moreover, this
approximation algorithm performs
substantially better in practice than the naive
heuristic and previous approximation
algorithms
CONCLUSION
In this paper an overview of
establishing a low cost wireless mesh
network for rural areas is presented. In rural
areas nodes are connected using long
distance 802.11 wireless links which are
established using high-gain directional
antennae. The main problem is with the
topology construction for long distance
wireless communication.
An efficient approximation
algorithm is proposed for the topology
construction problem in rural mesh
networks. This work introduces a number of
open research problems in the topology
construction.
One immediate problem is to
consider the case of k > 2 vertex or edge
connectivity, similar to the power optimal
network construction for k-connectivity.
Another important research direction is the
geometric version of this problem. In
practice, all nodes within a certain distance
of each other can establish a link.
In this paper, the location of the
towers is assumed to be fixed (within a
village). A variant of the problem would
make the location of the tower to be a
variable.
This method has added flexibility
than the previous method which would
result in reduced cost. The numerical
experiments demonstrate that the proposed
method with constant time approximation
algorithm performs well within its worst
case performance bounds, and outperforms
the naive heuristic by a substantial margin.
References
[1] I. Akyildiz and X. Wang, "A survey on
wireless mesh networks," IEEE
Communications Magazine, vol. 43, no.
9, pp. S23-S30, Sept. 2005.
[2] M. Lee, J. Zheng, Y.-B. Ko, and D.
Shrestha, "Emerging standards for
wireless mesh technology," IEEE
Wireless Communications, vol. 13, no.
2, pp. 56-63, April 2006.
[3] Pravin Bhagwat, Bhaskaran Raman, and
Dheeraj Sanghi," Turning 802.11 Inside-
Out", In HotNets-II, Nov 2003.
[4] R. Ramanathan and R. Rosales-Hain,
"Topology control of multihop wireless
networks using transmit
poweradjustment," vol. 2, 2000.
[5] Eric Brewer, Michael Demmer, Bowei
Du, Kevin Fall, Melissa Ho, Matthew
Kam, Sergiu Nedevschi, Joyojeet Pal,
Rabin Patra, and Sonesh Surana. The
Case for Technology for Developing
208
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
Regions. IEEE Computer, 38(6):25-38,
June 2005.
[6] P. Dutta, S. Jaiswal, K. Naidu, D.
Panigrahi, R. Rastogi, and A. Todimala.
Villagenet: A low-cost, 802.11 -based
mesh network for rural regions. In
Wireless Systems: Advanced Research
and Development Workshop
(WISARD), 2007.
[7] R. Patra, S. Nedevschi, S. Surana, A.
Sheth, L. Subramanian, and E. Brewer.
WiLDNet: Design and Implementation
of high performance wifi based long
distance networks. InNSDI, 2007.
[8] B. Raman and K. Chebrolu. Design and
evaluation of a new MAC for long
distance 802.11 mesh networks. In
Mobicom, 2005.
[9] D. S. Lun, N. Ratnakar, M. Medard, R.
Koetter, D. R. Karger, T. Ho, E. Ahmed,
and F. Zhao, "Minimum-cost multicast
over coded packet networks," IEEE
Transaction on Information Theory, vol.
52, no. 6, pp. 2608-2623, June 2006.
[10] K. Lu, Y. Qian, H.-H. Chen, and S.
Fu, "WiMAX Networks: From Access
To Service Platform," IEEE Network,
2008, accepted.
[11] ilogcplex.
http://www.ilog.com/products/cplex/ .
[12] M. T. Hajiaghayi, N. Immorlica, and
V. S. Mirrokni. Power optimization in
fault-tolerant topology control
algorithms for wireless multi-hop
networks. In MOBICOM, 2003.
[13] M. T. Hajiaghayi, G. Kortsarz, V. S.
Mirrokni, and Z. Nutov. Power
optimization for connectivity problems.
In IPCO, 2005
[14] Z. Nutov. Approximating minimum
power covers of intersecting families
and directed connectivity problems. In
APPROX-RANDOM, 2006.
[15] S. Sen and B. Raman,"Long distance
wireless mesh network planning:
Problem formulation and solution", In
WWW, 2007.
[16] P. N. Klein and R. Ravi. A nearly
best-possible approximation algorithm
for node-weighted steiner trees. J.
Algorithms, 19(1), 1995.
Author Biographies
Mrs. V. Anuradha graduated with B.Sc
Computer Science in the year 1995 and
completed M.C.A at Madras University in
the Year 2000. Completed her M.Phil in the
year 2003 and also got Guide Approval for
M.Phil in Bharathiar University, Peiyar
University, Bharathiar University and
currently doing her Ph.d.Hear area of
interest is Networks and data mining. Mrs.
V. Anuratha have guided 20 M.Phil scholors
and she have participated and presented
many papers in the national and
international conferences and etc..
Currently she is working as a H.O.D - PG
Department of Computer Science at Sree
Saraswathi Thyagaraja College with a
decade a teaching experience
Co author biography
Dr. P. Sivaprakasam :
He have completed his M.Sc(c.s) in the year
1986, M.Phil in the year 1995 and Ph.d in
2005 on the topic "An Analysis of Web
performance and caching". He have 19 years
of teaching experience. He have published 6
papers at national level and 3 at international
level. He was also sanctioned with 2 UGC
research projects.
He is now currently working as a Associate
professor in Computer Science at Sri Vasavi
college of Arts and Science. His areas of
interest are Internet, Computer Networks,
Service Oriented Architecture.
209
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
Reinforcement Learning by Comparing Immediate
Reward
Punit Pandey
Department Of Computer Science and Engineering,
Jaypee University Of Engineering And Technology
DeepshikhaPandey
Department Of Computer Science and Engineering,
Jaypee University Of Engineering And Technology
Dr. Shishir Kumar
Department Of Computer Science and Engineering,
Jaypee University Of Engineering And Technology
Abstract — This paper introduces an approach to Reinforcement
Learning Algorithm by comparing their immediate rewards
using a variation of Q-Learning algorithm. Unlike the
conventional Q-Learning, the proposed algorithm compares
current reward with immediate reward of past move and work
accordingly. Relative reward based Q-learning is an approach
towards interactive learning. Q-Learning is a model free
reinforcement learning method that used to learn the agents. It is
observed that under normal circumstances algorithm take more
episodes to reach optimal Q-value due to its normal reward or
sometime negative reward. In this new form of algorithm agents
select only those actions which have a higher immediate reward
signal in comparison to previous one. The contribution of this
article is the presentation of new Q-Learning Algorithm in order
to maximize the performance of algorithm and reduce the
number of episode required to reach optimal Q-value.
Effectiveness of proposed algorithm is simulated in a 20 x20 Grid
world deterministic environment and the result for the two forms
of Q-Learning Algorithms is given.
Keywords-component; Reinforcement Learning, Q-Learning
Method, Relative Reward, Relative Q-Learning Method.
I.
Introduction
Q-Learning algorithm proposed by Watkins [2,4] is a model
free and online reinforcement learning algorithm. In
reinforcement learning selection of an action is based on the
value of its state using some form of updating rule. There is an
interaction between agent and environment where the agent has
to go through numerous trials in order to find out the best
action. An agent chooses that action which has maximum
reward obtained from its environment. The reward signal may
be positive or negative depends on the environment.
Q-learning has been used in many applications because it
does not require the model of environment and is easy to
implement. State-action value, a value for each action from
each state, converges to the optimal value as state-action pairs
are visited many times by the agent.
In this article we propose a new relative reward strategy for
agent learning. Two different form of Q-Learning method is
considered here as a part of study. First form of Q-Learning
method uses a normal reward signal. In this algorithm Q-value
evaluates whether things have gotten better or worse than
expected as a result of an action selection in the previous state.
The action selected by agents is most favorable which has
lower TD error. Temporal difference is computed on the basis
of normal reward gain by agents from its surroundings. An
estimated Q-value in the current state is than determined using
Temporal Difference. Agent actions are generated using the
maximum Q-values. The second form of Q-Learning
algorithm is an extension towards a relative reward. This form
of Q-Learning method utilizes the relative reward approach to
improve the learning capability of algorithm and decreases the
number of iteration. In this algorithm only those action is
selected which has a better reward from its previous one.
This idea comes from psychological point of views that
human beings tend to select only those action which has higher
reward value. However, this algorithm is not suitable for multi
agent problems. To demonstrate effectiveness of the proposed
Q-Learning algorithm, Java applet is utilized to simulate a robot
that reaches to a fixed goal. Simulation result confirms that the
performance of proposed algorithm is convincingly better than
conventional Q-learning.
This paper is organized as follows: Basic concept of
reinforcement learning is presented in section 2. Section 3
describes about the conventional Q-Learning method. Section 4
presents a new Relative Q-Learning in context of relative
immediate reward. Section 5 describes Experimental setup &
results and concluding remarks follow in Section 6.
II. Reinforcement Learning
Reinforcement learning (RL) is a goal directed learning
methodology that is used to learn the agents. In Reinforcement
learning [1,5,6,7,8,9] the algorithm decide what to do and how
to map situations to actions so that we maximize a numerical
reward signal. The learner is not advised which actions to take,
but instead it discover which actions provide the maximum
reward signal by trying them. Reinforcement learning is
defined by characterizing a learning problem. Any algorithm
that can able to solve the defined problem, we consider to be a
reinforcement learning algorithm. The key feature of
reinforcement learning is that it explicitly considers the whole
problem of a goal-directed agent interacting with an uncertain
environment. All reinforcement learning agents [3,10,11,12]
have explicit goals, can sense aspects of their environments,
and can choose actions to influence their environments. In
210
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
reinforcement learning agent prefer to choose actions that it has
tried in the past and found to be effective in producing
maximum reward. The agent has to exploit based on what it
already knows in order to obtain reward and at the same time it
also has to explore in order to make better action selections in
the future. Reinforcement learning has four elements policy,
reward function, value function and model of environment.
(stale)
: ■=««■: :
Learner
:;:lcn'i
Environment
agent because it alone is sufficient to take the decision on
further action.
III. Q-Learning
Q-learning is a form of model- free reinforcement learning [2]
(i.e. agent does not need an internal model of environment to
work with it). Since Q-learning is an active reinforcement
technique, it generates and improves the agent's policy on the
fly. The Q-learning algorithm works by estimating the values
of state-action pairs.
The purpose of Q-learning is to generate the Q-table, Q(s,a),
which uses state-action pairs to index a Q-value, or expected
utility of that pair. The Q-value is defined as the expected
discounted future reward of taking action a in state s, assuming
the agent continues to follow the optimal policy. For every
possible state, every possible action is assigned a value which
is a function of both the immediate reward for taking that
action and the expected reward in the future based on the new
state that is the result of taking that action. This is expressed by
the one-step Q-update equation [2,4,10,13,14].
Q(s, a)
Q(s, a) + a [r+ y * max a Q(s', a') - Q(s, a)]
(i)
Figure 1.1: .Reinforcement learning
ao
*2
So
Si
S 2
ro
T2
Goal: learn to choose actions that maximize:
ro + 7 ri + f n + . . . , where < 7 < 1
Figure 1. 2 : Reinforcement learning
Model of the environment is an optional element because
reinforcement learning also supports the model free algorithms
like Q-learning.
A policy for our agent is a specification of what action to take
for every Input. In some cases policy may be a simple function
or look-up table or sometime it can be an extensive
computation. The policy is the core of reinforcement learning
AGENT
action seletor
Q- factor Table
Input
Reinforcement , ,
action
WORLD
Figure 2: Structure of the Q-Learning agent
Where a is the learning factor and y is the discount factor.
These values are positive decimals less than 1 and are set
through experimentation to affect the rate at which the agent
attempts to learn the environment. The variables s and a
represent the current state and action of the agent, r is the
reward from performing s' and a', the previous state and
action, respectively.
The discount factor makes rewards earned earlier more
valuable than those received later. This method learns the
values of all actions, rather than just finding the optimal
policy. This knowledge is expensive in terms of the amount of
information that has to be stored, but it does bring benefits. Q-
learning is exploration insensitive, any action can be carried
out at any time and information is gained from this experience.
The agent receives reinforcement or reward from the world,
211
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
and returns an action to the world round and round as shown
below:
A Elementary parts of Q-learning:
Environment:
Q-learning based on model-free mode of behavior i.e the
environment is continuously changing. Agent does need to
predict future state. Environment can be either deterministic or
non-deterministic. In deterministic environment application of
single state lead to a single state where as in nondeterministic
environment application of a single action may lead to a
number of possible successor states. In case of non-
deterministic environment, each action not only labeled with
expected immediate reward but also with the probability of
performing that action. For the sake of simplicity we are
considering deterministic environment in this thesis work.
Reward Function:
A reward function defines the goal in a reinforcement learning
problem, it maps each perceived state (or state-action pair) of
the environment to a single number, a reward, indicating the
intrinsic desirability of that state. A reinforcement learning
agent's sole objective is to maximize the total reward it
receives in the long run. The reward function defines what the
good and bad events are for the agent.
Action-value function:
The Q-learning learning is based upon Quality-values (Q-
values) Q(s,a) for each pair (s,a). The agent must cease
interacting with the world while it runs through this loop until
a satisfactory policy is found. Fortunately, we can still learn
from this. In Q-learning we cannot update directly from the
transition probabilities-we can only update from individual
experiences. In 1 step Q-learning, after each experience, we
observe state s', receive reward r, and update:
Q(s, a) = r+ y maxa' Q(s', a') (2)
B. Q-learning Algorithm:
Initialize Q(s, a) arbitrarily
Repeat (for each episode)
Choose a starting state, s
Repeat (for each step of episode):
Choose a from s using policy derived from Q
Take action a, observe a immediate reward r, next state s'
Q(s, a) <— Q(s, a) + a [r+ y * maxa' Q(s', a') - Q(s, a)]
s<— s' ;
Until state s' match with the Goal State
Until a desired number of episodes terminated
State
Action
Figure 3 : Q-Learning Architecture
IV. Relative Q-Learning
This section introduces a new approach Relative reward to
conventional Q-learning that makes Relative Q-Learning.
Conventional Q-learning has been shown to converge to the
optimal policy if the environment is sampled infinitely by
performing a set of actions in the states of the environment
under a set of constraints on the learning rate a. No bounds
have been proven on the time of convergence of the Q-learning
algorithm and the selection of the next action is done randomly
when performing the update. This simply mean that the
algorithm would take a longer time to converge as a random set
of states are observed which may or may not bring the state
closer to the goal state. Furthermore, it means that this function
cannot be used for actually performing the actions until it has
converged as it has a high chance of not having the right value
as it may not have explored the correct states. This is especially
a problem for environments with larger state spaces. It is
difficult to explore the entire space in a random fashion in a
computationally feasible manner. So by applying below
mention method and algorithm we try to keep the Q-learning
algorithm near to its goal in less time and less number of
Episode.
A Relative Reward
Relative reward is a concept that compares (current reward
with the previous received reward) two immediate rewards.
The objective of the learner is to choose actions maximizing
discounted cumulative rewards over time. Let there is an agent
in state st at time t, and assume that he chooses action at. The
immediate result is a reward rt received by the agent and the
state changes to st+1. The total discounted reward [2,4]
received by the agent starting at time t is given by:
r(t)=r t +yr t+1 +y 2 r t+ 2+ +y n r t+n +
Where y is discount factor in the range of (0: 1).
(3)
212
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
The immediate reward is based upon the action or move
taken by an agent to reach the defined goal in each episode.
The total discounted reward can maximize in less number of
episode if we select the higher immediate reward signal from
previous.
B. Relative Reward based Q-Learning Algorithm
Relative reward based Q-learning is an approach towards
maximizing the total discounted rewards. In this form of Q-
learning we selected the maximum immediate reward signal by
comparing it with previous one. This is expressed by the new
Q-update equation.
Q(s, a) = Q(s, a) + a [max(r(s,a),r(s\a'))+ y maxa' Q(s', a')
-Q(s,a)]
Algorithm:
Initialize Q(s, a) arbitrarily
Repeat (for each episode)
Choose a starting state, s
Repeat (for each step of episode):
Choose a from s using policy derived from Q
Take action a, observe a immediate reward r, and next state
s'
Q(s, a) = Q(s, a) + a [max(r(s,a),r(s',a'))+ y maxa' Q(s', a')
-Q(s,a)]
s<— s' ;
Until state s' match with the Goal State
Until a desired number of episodes terminated
V. Experiments & Results
The Proposed Relative Q-Learning was tested on 10 x 10
and 20 x 20 grid world environment. In the Grid World Square
There are four possible actions for the agent as it is a
deterministic environment given in figure 4.
IJtMSl
*
| Start | | Reset |
In order to consider the situation of encountering a wall, the
agent has no possibility of moving all the way in the given
direction. When the agent enters into goal states, it receives 50
as a reward. We are also providing the immediate reward value
by incrementing or decrementing the Q-value marked with S
represent the start state and G represent the goal state. The
purpose of the agent is to find out the optimum path to arrive at
the goal state starting from the start state, and to maximize the
reward it receives.
Conventional Q-Learning/ Random Strategy
60 1
1 38 75 112 149 186 223 260 297 334 371 408 445 482 519 556
Episode
Figure5: Conventional Q-Learning.
Relative Reward Basd Q-Learning/Random Strategy
— Seriesl
1 42 83 124 165 206 247 288 329 370 411 452 493 534 575 616
Episode
Figure4: A 10 x 10 Grid World Environment
Figure6: Relative Q-Learning
We have executed 500 episodes to converge the Q-value.
The grid world is a deterministic environment so the value of
learning a and discount rate Y were set to 0.8. Figure 5 &
213
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
Figure 6 shows the relationship between Q- Values and the
number of episode where x axis represents the number of
episode and y axis represents the Q-values. Figure 5 represents
the result of conventional Q-Learning where we can see that Q-
value converges after executing 500 episodes but in figure 6
Relative Q-learning takes 300 episode
So we can say that convergence rate of relative Q-learning
is faster than conventional Q-learning.
VI. Conclusion & future work
This paper proposed an algorithm which compares the
immediate reward signal with its previous one. The agent will
immediately return back to previous state if it will receive the
lower reward signal for that particular move. If conventional
Q-learning was applied in the real experiment, a lot of
iterations were required to reach the optimal Q values. The
Relative Q-learning algorithm was proposed for environment
which used small amount of episodes to reach the convergence
of Q-values. This new concept allows the agent to learn
uniformly and helps in such a way so that it will not deviate
from its goal. Part of future work may be included to verify the
proposed algorithm in nondeterministic environment.
References
[I] J.F. Peters, C. Henry, S. Ramanna, Reinforcement learning with pattern-
based rewards, in proceding of forth International IAS TED Conference.
Computational Intelligence (CI 2005) Calgary, Alberta,Canada, 4-6 July
2005,267-272
[2] Technical Note Q,-Learning Christopher J.C.H. Watkins and Peter
Dayan Centre for Cognitive Science, University of Edinburgh, Scotland
Machine Learning, 8, 279-292 (1992)
[3] J.F. Peters, C. Henry, S. Ramanna, Rough Ethograms: Study of
Intelligent System Behavior. In:M.A.Klopotek, S. Wierzchori ,
K.Trojanowski(Eds), New Trends in Intelligent Information Processing
and Web Mining (IIS05), Gdansk, Poland, June 13-16 (2005),1 17-126.
[4] C. Watkins, "Learning from Delayed Rewards", PhD thesis, Cambridge
University, Cambridge, England, 1989
[5] J.F.Peters,K.S.Patnaik,P.K.Pandey,D.Tiwari, "Effete of temperature on
swarms that learn", In Proceeding of IASCIT-2007,Hyderabad,INDIA
[6] P.K.Pandey,D.Tiwari, " Temperature variation on Q-Learning",In
Proceeding of RAIT in FEB 2008,ISM Dganbad
[7] P.K.Pandey,D.Tiwari, " Temperature variation on Rough Actor-Critic
Algorithm", Global Journal Computer Science and Technology, Vol 9,
No 4 (2009), Pennsylvania Digital Library
[8] L.P. Kaelbling, M.L. Littman, A.W. Moore, Reinforcement learning: A
survey Journal of Artificial Intelligence Research, 4, 1996, 237-285.
[9] R.S. Sutton, A.G. Barto, and Reinforcement Learning: An Introduction
(Cambridge, MA: The MIT Press, 1998).
[10] C. Gaskett, Q-Learning for Robot Control. Ph.D. Thesis, Supervisor:
A.Zelinsky, Department of Systems Engineering, The Australian
National University, 2002.
[II] Thrun. S.and Schwartz.A.(1993),Issues in using function approximation
for reinforcement learning, in Proceeding of the 1993 Connectionist
Models Summer School,Erblaum Associates. Nj.
[12] Richard S. Sutton, Reinforcement Learning Architectures, GTE
Laboratories Incorporated, Waltham, MA 02254.
[13] Tom 0'Neill,Leland Aldridge,Harry Glaser, Q-Learning and Collection
Agents, Dept. of Computer Science, University of Rochester
[14] Vanden Berghen Frank, Q-Learning, IRIDIA, Universit Libre de
Bruxelles
214 http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
Information realization with statistical predictive
inferences and coding form
D.Mukherjee
Sir Padampat Singhania University
Udaipur-3 13601 ,Raj asthanjndia
P.Chakrabarti* , A.Khanna , V.Gupta
Sir Padampat Singhania University
Udaipur-3 13601 ,Raj asthan,India
Abstract — The paper deals with information realization in case of
grid topology. Nodal communication strategies with clusters has
also been cited. Information prediction has been pointed out with
relevant statistical method, forward sensing, backward sensing
and cumulative frequency form. Binary tree classifier theory has
been applied for information grouping. The paper also deals with
comparison analysis of information coding.
Keywords- grid topology forward sensing
binary tree classifier, information coding
backward sensing,
I. INFORMATION MERGING IN GRID IN
DIAGONAL APPROACH
In order to solve complex problems in artificial intelligence
one needs both large amount of knowledge & some
mechanisms for manipulating that knowledge to create
solutions to new problems .Basically knowledge is a mapping
of different facts with help appropriate functions for e.g. Earth
is a planet. Can be realized as a function -planet (Earth).
Information merging can be realized as combining different
pieces of information to arrive at a conclusion. The different
information elements can be related in different ways i.e.
either in hierarchy or in form of a graph or even a mesh.
Consider a mesh of size m X n i.e. m rows & n columns then
if each intersection point has a information element placed on
it then one way of merging element A with B can be covering
a path of length (5XN) (here m= 8 & n=9). If weight of
covering each path is considered same then in case of diagonal
approach we can find a path of diagonal nature of length 5 V2
and then travelling a length (N-5) in linear fashion thus finding
a shortest path the same can also be determined by graph
algorithms like Dijkstra's or kruskal's algorithm for
minimum spanning tree. If each path is considered to be of
zero weight then interestingly there is no sense travelling a
path from A to B i.e. we can directly merge the two points
i.e. we take point A &directly merge it with point B in such a
case we need to have some stack like mechanism to determine
the order in which the nodes arrive & are merged.
*
*
*
*
4 ►
B
Mrows (m- l)paths
* shows path of traversal
N columns (n-1) paths
Figl: Information merging in mesh/grid
The above concept can be realized in DDM(Distributed Data
Mining) where large amount of geographically scattered
knowledge is merged & is mined to derive conclusions &
make decisions for e.g. GIS i.e. the Geographical Information
System which uses cartography(art of making maps) with
various information elements(sources) to derive decision
support results like which route to choose for a given
destination.
II. INFORMATION MERGING IN CLUSTER NETWORKS
This section mainly focuses on the nodal communication
between the farthest node in a N*N structurefl] and
information realization indicates nodal message . Let us
assume each cluster to be consisting of 16 nodes and then try
to communicate between the source and the destination node
as described in the figl. The point to be noted here is that to
establish the communication link between the adjacent
elements or units of the cluster we have to have the
communication in just reverse order in the 2 adjacent
elements. The order of the communication is
215
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
The condition can be visually imagined as follows:
Now let us first talk about the case when there is only one
element i.e, 1*1. In this particular case if we want to
communicate between the farthest node then there will be only
1 node in between the source and the destination which can be
further visualized as follows:
If we denote it by using the function f(x)then the value of
f(x)will be l.f(x)=l; The intermediate node is 11. Now let us
consider the case 2*2 matrix the value here will be
f(x)= 1+2=3; The intermediate nodes are 1(2,3),2(4).
n n
For the case for the 3*3 matrix the value of the function
f(x)= 1+2+2=5;
i m x in
nodes, i.e, f(x)=3+4+4;In case of 4*4 matrix to communicate
between the farthest node we need 7 nodes, i.e,
f(x)=3 +4+4+4 ;In case of 16 elements in a ring , we can
proceed as follows. Let us consider the case of 1*1 matrix to
communicate between the farthest node we need 3 nodes, i.e,
f(x)=7.In case of 2*2 matrix to communicate between the
farthest node we need 7 nodes, i.e, f(x)=7+8;In case of 3*3
matrix to communicate between the farthest node we need 7
nodes, i.e, f(x)=7+8+8;In case of 4*4 matrix to communicate
between the farthest node we need 7 nodes, i.e,
f(x)=7+8+8+8;Now the total number of nodes can be derived
by the general formula as (N/2-l)+(M-l)*(N/2) where N =
number of nodes present in the unit or element, M =
dimension of the square matrix. The data can be represented in
the tabular form as follows:
No. of
nodes
1*1
2*2
3*3
4*4
4
1
3
5
7
8
3
7
11
15
16
7
15
23
31
nz
nz
Fig. 2: Nodal communication in cluster
The x-axis represents the M*M matrix where M varies from 1
to 3. The y-axis represents the number of optimum
communication nodes required in the establishing the path
between the source node and the farthest node. The number of
nodes per element is indicated by the 3 colors.
Similarly for the 4*4 matrix we can get the value of
f(x)=l+2+2+2.
Here in this case we were having only 4 elements in a ring
.Suppose we have 8 elements in the ring in that case we have
to compute the number of nodes required to communicate or
establish the connection between the farthest nodes.
Justification - Let us consider the case of 1*1 matrix to
communicate between the farthest node we need 3 nodes, i.e,
f(x)=3.In case of 2*2 matrix to communicate between the
farthest node we need 7 nodes, i.e, f(x)=3+4;In case of 3*3
matrix to communicate between the farthest node we need 7
III. STATISTICAL INFERENCE OF FUTURISTIC
VALUES
In statistical inferences the input & output of a situation are
related with a certain relation or function based on which we
infer futuristic values. Consider a real-time situation in which
a given input parameter is observed over time between instants
Tl & T2 given the relation [2]
M t = a.e* then M avg = V(M ti . M t2 )
216
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
Case 1:
If we take observations at equal instants of time then
ti
tl+k
M t i = a.e
M t2 = a.e
M t3 = a.e tl+2k
General term M tn = a.e tl+(n " 1)k i.e. the values of output M
forms a G.P. series of increasing order common ratio as e k .
C. Cumulative frequency based information sensing
OBSERVATIONS
INFORMATION INVOLVED
gi
UMmag
g2
13,15
g3
14,15,16
g4
12,13,15
g5
11,12
g6
11,12,13,16
Case 2:
Table 1 : Association of information against each observation
ti
t\ =>M tl = a.e
tl + kl => M f2 =a.e'
tl+kl
t2 + k2
tl+kl+k2
tl + (kl + k2) => M t2 = a.e t2+k2 = M t2 =
tl+Ktotal
i.e.
If we take observation at unequal timing interval in that case
Tl
T2
T3
a.e
General term Tn = Tl+(kl+k2+k3+. . .+kn)
Tn = tn-1 + kn-1 = tl + (kl + k2 + k3+...+kn-l) =>
M tn = *.e tn - l+kn - 1 = M tn = a . e ( tl+kl+k2+k3+ - +kn - 1 > = a.e'
now any futuristic value say at instant tn is
M tn = a.e^.e^ '* 1 (observed value)
Given M t = a.e 1 , taking log on both sides we have,
ln(M t ) = ln(a) + 1
i.e. h^Mtn) = ln(a) + tn
ln(M ta ) = ln(a) + tl+kl+k2+k3+...+kn-l
Thus we have obtained a log linear model for the above
function M t = a.e t using which we can calculate or predict the
futuristic values for increased ranges.
Y = m.X + C
If we try to minimize the value of Ktotal we can do so by
making kl=k2=k3=...=kn-l which is same as Case 1.
IV. PROJECTION OF SENSED INFORMATION
Let 1= {ii,i 2 ,...i n } be the set of sensed information. In the
process of feature appropriate observation, forward selection ,
backward elimination and decision based induction methods
are applied.
A. Forward selection based information sensing
Let 1= {ii , i 2 ,....,i n }be the set of information estimates of
various trends noted after observation in respective timing
instants Y = {yi,y2,--.y n }- The accuracy measurement is to be
calculated first based on comparison analysis. The minimum
deviation reflects high accuracy level of prediction and that
information will be selected. In this manner, { } , {best
information} , {first two} . . . .will be selected.
B. Backward elimination based information sensing
Using backward elimination , in each stage each information is
eliminated and thereby after the final screening stage the
projected set reveals the final optimum information space.
Features
Initial
value
Count
Value
(Value) 2
ii
0.1
3
0.3
0.09
12
0.2
3
0.6
0.36
13
0.3
4
1.2
1.44
14
0.4
2
0.8
0.64
15
0.5
3
1.5
2.25
16
0.6
3
1.8
3.24
Table 2 : Determination of count and value
Now CF = ( x , y , z )
where x = number of elements , y = linear sum of the elements
and z = sum of the square of the elements [3]
V. BINARY TREE BASED GAIN CLASSIFIER
In this section information represents gain analysis. A
search[4] can be formed based on the initial search term and
its gradual sub term while the process of matching. Thereby
the level is increased, in initial search term is the root and the
final term fully matching with the context of the users' desire
is a leaf node.
GO
Gl,l Gl,2
G2,l G2,2 G2,3 G2,4
A A KK
G3,l G3,2 G3,3G3,4 G3,5 G3,6 G3,7 G3,8
LEVEL
LEVEL 1
LEVEL 2
LEVEL 3
Fig3 : Binary tree based gain classifier
In the above figure, GO is the root that is initial search term. If
a user wants to analyze further gain classification, then
identify each search term as a binary code and by giving the
code number he can analyze the position of gain estimate in
the model . The concept of coding is as follows:
217
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
Value = if the search term is a left child of parent node
= 1 otherwise
C3
C4 :
{10}
{11}
N
Theorem: In the process of coding, X l/2 L i =1, where
i=l
Li is the length of code of ith leaf node in the tree, N is total
number of leaf nodes and Ki<N.
Proof:
From fig.3 codes of leaf nodes are as follows:
Nodes
Respective code
G3,l
000
G3,2
001
G3,3
010
G3,4
011
G3,5
100
G3,6
101
G3,7
110
G3,8
111
So, N=8. Each leaf node has identical code length i.e. 3.
Therefore, 1/2^=1/2=1/8, 1/2= 1/8, ...1/2=1/8
We now design a binary tree based classifier taking some
parameters for examination purpose and represent each point
on the basis of a code generated by arithmetic coding. Finally,
represent the same on the basis of set theory .We assume that
the gain set available is G ={ gl,g2,g3,g4 }.The parameters
based on which the examination is to be carried are the
elements of the set P = { pl,p2,p3 }.The result of the
examination are denoted in the form of Boolean variables such
that the outputs are denoted as:
NO =
YES=1
At the initial timing instant, the parameter pi is applied for
testing purpose. Hence, in the initial stage, there will be at
least one class while a maximum of two classes. In the second
level, the parameter p2 is applied and accordingly the classes
are defined. In the final stage, the parameters p3 is applied.
If we assume the classifier as a binary tree representation, we
can apply arithmetic coding to each class such that a 'NO' of a
particular exam is denoted by '0' and a 'YES' is denoted by
T.In the initial stage, the class which contains the elements
for negative supply of pi is CI ={ dl,d2}, while, C2 = { d2,d4
}. In this manner, the tree is to be constructed such that the
code word for each class is denoted by ijk where i C { 0,1 } , j
G{0,1 } and kC{0,l }.
For pi:
Forp2:
CI
C2
Cl =
C2 =
= { dl,d2 }
= { d2,d4 }
{00}
{01}
For p3 :
Class
ijk
False
True
CI
000
pl,p2,p3
-
C2
001
pl,p2
P 3
C3
010
pl,p3
p2
C4
011
Pi
P 2,p3
C5
100
p2,p3
pl
C6
101
p2
pl,p3
C7
110
p3
pl,p2
C8
111
-
pl,p2,p3
In the initial stage, classes are CI, C2 based on the parameter
pl. In the second stage, the classes are C1,C2,C3 based on p2.
In the last stage, classes are C1,C2...C8 based on p3.This
means that if we assume that 'n' is the number of parameters
involved in the system for examination purpose. Then, the
maximum length of code word for a particular class is 'n'. The
number of classes is 2 n , provided that the classes are distinct in
nature.
VI. CODED INFORMATION SENSING
Let original message is "FATHER". For the first alphabet,
lvalue = Imposition of that)+ n /100). Hence it's offset value =
ceiling of (the product of |u va i ue and 10). The weight is given by
its position in alphabet string[5].
Therefore totalvalue = offset value * weight. From the next
character onwards, |u va i ue _next = l/(mod value of (position of
next - position of previous ) + 7i/100). Hence totalvalue is
calculated in similar manner. Now, bias value will be equal to
total number of characters in the message. Compute netvalue
as (totalvaluefirst char + total valuelast char)- (bias value)
and let it be x (say).
Mode Operation
0<x<100 Reverse the message.
1 00<x<l 50 Circular left shift of message by n/2 bits
where n= bias value.
1 50<x<200 Circular right shift of message by n/2 bits
Iteration 1: |u F = l/((position of 'F' in alphabet list) + 71 /100) =
l/((6)+ 71 /100) = 0.165798547. Offset value = ceiling of
(0.165798547*10) = 2. Weight = position of T' in alphabet list
= 6. Thus, totalvalue = 2*6 =12.
Iteration 2: [i A = l/(|(position of 'A' - position of 'F')|+ 7i/100)
= 1/(|(1-6)|+ tt/100) = 1/5.031415927 = 0.198751209. Offset
value = ceiling of (0.198751209*10) = 2. Weight = 1. Thus,
totalvalue = 2*1 =2.
Iteration 3: \i T = l/(|(position of 'T' - position of 'A')|+ ti/100)
= 1/(|(20-1)|+ tt/100) = 1/19.03141593 = 0.052544697. Offset
value = ceiling of (0.052544697*10) = 1. Weight = 20. Thus,
total value =1*20 = 20.
218
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
Comparator
Encryption
system
Output
Cipher
wn
Ci = offset value for i = 1 to n , wi = weight , wb = bias value
Fig 4: Coding Model
Iteration 4: ju H = l/(|(position of 'H' - position of T')|+ tt/100)
= 1/(|(8-20)|+ ti/100) = 1/12.03141593 = 0.083115736. Offset
value = ceiling of (0.083115736*10) = 1. Weight = 8. Thus,
total_value=l*8 = 8.
Iteration 5: \i E = l/(|(position of 'E' - position of 'H')|+ tt/100)
= 1/(|(5-8)|+ tt/100) = 1/3.031415927 = 0.32987885. Offset
value = ceiling of (0.32987885*10) = 4. Weight = 5. Thus,
totalvalue = 4*5 = 20.
Iteration 6: \i R = l/(|(position of 'R' - position of 'E')|+ tt/100)
= 1/(|(18-5)|+ tt/100) = 1/13.03141593 = 0.076737632. Offset
value = ceiling of (0.076737632*10) = 1. Weight = 18. Thus,
total_value=l*18 = 18.
Now, wb= bias value = number of bits in FATHER= 6. So
net_value= accumulated sum of all totalvalue - wb =
(12+2+20+8+20+18) - 6 = 74. It falls in the range 0<x<100.
So, "FATHER" is reversed.
Therefore resultant cipher is "REHTAF".
VII. CONCLUSION
The paper points out information merging in grid and cluster
network models. Statistical means of information prediction as
well as forward, backward and cumulative frequency based
schemes have been analyzed . Binary tree based information
classification and coded information have been justified with
relevant mathematical analysis.
REFERENCES
[1] A.Kumar , P.Chakrabarti , P.Saini , V.Gupta /'Proposed techniques of
random walk, statistical and cluster based node realization" communicated to
IEEE conf. Advances in Computer Science ACS 2010 , India , Dec 10
[2] P.Chakrabarti , S.K.De , S.C.Sikdar , "Statistical Quantification of Gain
Analysis in Strategic Management" published in IJCSNS ,Korea , Vol 9 Nol 1
,pp.3 15-3 18, 2009
[3]P.Chakrabarti, "Data mining- A Mathematical Realization and cryptic
application using variable key" published in International journal , Advances
in Information Mining , Vol 2 No 1, pp- 18-22,20 10
[4] P.Chakrabarti, P.S.Goswami, "Approach towards realizing resource
mining and secured information transfer" published in international journal of
IJCSNS, Korea , Vol 8 No.7, pp345-350, 2008
[5] P.Chakrabarti , "Attacking Attackers in Relevant to Information Security"
Proceedings of RIMT-IET, Mandi Gobindgarh. pp 69-71, March 29, 2008
About authors:
Debasis Mukherjee (20/08/80) is pursuing Ph.D. from USIT,
GGSIPU, Delhi, India from2010. He received the M. Tech.
degree in VLSI Design from CDAC Noida in 2008 and
bachelor degree in Electronics and Instrumentation
Engineering from BUIE, Bankura, West Bengal, India in
2003 .He achieved first place in district in "Science Talent
Search Test" 1991. He has some publications of repute in
IEEE conferences.
Dr.P.Chakrabarti(09/03/81) is currently serving as Associate
Professor in the department of Computer Science and
Engineering of Sir Padampat Singhania University,Udaipur.
Previously he worked at Bengal Institute of Technology and
Management , Oriental Institute of Science and Technology,
Dr. B.C. Roy Engineering College, Heritage Institute of
Technology, Sammilani College. He obtained his Ph.D(Engg)
degree from Jadavpur University in Sep09,did M.E. in
Computer Science and Engineering in 2005, Executive MBA
in 2008and B.Tech in Computer Science and Engineering in
2003 .He is a life member of Indian Science Congress
Association , Calcutta Mathematical Society , Calcutta
219
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
Statistical Association , Indian Society for Technical
Education , Cryptology Research Society of India,
IAENG(HongKong), CSTA(USA), annual member of
Computer Society of India, VLSI Society of India ,
IEEE(USA), senior member of IACSIT(Singapore) and
selected member of The IAENG Society of Artificial
Intelligence , Computer Science , Data Mining. He is a
Reviewer of International journal of Information Processing
and Management (Elsevier) , International Journal of
Computers and Applications , Canada and International
Journal of Computer Science and Information
Security(IJCSIS,USA), editorial board member of
International Journal of Engineering and Technology,
Singapore and International Journal of Computer and
Electrical Engineering. He has about 100 papers in national
and international journals and conferences in his credit and
two patents(filed). He has several visiting assignments at BHU
Varanasi , IIT Kharagpur , Amity University,Kolkata , et al.
A.Khanna and V.Gupta are the third year students of
Information Technology and Computer Science & Engg.
branch respectively of Sir Padampat Singhania University.
220 http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
Scaling Apriori for Association Rule Mining
using Mutual Information Based Entropy
S.Prakash,Research Scholar
Sasurie College of Engineering
Vij ayamangalam,Erode(DT)
Tamilnadu, India.Ph.09942650818
Mail:prakash_ant2002@yahoo.co.in
Abstract - Extracting information from large datasets is
a well-studied research problem. As larger and larger
data sets become available (e.g., from customer
behavior data from organizations such as Wal-Mart) it
is getting essential to find better ways to extract
relations (inferences) from them. This thesis proposes
an improved Apriori algorithm to minimize the number
of candidate sets while generating association rules by
evaluating quantitative information associated with
each item that occurs in a transaction, which was
usually, discarded as traditional association rules focus
just on qualitative correlations. The proposed approach
reduces not only the number of item sets generated but
also the overall execution time of the algorithm. Any
valued attribute will be treated as quantitative and will
be used to derive the quantitative association rules
which usually increases the rules' information content.
Transaction reduction is achieved by discarding the
transactions that does not contain any frequent item set
in subsequent scans which in turn reduces overall
execution time. Dynamic item set counting is done by
adding new candidate item sets only when all of their
subsets are estimated to be frequent. The frequent item
ranges are the basis for generating higher order item
ranges using Apriori algorithm. During each iteration
of the algorithm, use the frequent sets from the previous
iteration to generate the candidate sets and check
whether their support is above the threshold. The set of
candidate sets found is pruned by a strategy that
discards sets which contain infrequent subsets. The
thesis evaluate the scalability of the algorithm by
considering transaction time, number of item sets used
in the transaction and memory utilization. Quantitative
association rules can be used in several domains where
the traditional approach is employed. The unique
requirement for such use is to have a semantic
connection between the components of the item-value
pairs. The proposal used mutual information based on
entropy to generate association rules from non-
biological datasets.
Key words- Apriori, Quantitative attribute, Entropy
Dr.R.M.S.Parvathi M.E.(CSE),Ph.D.
Principal
Sengunthar College of Engg.for Women
Tiruchengode. Tamilnadu, India
rmsparvathi @ india.com
I. INTRODUCTION
Data mining, also known as knowledge
discovery in databases, has been recognized as a new
area for dataset research. The problem of discovering
association rules was introduced in latter stages.
Given a set of transactions, where each transaction is
a set of items, an association rule is an expression of
the from X + Y, where X and Y are sets of items. The
problem is to find all association rules that satisfy
user- specified minimum support and minimum
confidence constraints.
Conceptually, this problem can be viewed as
finding associations between the "1" values in a
relational table where all the attributes are Boolean.
The table has an attribute corresponding to each item
and a record corresponding to each transaction. The
value of an attribute for a given record is "1" if the
item corresponding to the attribute is present in the
transaction corresponding to the record, "O" else.
Relational tables in most business and
scientific domains have richer attribute types.
Attributes can be quantitative (e.g. age, income) or
categorical (e.g. zip code, make of car). Boolean
attributes can be considered a special case of
categorical attributes. This thesis defines the problem
of mining association rules over quantitative attribute
in large relational tables and present techniques for
discovering such rules. This is referred as the
Quantitative Association Rules problem.
The problem of mining association rules in
categorical data presented in customer transactions
was introduced by Agrawal, Imielinski and Swami
[1][2]. This thesis work provided basic idea to several
investigation efforts [4] resulting in descriptions of
how to extend the original concepts and how to
increase the performance of the related algorithms.
221
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
The original problem of mining association
rules was formulated as how to find rules of the form
setl -> set2. This rule is supposed to denote affinity
or correlation among the two sets containing nominal
or ordinal data items. More specifically, such an
association rule should translate the following
meaning: customers that buy the products in setl also
buy the products in set2. Statistical basis is
represented in the form of minimum support and
confidence measures of these rules with respect to the
set of customer transactions.
The original problem as proposed by
Agrawal et al.[2] was extended in several directions
such as adding or replacing the confidence and
support by other measures, or filtering the rules
during or after generation, or including quantitative
attributes. Srikant and Agrawal describe a new
approach where quantitative data can be treated as
categorical. This is very important since otherwise
part of the customer transaction information is
discarded.
Whenever an extension is proposed it must
be checked in terms of its performance. The
algorithm efficiency is linked to the size of the
dataset that is amenable to be treated. Therefore it is
crucial to have efficient algorithms that enable us to
examine and extract valuable decision-making
information in the ever larger databases.
This thesis present an algorithm that can be
used in the context of several of the extensions
provided in the literature but at the same time
preserves its performance. The approach in our
algorithm is to explore multidimensional properties
of the data (provided such properties are present),
allowing to combine this additional information in a
very efficient pruning phase. This results in a very
flexible and efficient algorithm that was used with
success in several experiments using quantitative
databases with performance measure done on the
memory utilization during the transactional pruning
of the record sets.
II. LITERATURE REVIEW
Various proposals for mining association
rules from transaction data were presented on
different context. Some of these proposals are
constraint-based in the sense that all rules must fulfill
a predefined set of conditions, such as support and
confidence [6,7,8]. The second class identify just the
most interesting rules (or optimal) in accordance to
some interestingness metric, including confidence,
support, gain, chi-squared value, gini, entropy gain,
laplace, lift, and conviction [9,6]. However, the main
goal common to all of these algorithms is to reduce
the number of generated rules.
A)Existing Scheme
The thesis extend the first group of
techniques since it do not relax any set of conditions
nor employ a interestingness criteria to sort the
generated rules. In this context, many algorithms for
efficient generation of frequent item sets have been
proposed in the literature since the problem was first
introduced in [10]. The DHP algorithm [11] uses a
hash table in pass k to perform efficient pruning of
(&+7)-item sets. The Partition algorithm minimizes
I/O by scanning the dataset only twice. In the first
pass it generates the set of all potentially frequent
item sets, and in the second pass the support for all
these is measured. The above algorithm are all
specialized techniques which do not use any dataset
operations. Algorithms using only general purpose
DBMS systems and relational algebra operations
have also been proposed [9.10].
Few other works trying to solve this mining
problem for quantitative attributes. In [5], the authors
proposed an algorithm which is an adaptation of the
Apriori algorithm for quantitative attributes. It
partitions each quantitative attribute into consecutive
intervals using equi-depth bins. Then adjacent
intervals may be combined to form new intervals in a
controlled manner. From these intervals, frequent
item sets (c.f. large item sets in Apriori Algorithm)
will then be identified.
Association rules will be generated
accordingly. The problems with this approach is that
the number of possible interval combinations grows
exponentially as the number of quantitative attributes
increases, so it is not easy to extend the algorithm to
higher dimensional cases. Besides, the set of rules
generated may consist of redundant rules for which
they present a "greater-than-expected-value" interest
measure to identify the interesting ones.
Some other efforts that exploit quantitative
information present in transactions for generating
association rules[12]. In [5], the quantitative rules are
generated by discrediting the occurrence values of an
attribute in fixed-length intervals and applying the
standard Apriori algorithm for generating association
rules. However, although simple, the rules generated
by this approach may not be intuitive, mainly when
there are semantic intervals that do not match the
partition employed.
222
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
Other authors [5] proposed novel solutions
that minimize this problem by considering the
distance among item quantities for delimiting the
intervals, that is, their "physical"' placement, but not
the frequency of occurrence as a relevance metric.
To visualize the information in the massive
tables of quantitative measurements we plan to use
clustering and mutual information based on entropy.
Clustering is an old studied technique used to extract
this information from customer behavior data sets.
This follows from the fact that related customer
purchase through word of mouth have similar
patterns of customer behavior. Clustering groups
records that are "similar" in the same group. It suffers
from two major defects. It does not tell you how the
two customer buyin behavior/clusters are exactly
related. Moreover, it gives you a global picture and
any relation at a local level can be lost.
B) Proposed Scheme
The proposed scheme comprises of two
phases. The first phase of the thesis concerns about
the quantitative association rule mining with the
enhancement on Apriori algorithm. The second phase
of the thesis deals with the reduction of memory
utilization during the pruning phase of the
transactional execution.
The algorithm for generating quantitative
association rules starts by counting the item ranges in
the data set, in order to determine the frequent ones.
These frequent item ranges are the basis for
generating higher order item ranges using an
algorithm similar to Apriori. Take into account the
size of a transaction as the number of items that it
comprises.
a) Define an item set m as a set of items of size m
b) Specify frequent (large) item sets by Fm
c) Specify candidate item sets (possibly frequent) by
Lm.
A n range set is a set of n- item ranges, and
each m-item set has a n-range set that stores the
quantitative rules of the item set. During each
iteration of the algorithm, the system uses the
frequent sets from the previous iteration to generate
the candidate sets and check whether their support is
above the threshold. The set of candidate sets found
is pruned by a strategy that discards sets which
contain infrequent subsets. The algorithm ends when
there are no more candidates' sets to be verified.
The enhancement of Apriori is done by
increasing the efficiency of candidate pruning phase
by reducing the number of candidates that are
generated to further verification. The proposed
algorithm use quantitative information to estimate
more precisely the overlap in terms of transactions.
The basic elements considered in the development of
the algorithm are number of transactions, average
size of transaction, average size of the maximal large
item sets, number of items, and distribution of
occurrences of large item sets.
The second phase of the thesis claimed
improvement over A priori by considering memory
consumption for data transaction. This part of the
algorithm generate all candidates based on 2-frequent
item sets on sorted dataset and already generates all
frequent item sets that can no longer be supported by
transactions that still have to be processed. Thus the
new algorithm no longer has to maintain the covers
of all past item sets into main memory. In this way,
The proposed level-wise algorithms accesses a
dataset less often than Apriori and require less
memory because of the utilization of additional
upward closure properties.
C)Mutual Information based entropy
The mutual information I (X, Y ) measures
how much (on average) the realization of random
variable Y tells us about the realization of X, i.e.,
how by how much the entropy of X is reduced if we
know the realization of Y .
I(X; Y ) = H(X) - H(XIY )
For example, the mutual information between a cue
and the environment indicates us how much on
average the cue tells us about the environment. The
mutual information between a spike train and a
sensory input tells us how much the spike train tells
us about the sensory input. If the cue is perfectly
informative, if it tells us everything about the
environment and nothing extra, then the mutual
information between cue and environment is simply
the entropy of the environment:
I(X; Y ) = H(X) - H(XIY ) = H(X) - H(XIX) = H(X).
In other words, the mutual information
between a random variable and itself is simply its
entropy: I(X;X) = H(X). Surprisingly, mutual
information is symmetric; X tells us exactly as much
about Y as Y tells us about X.
223
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
III. QUANTITATIVE ASSOCIATION
RULE MINING - MUTUAL
INFORMATION BASED ENTROPY
The proposal of this work use mutual
information based on entropy for generating
quantitative association rules. Apart from the usual
positive correlations between the customers, this
criterion would also discover association rules with
negative correlations in the data sets. It is expected to
find results of the form attrib 1/ attrib 2 -> A attrib3,
which can be interpreted as follows: Attrib 1 and
Attrib2 are co expressed and have silencing effect on
attrib 3. Then compare the results from our
experiments to those obtained from clustering.
First tune various parameters (like support,
support fraction, significance level), of the auto
performance dataset. This was because even with
binary data, 2468 attributes may lead to the power(2,
2468) relations (which the software was not designed
to handle). Here, it is needed to know that for the
problem under consideration, the auto are attributes,
as needed to find relationships among them. To
overcome this problem used another approach, in
which data attributes were already known to be
related, using the results obtained from clustering.
This decreases the number of attributes to
manageable levels (both for program). The proposed
work used the approach above to find the
relationships (positive, negative) among the
attributes.
Algorithm Steps
a. Find all frequent item sets (i.e., satisfy
minimum support)
b. Generate strong association rules from the
frequent item sets (each rule must satisfy minimum
support and minimum confidence).
c. Identify the quantitative elements
d. Sorting the item sets based on the
frequency and quantitative elements
e. Merge the more associated rules of item
pairs
f. Discard the infrequent item value pairs
g. Iterate the steps c to f till the required
mining results are achieved
h.
Let I = { il, i2 ... im} be a set of items, and
T a set of transactions, each a subset of I. An
association rule is an implication of the form A=>B,
where A and B are non-intersecting The support of
A=>B is the percentage of the transactions that
contain both A and B . The confidence of A=>B is the
percentage of transactions containing A that also
contain B (interpret as P(BIA)). The occurrence
frequency of an item set is the number of transactions
that contain the item set.
IV. IMPLEMENTATION OF QUANTITATIVE
APRIORI
The function op is an associative and
commutative function. Thus, the iterations of the
foreach loop can be performed in any order. The
data- structure Reduc is referred to as the reduction
object. The main correctness challenge in
parallelizing a attribute like this on a shared memory
machine arises because of possible race conditions
when multiple processors update the same element of
the reduction object.
The element of the reduction object that is
updated in a loop iteration is determined only as a
result of the processing. In the a priori association
mining algorithm, the data item read needs to be
matched against all candidates to determine the set of
candidates whose counts will be incremented. The
major factors that make these loops challenging to
execute efficiently and correctly are as follows:
It is not possible to statically partition the
reduction object so that different processors update
disjoint portions of the collection. Thus, race
conditions must be avoided at runtime.
The execution time of the function process
can be a significant part of the execution time of an
iteration of the loop. Thus, runtime preprocessing or
scheduling techniques cannot be applied.
The updates to the reduction object are fine
grained. The reduction object comprises a large
number of elements that take only a few bytes, and
the for each loop comprises a large number of
iterations, each of which may take only a small
number of cycles.
{ * Outer Sequential Loop *}
While() {
{* Reduction Loop*}
Foreach(element e ) {
( i, val) = process (e);
Reduc (i) = Reduc (i) op val;
i
Fig 1 : Pseudo code
The consumer behavior auto databases
obtained data from UCI Machine Learning
Repository. The data obtained was about CPU-
performance and automobile mileage. The data was
discretized into binary values. For these data sets the
224
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
discretization was done in accordance with
interpretation required. This discretization was done
automatically using the written software. This
software also formatted the data into the format
required by the program. A finer level of
discretization (or supporting the real values) would
have been more appropriate, but the used approach
also gave much of the useful results.
V. EXPERIMENTAL RESULTS FROM APRIORI
The process of executing quantitative
Association rule mining for the auto manufacturer
power evaluation data is given below
a) Get data (file: auto-mpg.data): 9 attributes, 398
samples.
b) Remove unique attributes (IDs). Here, car_name
attribute has been removed.
c) Remove those samples (total 5) that contain "?"
(missing data) as a value for some of their attributes
(so, we are left with 8 attributes and 393 samples).
d) Discretize real-valued attributes based on their
average values (which is (maximum attribute value +
minimum attribute value) / 2)
e) Run the program to generate association rules
using mutual information based on entropy metric.
The experiment focused on evaluating all
quantitative a priori techniques. Since we were
interested in seeing the best performance, we used
banking data set samples. We used a 1 GB dataset. A
confidence of 90% and support of 0.5 is used.
Execution times using 1, 2, 3, and 4 threads
are presented on the processor. With 1 thread, Apriori
does not have any significant overheads as compared
to the sequential version. Therefore, this version is
used for reporting all speedup s. Though the
performance of quantitative a priori is considerably
lower than a priori, they are promising for the cases
when sufficient memory for supporting full
replication may not be available. We consider four
support levels, 0.1%, 0.05%, 0.03%, and 0.02%. The
execution time efficiency is improved for the
quantitative a priori on frequent item set evaluation
with the support count (Graph 1).
1 10
i.i.
1.4
1.2
a ■]
il.rt
II II
rmir
1 1 ofl
1 1
■'.I.-
11.11=..
u.u.?s
' '.' V :,
Support LevGl
Graph 1 : Support VS Time on Quantitative and qualitative A priori
The thread execution on the quantitative a
priori and qualitative a priori are evaluated for the
same data set (Graph 2). Here the initial thread
requires more time, however consequent threads
shows better scalable performance of quantitative
Apriori.
■i
.*;.■■■ ■
.*
.-. 2500
if!
■I'
z
h 2000
1
;,.„,
1
fr-
man
1 1
fr-int
n
Dfl-
man
□ afl-
int
i i
csl
-man
■
■il
-int
IU
1 thread
'■ WW- ■■)■[■
4 threads
No. of threads
Graph 2: Thread Vs Time for a priori execution on rule set
generation
225
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
By observing the output from the program it is seen
that a few relationships between the attributes had
high values of mutual information. Namely, the
highest Mi-values were obtained for:
a) displacement and horsepower. Further, by
observing the entropy values we may notice that
there are very few cars that have small displacement
and high horsepower.
b) displacement and weight. Further, by observing
the entropy values we may notice that there are very
few cars that have large displacement and light
weight.
c) cylinders and weight. Further, by observing the
entropy values we may notice that there are very few
cars that have small number of cylinders and heavy
weight.
d) horsepower and weight. Further, by observing the
entropy values we may notice that there are very few
cars that have large horsepower but heavy weight.
VI. CONCLUSION
The thesis have defined a new a rule set
namely the informative rule set that presents
prediction sequences equal to those presented by the
association rule set using the confidence priority. The
informative rule set is significantly smaller than the
association rule set, be especially when the minimum
support is small.
The proposed method has some merit in
extracting information from huge data sets by
pruning the initial information (to bring it down to
the manageable levels) and then finding the
association rules among the attributes. Further, the
approach is used to predict the relationships among
the silencer auto power, weight, model, year etc.,
could be extended to unknown function.
The proposed scheme have characterized the
relationships between the informative rule set and the
non-redundant association rule set, and revealed that
the informative rule set is a subset of the non-
redundant association rule set. The thesis considers
the upward closure properties of informative rule set
for omission of uninformative association rules, and
presented a direct algorithm to efficiently generate
the informative rule set without generating all
frequent item sets.
efficiency improvement results from that the
generation of the informative rule set needs fewer
candidates and dataset accesses than that of the
association rule set rather than large memory usage
like some other efficient algorithms.
REFERENCES
[I] R. Agrawal, T. Imielinski, and A. Swami. Dataset mining: A
performance perspective. In IEEE Transactions on Knowlegde and
Data Engineering, December 1993.
[2] R. Agrawal, T. Imielinski, and A. Swami. Mining association
rules between sets of items in large databases. In Proc. of the
ACM SIGMOD Washington, D.C., pages 207-216, May 1993.
[3] R. Miller and Y. Yang. Association rules over interval data. In
ACM SIGMOD Conference, Tucson, Arizona, pages 452 - 461,
May 1997.
[4] J. Park, M. Chen, and P. Yu. An effective hash based algorithm
for mining associative rules. In ACM SIGMOD Conference, San
Jose, CA, pages 175 - 186, May 1995.
[5] R. Srikant and R. Agrawal. Mining quantitative association
rules in large relational tables. In Proceedings of the ACM
SIGMOD Conference on Management of Data, pages 1-12,
Montreal, Canada, June, 1996.
[6] R. Agrawal, T. Imielinski, and A. Swami. Dataset mining: A
performance perspective. In IEEE Transactions on Knowlegde and
Data Engineering, December 1993.
[7] R. Agrawal, H. Mannila, R. Srikant, H. Toivonen, and
A.Verkamo. Fast discovery of association rules. In Advances in
Knowledge Discovery and Data Mining, San Jose, CA, pages
307-328,1996.
[8] R. Bayardo, R. Agrawal, and D. Gunopulos. Constraint- based
rule mining in large, dense databases. In 15th International
Conference on Data Engineering, Sydney, Australia, pages 188
- 197, March 1999.
[9] R. Bayardo and R. Agrawal. Mining the most interesting rules.
In 5th ACM SIGKDD International Conference on Knowledge,
San Diego, CA, Pages 145 - 154, August 1999.
[10] R. Agrawal, T. Imielinski, and A. Swami. Mining association
rules between sets of items in large databases. In Proc. of the
ACM SIGMOD Washington, D.C., pages 207-216, May 1993.
[II] J. Park, M. Chen, and P. Yu. An effective hash based
algorithm for mining associative rules. In ACM SIGMOD
Conference, San Jose, CA, pages 175 - 186, May 1995.
[12] Prakash.S and R.M.S.Parvathi. " Scaling Apriori for
Association Rule Mining Efficiency ".Proceedings of the Fourth
International Conference, Amrutvani College of Engineering,
Sangamner,Maharatra.pp 29, March 2009.
The informative rule set generated in this
thesis is significantly smaller than both the
association rule set and the non-redundant association
rule set for a given dataset that can be generated more
efficiently than the association rule set. The
226
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
AUTHORS PROFILE
Prof. S. Prakash has completed his M.E.,
(Computer Science and Engineering) in
K.S.R.College of Technology , Tamilnadu,
India in 2006. Now he is doing research in
the field of Association Rule Mining
algorithms. Currently, he is working as
Assistant Professor in the department of
Information Technology, Sasurie College of Engineering,
and Tamilnadu. India. He has completed 9 years of
teaching service.
Dr. R.M.S. Parvathi has completed her
Ph.D., degree in Computer Science and
Engineering in 2005 in Bharathiar
University, Tamilnadu, India. Currently she
is a Principal and Professor , Department of
CSE in Sengunthar College of Engineering
for Women, Tamilnadu, India, She has completed 20 years
of teaching service. She has published more than 28 articles
in International / National Journals. She has authorized 3
books with reputed publishers. She is guiding 20 Research
scholars. Her research areas of interest are Software
Engineering, Data Mining, Knowledge Engineering, and
Object Oriented System Design.
227
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
Clustering of High- Volume Data Streams
In Network Traffic
M.Vijayakumar
Department of Computer Science and Engineering
Sasurie College of Engineering,
Vijayamangalam, Erode(Dt), Tamilnadu, India.
tovijayakumar @ gmail.com
Dr.R.M.S.ParvathiM.E.(CSE),Ph.D.
Principal,
Sengunthar College of Engineering, for Women,
Tiruchengode,Tamilnadu, India.
rmsparvathi @ india.com
Abstract — The thesis concerned with the problem of
mining network traffic data discovering useful
associations, relationships, and groupings in large
collections of data. Mathematical transformation
algorithms have proven effective at reducing the content
of multilingual, unstructured data into a vector that
describes the content. Such methods are particularly
desirable in fields undergoing information explosions,
such as network traffic analysis, bio-informatics, and the
intelligence community. In response, traffic mining
methodology is being extended to improve performance
and sufficiently scalable.The usage of data flow collected
from site routers for various analysis i.e., network
performance characterization, investigating computer
security incidents and their prevention, network traffic
statistics, and others. Currently, the data flow analysis is
built as a distributed system to collect data from multiple
routers, both at the edge of the site network as well as
from local routers and multilayer switches. Average per
day volume is about 2GBytes of raw data. Despite a high
volume of collected information, some analysis is
conducted in near real time to satisfy demands of users
communities for quick results.The proposed work present
an efficient clustering means to analyze experimental
results for traffic data streams nature (symmetric and
asymmetric). As summary, this paper describe a system
designed to satisfy three primary goals i.e., real-time
concept mining of high-volume data streams, dynamic
data flow into a relational hierarchy; and adaptive
reorganization of the traffic data hierarchy in response to
evolving circumstances and network traffic time to time.
The proposed clustering network traffic data flow
collection and analysis system describe traffic
characterization and network performance estimation for
the data flow centre. The system checks the traffic
consistency for End To End circuits and Policy Based
Routing and finally, profiling of host's traffic to keep
track of their typical behavior to prevent accidental
blocking by site IDS system.
Keywords — Traffic analysis, network management ,
clustering, frequent Item set , hierarchical clustering.
I. INTRODUCTION
As part of an ongoing research project,
developed a novel algorithmic approach for extracting
traffic data from voluminous data streams[l][14]. The
approach is applicable to internet data servers, which
can be automatically identified and converted into a
common structure.
Here, report on an extension of the traffic data
clustering work, which clusters data flow hierarchically
based on the nature requested by users, (demand, load).
The new method represents the streaming hierarchical
clustering of documents [2] [18], which can be seen as a
subfield of the nascent discipline of "streaming AI",
"evolutionary clustering", or "AI in hardware".
The proposed system is a High Speed Content
classification system that works in three stages to
classify flows of TCP traffic. A TCP flow is half of a
TCP conversation. A connection from a client to a Mail
server with SMTP (simple mail transfer protocol) is an
example of a flow[4][26]. The connection from the
Mail server back to the client is considered a separate
flow. It extracts words then builds a vector
representation and then scores against known nature of
the data flow. The scores of the completed flows are
passed out of the system for evaluation.
The base data list from stage one used for
counting in stage two. Each dimension is represented
by 4bits[3]. Counting for a dimension saturates at 15.
When the flow ends the count vector representing the
flow is passed on to stage three. In the third stage a
vector representing a flow is scored against vectors
representing known traffic data[10][24][25]. The data
vectors are called the Score Table (ST) and are
reconfigurable at run time like the WMT. The ST is
derived from a set of documents. The output of the
228
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
system is a set of scores and the count array of the
flow.
Evaluation of the scores determines the
classification of the flow against the known traffic data.
However, simply classifying the flow as the data with
greatest score is not adequate. A forced classification
of all flows will be undesirable in most applications. A
threshold provides a confidence level to the
classification of flows. Any traffic data that is not
classified is considered unknown to the system.
Clustering these unclassified flows is the focus of the
Streaming Clustering work in this thesis.
II. HIERARCHICAL PARTITIONING
The notion that natural categories tend to be
organized hierarchically as a general principle dates
back at least to Aristotle, and has been refined by
Simon[5][ll][12]. Considering that libraries,
governments, Internet newsgroups and taxonomies, and
the human neocortex all organize and process
information hierarchically, it is certainly a natural
methodology for organizing unknown content. A
logical assumption for document clustering is that like
documents will group together. Groupings with large
number of items are more general.
As the number of items starts to dwindle, the
topic of a grouping becomes more specific. For
example, the topic of "cars" is very general, while
"Cadillac" is more specific and would be part of the
general topic of "cars." Two standard approaches can
be taken to the problem organizing a collection of
documents represented as fixed length vectors of high
dimensionality hierarchically agglomerative (bottom-
up), and divisive (top-down) [6] [13] [22].
Agglomerative -Cluster Algorithm
Assign each document to a unique cluster
While more than one cluster exists
Merge the two closest clusters
Return the single remaining cluster as the root
Divisive-Cluster Algorithm
If the collection contains a single traffic data
Return it in a single cluster
Divide the collection into two parts
Call divisive-cluster on each part
The final result for both procedures is a binary
tree containing a single data in every leaf, where close
leaves (measured by tree-traversal distance) are related.
Both methods have their general advantages and
drawbacks, but for the specific problem of data
clustering where traffic data are represented via a flow
sequence approach, empirical studies have shown that
the top-down divisive approach produces consistently
superior results.
The generally accepted explanation for these
results is that local neighborhood information (which
pairs of documents contain most similar word
distributions) is often misleading, deceiving
agglomerative clustering into making bad merge
decisions early on. Divisive clustering can often avoid
these mistakes by first considering the global statistics
of collections of traffic data, which are more robust.
Thus, a divisive hierarchical clustering approach is
expected to give us the best results; the remaining
decision that must be made is how to divide a
collection of documents.
The centroid gives an indication of the overall
makeup of a cluster, and is used in many clustering
algorithms. Assuming that our set of documents, V,
represents some cohesive grouping (e.g., a collection of
postings from a single internet newsgroup), its centroid
provides us with an indication of what concepts the
"average" document refers to the value in some
dimension will be between zero and one, and denote
the probability of a random document from the
collection scoring a hit in that dimension. It is natural
to consider measuring the affinity of a traffic data to a
given collection by comparing its distribution of hits to
our expectations normalized by data flow vector
magnitude, as data with higher magnitudes have more
opportunities for hits.
The heuristic approach that is taken to attempt
this is quite simple. A set of data- vectors is randomly
partitioned (a fair coin is flipped for each document to
assign it to either cluster one or cluster two). Data are
iteratively transferred between the two clusters, one at
a time, as long as doing so strictly increases the
division quality. Requiring a strict increase in division
quality ensures that the partitioning procedure is
guaranteed to terminate [7] [22].
The system will compare hierarchical
partitioning to flat and hierarchical variants of the
popular k- means clustering algorithm[8], and to a
baseline hierarchical clustering algorithm (bisection k-
means ) from the synthetic dataset. These traffic data
sets consist largely of flow of data at the network
points, demand, and load etc.,
III. K-MEANS CLUSTERING
The k-means clustering algorithm separates
input data into K groups. The number of groups, or K,
is set prior to running the clustering algorithm. Each
document in the data is assigned to a cluster. These
assignments are used to calculate the cluster centroid or
center[9][21][23]. The cluster centroids are then
229
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
utilized to determine the distance between each
centroid and a data element. The algorithm seeks to
minimize the inner cluster distance (i.e. form tight
groups of similar data) and minimize the inter cluster
distance (i.e. the groupings are non-overlapping).
The distance calculation can be performed in
any number of ways. Three common methods of
distance calculation include the Minkowski,
Manhattan, and Euclidean Distance metrics. The cosine
theta distance has also been used with k-means in order
to cluster high dimensional data. The algorithm is a
cyclical algorithm that performs in the following
manner
a) Initially assign document in the data to K groups
b) Calculate the cluster centroids based on assignments
c) For each document in the data
i) Recalculate distances from document to all
centroids and find closest centroid
ii) Change document assignment to closest centroid
and update the centroids that the document
used to reside and currently resides
d) Repeat step 3 until either no changes are made to
document assignments or the epoch limit is reached.
Bisection k-means is a variant of the k-
means algorithm. It starts with a single cluster and
continually selects a cluster to split into 2 sub-clusters
until the requested number of clusters is achieved.
a) Pick a cluster to split.
b) From the selected cluster, use k-means to cluster
the elements into 2 sub clusters.
c) Repeat steps 1 and 2 until the desired number of
clusters is found.
The selection of the cluster to bisect can be
done in a number of ways. For choosing the cluster
with the most elements was sufficient to find good
clustering, use the same heuristic here. In order to
objectively compare the results of a hierarchical
clustering algorithm to a flat clustering, need a means
of automatically flattening a full binary cluster tree to a
set of k clusters (given some particular k). A simple
heuristic for doing so is to choose the (non-
overlapping) subtrees that make the highest quality
clusters, as defined in the previous section. So for k=2,
will choose the left and right subtrees of the root. For
k=3 we will take the lowest scoring of our current
clusters (that is not already a leaf) and expand it into
two clusters by replacing it with its two children (recall
that divisive hierarchical clustering always produces a
binary tree). For bisection k-means, stop tree creation
after k leaves have been formed and take them as our
clusters.
The ground-truth newsgroups (horizontal axis)
are ordered so that the far right column of the
confusion matrix is the chaff, and the ordering of the
remaining newsgroups is arbitrary. The clusters
(vertical axis) are ordered based by their most frequent
newsgroup, with color showing purity (blue is lowest,
red highest). A perfect clustering would hence be
denoted by a crisp diagonal red line. K-means is clearly
inferior to the two hierarchical approaches, placing the
majority of the documents into two large clusters (the
two horizontal red lines near the bottom of the plot).
Bisection k-means and hierarchical partitioning
produce comparable results; however half of the
clusters created by bisection k-means are dominated by
chaff, whereas hierarchical partitioning creates only ten
such "junk" clusters. This is preferable from a human
analyst's point of view, as it allows uninteresting
document sets to be identified and discarded more
quickly.
IV. EXPERIMENTAL RESULTS
This part of the thesis describes experimental
results with the streaming hierarchical partitioning
algorithm. Results are presented on ISP data, streamed
according to a regime designed to simulate nature of
the traffic data are randomly collected, but to begin,
only data from half of the ISP server are evaluated. At
uniform intervals, a ISP is gradually introduced into the
distribution (and hence the old newsgroup density
gradually reduced).
The traffic flow appears uniformly throughout
the entire data stream. In order to run streaming
hierarchical partitioning clustering, need to set two
parameters, the maximal number of traffic data that are
capable of collecting at a time, and how often the set of
data flowing in the server working memory will be
reclustered. In actual deployment of course, will want
to store as many traffic data as possible in fast memory,
and recluster as often as possible, and the bandwidth of
the data stream. Given that our dataset contains about
10GB of data should prove illustrative.
To analyze the quality of hierarchical
streaming clustering, there are two basic factors
consider, traffic quality as in the non- streaming case,
how meaningful are the clustered data discovered by
the algorithm, and traffic discovery as drift occurs ,
does the algorithm effectively identify new flow of
traffic data.
In a controlled experiment, where know the
ground-truth labeling of the data, traffic quality can be
measured by considering how many non-chaff data are
230
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
assigned to clusters that are nearly "pure" (at least 90%
of the documents originating in a single newsgroup).
Furthermore, only consider clusters with more than 10
documents in computing this measure, henceforth
referred to as a purity score.
Similarly, traffic discovery over time can be
measured by considering, at any given time, how many
pure clusters have been created corresponding to
unique non-chaff labels, up until this time. The
measure is hence cumulative, and will henceforth be
referred to as a discovery score. In order to understand
how effective the data insertion and removal heuristics
are at augmenting hierarchical partitioning for
streaming clustering. In this procedure, full hierarchical
partitioning clustering will still be carried out at the
same regular intervals. However, when new data are
congested, traffic will be chosen entirely at random to
be removed. This methodology will henceforth be
referred to as naive streaming clustering.
In order to provide a fair comparison between
naive and non-naive streaming clustering, scores are
only computed immediately after batch clustering has
been completed. That is, naive streaming clustering is
not penalized for having poor results in between batch
clustering. Both methods are equally effective in terms
of purity scores as expected, since this score is
essentially determined by the batch clustering, which is
identical.
V. PERFORMANCE EVALUATION
Analysis of flow data gathered from multiple
routers is the source of valuable information for various
tasks. We distinguish three major areas where that
information is used. These areas are computer security,
performance analysis for data movement applications
and verification of traffic consistence across site
network infrastructure. We have considerable
experience to use flow data in two first areas. The last
one, verification of traffic consistency is relatively new
to us. It is motivated by recent deployment End To End
circuits for LHC/CMS experiment. These circuits have
predictable performance characteristics and dynamic
capabilities to use alternative WAN path available for
the data centre.
Traffic is analyzed for different time intervals,
typically lmin, 5min, 15min, lhour, or 1 day. A start
date and time can be also specified to generate historical
results. Tagging of raw data can be done at any time, but
for most applications, it runs when new flow sets are
created and become available on NAS server. New
intervals can be added as needed. The next level is the
target, which represents defined list of IP blocks assigned
to particular users group or community. The system Tier
1 LAN is an example of a user group address block.
ill Data
Actual Traffic
»
Monthly
Daily
Monthly
Bord9r
600MB
17GB
15IBP)
300TB PTB)
StaUi
1MB
KB
BOTPTB)
upep)
H
1,2GB
30GB
8GTB(120TB)
1.5PB(2.4PB)
m.
3GB
40GB
1
1
Table 1 : Volumes of flow data and actual traffic
A target's traffic is going to be inspected based on
set of predefined filters. For example, analyze appropriate
traffic to specific destinations. For each target there may be
several filters defined. Filters have unique names and are
stored in a database with time stamp allowing to track
changes. While inspecting traffic, we may use some
specific names for destinations or sources. For example,
we don't need to known all details of traffic to each node
of the destination cluster. Instead, we need to know traffic
to whole cluster, and use a predefined name for the address
block. The next level is the static definition of rules
specified in separated XML files. This level is identified
by name of XML configuration file. The content of
configuration files may be changed over time, due
different reasons, i.e. readdressing of nodes, moving them
to different subnets and so on. To track these changes, the
content of configuration files is stored in database with
time stamp. Anything that is not statically defined will be
resolved based on DNS level, i.e. top level, second, third
and so on, as needed.
ipaddr name
hosiCouni
octets
Flows
packets
duration
mmm |kmq-iw-i .sun.coM
53677
5523542
67220
125026
34003
221.215.63.230
39694
1634658
39595
39946
34995
83.11 7.22.1 V. c53?5t&7(.cable.wanadoc-.nl
32763
mm
327fi9
55019
34995
212.6I105.1S cd4J06912.cable.wanadoo.nl
32762
mm
32792
56649
34995
66.1 40.1 05.98 |ad5l>BG*1 4a-105-9B.dsl.lbcPrtx.swholl.not
32757
3602024
33433
59323
34993
8230.176.81 |hosl&1 -1 7fi-sta1ic.30-fl2-b.busiiies5.telecoinilalia.it
32750
mm
33223
50151
34995
90.1.63.17 ALille-153-1-100-17.w90-1.abo.wanadQO.fr
32735
3091255
33065
50815
34995
208.1 57.1 65.93 amg>bb<dsl<348.dsLairslruniconim.net
32733
6234200
62393
102200
34995
201.66,29.248 201^-29-Mclaje700.dSLbrasil:eleron.neLbr
3272&
3870511
32751
63451
34933
83.2B.73.1 75 |bgj175.neoplus.adsl.tpnstpJ
32724
3564D2Q
33161
SB544
34993
21 6. 1 71 .1 77.4S |acc-rssiel-2 16-171-1 77-46 . 1 77.1 71 .21 6.in -add r^rpa
32715
2981603
33021
49002
34993
Table 1: ISP Data traffic flow output
231
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
Knowing transfer rates for data movement
applications is very important. Often, neither SNMP
monitoring of router/switch interfaces, nor applications
themselves can provide that information. For this purpose,
use passive monitoring based on net flow information.
First, on the regular basis, we generate reports
with topN senders, receivers and conversations for
specified interval of time between the entities that we need
to monitor. The reports are produced using the results of
traffic tagging described earlier. The output of the WEB
based interface is very similar to Top Scanning reports.
Our objective is to detect conditions automatically
and notify if it becomes a long term situation. After
preliminary investigation, determined that flow data
collected from routers at the edge of the campus network
could be used for that purpose by comparing flow rates at
various points of network. Flow rates for a symmetric path
should be very close in both directions, inbound and
outbound. The proposed cluster tool is used interactively to
show asymmetric traffic conditions. Currently, we are
working on automated alerts.
traffic at work group router is almost symmetric. However,
at the border router observed only outbound flows (second
graph) and the high impact infrastructure providing
alternative path, only inbound flows are observed. Thus,
by comparing flow rates at the different routers along
potential traffic path successfully detect asymmetry
conditions.
VI. CONCLUSION
The proposed model describe a system for
extracting traffic data based on the nature requested by the
network administrator with an effective clustering system.
The implemented system, streaming hierarchical
partitioning, hierarchically clusters data streaming content.
The performance predictions include the quality
of clustering, and traffic data discovery. The streaming
hierarchical clustering algorithm was able to improve the
ability to discover traffic data of required nature. The
system has been prototyped and tested on a Xeon
processor as well as on a PowerPC. To implement
additional streaming functionality, some of the same
circuitry can be reused in particular the computation of
similarity of traffic data insertion.
In the future, plan to additionally move towards a
system that, Integrates clustering into our classification
system, continually searches for new and emerging traffic
data sets on larger inter networks, allows the resolution of
data to fade over time to allow for streaming with infinite
flow of data sets.
REFERENCES
Tine
Graph 1 : Detection of traffic asymmetry
IDBD 1
u
0.00 L
'i-
Id. DC IS
ll"
4Q.Q0. M
00.00 mM
mm Ifl
1DD.DD ^-|
iiddd 1
i
u
[2]
I.I
[3]
B
H
g
Ifl
in
Q
in
in
';:
pi
*
*
o
pi
A
Tt
■:■
■■■
o
pi
frj
n
in
•->
*
GQ
o
*
o
o
o
o
o
P
n
*
F-
"■
r
>
f*
*■
*
r*
^
r
N
n
N
N
M
'-
■-
■-
w
*
o
<D
'-'
O
10
<D
■-
■■■■
o
°
o
°
°
o
o
o
[41
Time
Graph 2: Detection of traffic asymmetry
The graph 1 and 2 visualize the idea of steering
production traffic. The traffic switched into a high impact
path is almost always symmetric in all routers. However,
traffic is going to be asymmetric at the border and the
point of presence. From graphs at the right we can see that
[5]
[6]
Abdun Naser Mahmood, Christopher Leckie, and Parampalli
Udaya " An Efficient Clustering Scheme to Exploit
Hierarchical Data in Network Traffic Analysis ". In IEEE
Transactions on Knowledge and Data Engineering, Vol 20,
No 6, June 2008.
Cao J., D. Davis, S. Vander Weil, and B. yu, "Time- Varying
Network Tomography", J. Am. Statistical Assoc, 2000.
Chadi Barakat, Patrick Thiran, Gianluca Iannaccone,
Christophe Diot, and Philippe Owezarski. Modeling internet
backbone traffic at the flow level. IEEE Transactions on
Signal Processing, 51(8):21 11-2124, August 2003. [2] Graham
Cormode, S.Muthukrishnan, and Irina
Dai , B.-R. J.-W. Huang, M.-Y. Yeh, and M.-S. Chen,
"Adaptive Clustering for Multiple Evolving Streams," IEEE
Trans. Knowledge and Data Eng., vol. 18, no. 9, pp. 1166-
1180, Sept. 2006.
Goldschmidt O., "ISP Backbone Traffic Inference Methods to
Support Traffic Engineering", Proc. Internet Statistics and
Metrics Analysis Workshop (ISMA '00), Dec. 2000.
Guha S, A. Meyerson, N. Mishra, R. Motwani, and L.
O'Callaghan, "Clustering datastreams: Theory and practice,"
IEEE Transactions on Knowledge and Data Engineering, vol.
15, no. 3, pp. 515-528, 2003.
232
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
[7] Guha S., N. Mishra, R. Motwani, and L. O'Callaghan,
"Clustering Data Streams", 41 st Annual Symposium on
Foundations of Computer Science, 2000.
[8] Henzinger M. R. , P. Raghavan, and S. Rajagopalan,
"Computing on Data Streams", External Memory Algorithms,
Boston: American Mathematical Society, 1999.
[9] Huang Z, M. Ng, H. Rong, and Z. Li, "Automated Variable
Weighting in k-Means Type Clustering," IEEE Trans. Pattern
Analysis and Machine Intelligence, vol. 27, no. 5, pp.657-668,
May 2005.
[10] Hun-Jeong Kang, Hong-Taek Ju, Myung-Sup Kim and James
W. Hong, "Towards Streaming Media Traffic Monitoring and
Analysis", APNOMS 2002, September 2002, Jeju, Korea.
[11] Jacobus van der Merwe, Ramon Caceres, Yang-hua Chu, and
Cormac Sreenan,"mmdump- A Tool for Monitoring Internet
Multimedia Traffic," ACM Computer Communication Re-
view, Vol. 30, No. 5, 2000.
[12] Jensen C , D. Lin, and B.C. Ooi, "Continuous Clustering of
Moving Objects," IEEE Trans. Knowledge and Data Eng., vol.
19, no. 9, pp. 1 161-1 173, Sept. 2007.
[13] Kuman A, M. Sung, J. Xu, and J. Wang, "Data Streaming
Algorithms for Efficient and Accurate Estimation of Flow Size
Distribution", Proc. ACM SIGMETRICS, 2004.
[14] Lakhina A, K. Papagiannaki, M. Crovella, C. Diot, E.
Kolaczyk, and N. Taft, "Structural Analysis of Network
Traffic Flows", Proc. ACM SIGMETRICS '04, June 2004.
[15] Lakhina A, M. Crovella, and C. Diot, "Characterization of
Network- Wide Anomalies in Traffic Flows", Technical
Report BUCS-2004-020, Boston Univ., 2004.
[16] Lockwood J. W, S. G. Eick, J. Mauger, J. Byrnes, R. P. Loui,
A. Levine, D. J. Weishar, and A. Ratner, "Hardware
Accelerated Algorithms for Semantic Processing of Document
Streams", 2006 IEEE Aerospace Conference, March 4-11,
2006.
[17] Lockwood J. W, S. G. Eick, D. J. Weishar, R. P. Loui, J.
Moscola, C. Kastner, A. Levine, and M. Attig,
"Transformation Algorithms for Data Streams, 2005 IEEE
Aerospace Conference", March 5-12, 2005.
[18] Li W , W.K. Ng, Y. Liu, and K.-L. Ong, "Enhancing the
Effectiveness of Clustering with Spectra Analysis," IEEE
Trans. Knowledge and Data Eng., vol. 19, no. 7, pp. 887-902,
July 2007.
[19] Madina A , K. Salamatian, N. Taft, I. Matta, and C. Diot, "A
Two Steps Statistical Approach for Inferring Network Traffic
Demands", revision of Technical Reports BUCS- TR-2003-
003,Mar.2004.
[20] Maulik U and S. Bandyopadhyay, "Performance Evaluation of
Some Clustering Algorithms and Validity Indices," IEEE
Trans. Pattern Analysis and Machine Intelligence, vol. 24,
no.12, pp. 1650-1654, Dec. 2002.
[21] Mouratidis K , D. Papadias, S. Bakiras, and Y. Tao, "A
Threshold-Based Algorithm for Continuous Monitoring of K
Nearest Neighbors," IEEE Trans. Knowledge and Data Eng.,
vol. 17, no. 11, pp. 1451-1464, Nov. 2005.
[22] Patrikainen A and M. Meila, "Comparing Subspace
Clusterings," IEEE Trans. Knowledge and Data Eng., vol. 18,
no. 7, pp. 902-916,July2006.
[23] Su M.-C. and C.-H. Chou, "A Modified Version of the It-
Means Algorithm with a Distance Based on Cluster
Symmetry," IEEE Trans. Pattern Analysis and Machine
Intelligence, vol.23, no. 6,pp. 674-680, June 2001.
[24] Subhabrata Sen and Jia Wang, "Analyzing Peer-to-Peer Traffic
Across Large Networks", IMW2002 Workshop, 2002,
Marseille, France.
[25] Vijayakumar M and R.M.S.Parvathi. " Concept Mining of
High- Volume Data Streams in Network Traffic using
Hierarchical Clustering'Mn Proceedings of the Fourth
International Conference, Amrutvani College of Engineering,
Sangamner,Maharatra. Proc. ITECH '09 pp 69 , March 2009.
[26] Wang J, D. Miller, and G. Kes dis, "Efficient Mining of the
Multidimensional Traffic cluster Hierarchy for Digesting
Visualization, and Anomaly Identification", IEEE J. Selected
Areas ofComm. Vol. 24, no. 10, pp. 1929-1941, Oct. 2006.
[27] Xu R and D. Wunsch II, "Survey of Clustering Algorithms,"
IEEE Trans. Neural Networks, vol. 16, no. 3, pp. 645-678,
2005.
[28] Yip K.Y.L., D.W. Cheng, and M.K. Ng, "HARP: A Practical
Projected Clustering Algorithm," IEEE Trans. Knowledge and
Data Eng., vol. 16, no. 11, pp. 1387-1397, Nov. 2004.
AUTHORS PROFILE
M.Vijayakumar has completed
his Bachelor of Engineering in
Computer Science in Bharathiar
University, Tamilnadu, India and
Master of Engineering , in
Computer Science in Anna
University,Chennai , Tamilnadu,
India. He has started his teaching
profession in the year 2004 in
Muthayammal Engineering
College,Tamilnadu. India., At
present, he is working as an Assistant Professor in the
department of Computer Science and Engineering in Sasurie
College of Engineering, Tamilnadu. India. He has published
15 research papers in National and International journals and
conferences. Currently he is a part time Research Scholar in
Anna university of Technology, Coimbatore. His areas of
interest are Data mining, Knowledge Engineering, Clustering
algorithms and Network security. He is a life member of ISTE.
Dr. R.M.S. Parvathi has
completed her Ph.D., degree in
Computer Science and
Engineering in 2005 in
Bharathiar University,
Tamilnadu, India. Currently she
is a Principal and Professor ,
Department of Computer
Science and Engineering in
Sengunthar College of
Engineering for Women,
Tamilnadu, India, She has completed 20 years of teaching
service. She has published more than 28 articles in
International / National Journals. She has authorized 3 books
with reputed publishers. She is guiding 20 Research scholars.
Her research areas of interest are Software Engineering, Data
Mining, Knowledge Engineering, and Object Oriented
System Design. She is a life member of ISTE and Indian
Computer Society.
233
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
A.R.Q. techniques using Sub-block retransmission for wireless networks
A.N.Kemkar, Member, ISTE and Dr. T.R.Sontakke MemberJSTE
Abstract — In this paper we mainly focus our
investigation on the throughput performance in
conjugation with sub-block transmission scheme.
The throughput of a wireless data
communications system depends on a number of
variables, one of it is length of the message
blocks. Over a noisy communication medium like
wireless medium used for mobile ad-hoc network,
our propose scheme performs effectively. In
propose scheme random length of the message is
divided in to fixed length blocks and applying
ARQ techniques if the error occurs. A threshold
model is used for fading channel, estimation and
CRC detection codes are used. Comparison of
transmission efficiency of proposed scheme with
varying channel condition is shown.
.Index Terms— FEC, Hybrid ARQ, BER.
1. INTRODUCTION:
Wireless channels are highly affected by
unpredictable factors such as co-channel
interference, adjacent channel interference,
propagation path loss, shadowing and multi path
fading. The unreliability of media degrades the
transmission quality seriously. Automatic Repeat
ReQuest (ARQ) and Forward Error Correction
(FEC) schemes are frequently used in wireless
environments to reduce the high bit error rate of
the channel.
As we have seen, the throughput efficiencies of
all the basic ARQ schemes are functions of the
packet size n. [1],[2],[3],[4],[5]. Our main result
is a mathematical technique for determining the
block size as a function of the other variables like
BER, signal-to-noise ratio.
A.N.Kemkar 1 ,S.R.T.M.U,Nanded.
+91-9819150392, ankemkar@gmail.com
Dr.T.R.Sontakke 2
Ex.Director - S.G.G.S.I.T.E.- Nanded
Principal,Sidhant college of Engineering
Pune.+91-
9822392766,trsontakke@gmail.com
In an attempt to improve throughput
performance, we have included an analysis using
forward error correcting (FEC) block codes (used
in Hybrid ARQ). The optimum amount of FEC
coding was found to be dependent upon the Block
length. As the Block length increases, the number
of correctable errors to optimize the throughput
also increases, mathematical expression is shown
in 2.2.
The paper is organized as follows. In Section 2
.Summary on the Related work and basic concept.
In section 3. scheme description and system
model. We consider the performance analysis of
the proposed scheme for simulation study in
section 4. followed by conclusion in Section 5
2.RELATED WORK :
2.1: Related Work- The efficiency of HARQ
scheme is compared with GBN schemes using
different lengths of IP Blocks. Further show that
usage of smaller Blocks and hybrid schemes leads
to an improved throughput. Differences between
pure and hybrid GBN schemes are also discussed.
[i]
When the channel is quiet the sub-block
retransmission scheme behaves like a
conventional ARQ or hybrid ARQ scheme. As
the channel becomes increasingly noisy, the data
block is divided into smaller sub-blocks for
transmission. Each sub-block is encoded for error
control by an appropriate shortened code of
which the code length is adapted to the
corresponding channel BER. [2] Further optimum
block size in accordance with the channel
conditions [4] A single code HARQ scheme was
proposed in which transmitter is operating in any
one mode with the degree of errors encounter.
The operating state is selected based on the
channel BER. Data bits are divided in blocks and
are encoded with shortened codes. During the
retransmission new coded blocks are combined
and at the receiver end proper decoding
techniques are used to separate retransmitted
blocks from the new blocks.
2.1. BASIC CONCEPT -
234
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
Analytical expression how Throughput
performance of the system varies with the size of
the block length and FEC. : Consider the
following two cases to verify the Throughput
performance of the system.
Case 1. Blocks are transmitted without FEC.
Case 2. Blocks are transmitted with FEC.
Our analysis includes the following simplifying
assumptions: 1. The CRC decoder detects all
errors in the output of the FEC decoder.2.
Transmission of acknowledgments from the
receiver to the transmitter is error free and
instantaneous.
System throughput (P) is the number of payload
bits per second received correctly:
„ K
RfiY)
(1)
where
b/s is the payload transmission rate
KR
L
Where f(Y) = Block success rate defined as
probability of receiving block correctly.
Probability is a function of signal to noise ratio.
Eb
No
(2)
p
In which Eb = —joules in received energy per bit.
R
where R= Transmission rate. Probability is a
function of signal to noise ratio.
= Eb = P_
N No*
(3)
Each Block, of length L bits, is a combination
of a payload [K] and overhead (L-K) . Because
the Block success rate,/(7) is a decreasing
function of,£ there is an optimum Block length,.
V . When
Lp l* excessive overhead in each Block limits
the throughput. When L > l* Block errors limits
the throughput.
For case 1. When there is no forward error
correction coding. In this case
f(7) = (l-Pe(r)) L (4)
where f(7) = block success rate defined as
probability of receiving block correctly.
Pe{ y) is block error rate.
Therefore, in a system without FEC, the
throughput as a function of L ,from (1)
T=^R(i-p e {y)) L (5)
Case 2.: Now instead of transmitting those L bits
with no error correction capability, we will now
add B error correcting bits and transmit a total of
bits L+B .Using a block code forward error
correction scheme, the minimum number of B
bits required to correct t errors is given by [5]
(6)
B > log 2
t
I
n=0
Now that we can correct 4 1 ' errors, our block
success rate,/(7) should be larger than its
previous value with no error correction. Recall
that /(/) with t= is given by:
f(y) = (l-p e (y)) L where P e (y) is the probability
of a bit error as a function of the SNR. Now, with
error correction capability, the Block success rate
for some arbitrary value oft is [7]
^-Pe[r)) L+B - n (7)
\P n e{7)
ft(r)= E
Our new equation for the throughput as a function
of the signal to noise ratio is:
P
T(Y):
L-C
L+B
N,
ft(r)
(8)
From (5) and (8) it is clear that throughput of the
system is a function of message block length.
Further (5) and (8) are used for pure and hybrid
ARQ techniques.
3. SCHEME DISCRIPTION AND SYSTEM
MODEL :
This paper presents a sub-block retransmission
scheme for ARQ . The data block is divided into
smaller sub-blocks for transmission. Each sub-
block is encoded by an appropriate error detection
codes. The encoded block is then transmitted. The
received block is checked for errors sub-block by
sub-block. The proposed scheme provides
improved throughput by retransmitting only the
sub-blocks in the occurrence of errors.
3.1 SYSTEM MODEL :
235
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
We consider an ad-hoc network with V nodes
and assume that each node is equipped with only
one antenna. A Point to Point protocol is used at
the medium access control layer. A Selective
repeat request ARQ mechanism is used.
Particularly, the source node transmits a data
packet with a C-bit CRC attached. The
destination node detects CRC and then sends an
acknowledgement that is either positive (ACK) or
negative (NACK) back to the source node. If the
packet is correctly detected by the destination
node (with ACK feedback), the source node
continues to transmit a new data packet and the
above process is repeated. Otherwise,
retransmission will start. A threshold model for
channel characterization is used for fading
channel.
4. PERFORMANCE ANALYSIS OF THE
PROPOSE SCHEME:
The performance analysis of the scheme is
measured in terms of throughput of the proposed
scheme. Further we show the comparison of
throughput with sub block and without sub block
transmission schemes.
Expression of throughput for ARQ for present
scheme:
K
V =
E[T]
(9)
where K =information bits in a block. E[.\
=Expectation of number of transmitted bits in a
given block.
T = Mn + ^Ti (10)
i=\
where M =number of sub blocks, n =number of
bits in a sub block, Ti =number of transmitted bits
for i th transmission.
oo
E[T]=Mn + Y,E[T i ] (11)
1=1
where E[T\ =Average number of transmitted
bits.
Out of M sub blocks if L sub blocks are
transmitted at the f 1 retransmission ,then random
variable, Ti takes the value Ln ,if L out of M
sub-blocks are retransmitted at the f 1
retransmission.
SIMULATION RESULTS: We evaluate the
performance of the proposed scheme
implemented with Matlab. We run the simulation
for two schemes i.e. with sub block transmission
and without sub block transmission. The
simulation parameters are shown in the table
1. Simulation run for 5000 total blocks. Result is
the average of independent experiments where
each experiment uses different randomly
generated uniform parameters. We use mean
values which are obtained independent
experiments as a basic data to get the result.
Simulation results are shown Table 2
Table 1: System Parameters:
Parameters
Notaion
Values
Signal to Noise Ratio
7
Varied
Total number of blocks
5000
Total sub block
M
32
Information bits in a block
K
16
Packet length
n
5000*32*16
Max. number of
Retransmissions
3
Number of sub blocks
retransmitted
L
Varied
Cyclic Redundancy Check
CRC
Varied
Bit error rate
BER
Varied
Packet error rate
PER
Varied
Throughput efficiency
V
Varied
Table 2 : Simulation Results: Following
Simulation results shows the comparison of
Throughput efficiency verses varied block size
verses changing channel condition in terms of
PER.
Table 2:
Block length
Packet error rate
Throughput
Without
sub block
With
sub
block
Without
sub
block
With
sub
block
Without
sub
block
With
sub
block
Whole
block is
transmitted
with out
sub
division.
Whole
block
is
divided
into
sub 4
blocks
0.1
0.1
0.9
0.99
0.3
0.3
0.86
0.96
0.5
0.5
0.66
0.94
0.7
0.7
0.59
0.93
0.9
0.9
0.57
0.9
1
1
0.5
0.89
236
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
5. CONCLUSION:
From the Fig.l shown below performance of the
proposed scheme. This paper has presented a sub-
block retransmission scheme for
1. With micro block ARQ
2. Without micro block ARQ
Proposed sub-block retransmission schemes
showed better overall performance compared to
the other competitive schemes by retransmitting
without sub-blocks in the occurrence of errors.
From Table 2 and the Fig.l it is clear that as the
packet error rate increases i.e. channel condition
gets deteriorated the throughput performance
goes down by more than 40% as compare to
proposed scheme. There fore we proposed sub-
block retransmission scheme with ARQ is more
reliable than existing i.e. without sub block
transmission scheme.
6. REFERENCES:
[1] W.W. Chu. Optimalmessage block size
for computer communications with error
detection and retransmission strategies.
IEEE Transactions on Communications,
COM-22:1516- 1525, October 1974.
[2] J.S. DaSilva, H.M. Hafez, and S.A.
Mahmoud. Optimal packet length for
fading land mobile data channels. In
Proceedings of ICC 1980, pages 61.3.1-
61.3.5, June 1980.
[3] R.L. Kirlin. Variable block length and
transmission efficiency. IEEE
Transactions on Communication
Technology, COM-17:350-355, June
1969.
Fig.l: Performance of proposed scheme vs.
existing schemes.
Throughput Vs. Packet E rro r Rate
) block
[4] E. Modiano. An adaptive algorithm for
optimizing the packet size used in
wireless ARQ protocols. Wireless
Networks, 5:279-286, July 1999.
[5] J.M.Morris. Optimal block lengths for
ARQ error control schemes. IEEE
Transactions on Communications, COM-
27:488-493, February 1979
0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.
Packet E rro r Rate
237
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, 2010
Performance Analysis of Delay in Optical Packet
Switching Using Various Traffic Patterns
A.Kavitha/Chettinad College of Engineering &
Technology
IT dept
Chettinad College of Engineering & Technology,
Karur, Tamilnadu, India
V.Rajamni/Indra Ganesan College of Engineering,
Indra Ganesan College of Engineering
Trichy, Tamilnadu, India
P.Anandhakumar/Madras Institute of Technology,
IT Dept
Chennai, Tamilnadu, India
Abstract — Quality of Service parameters are improved
for development of optical packet switching technology.
Delay is an important parameter in optical packet
switching networks and it affects the performance of the
network. In this paper, a mathematical model is presented
to evaluate the delay rate. Delay rates are analyzed for
fixed packet length and variable length packet for various
traffic patterns viz. Non-uniform, Poisson and ON-OFF
traffic models for various service classes using
Reservation Bit technique. The results are compared with
the existing port based First-Fit wavelength assignment
algorithm. Here delay rates are reduced by 29% in our
class based model than the port based model.
Keywords-component; Optical Packet Switching (OPS),
RB (Reservation Bit algorithm), FF (First-Fit Wavelength
assignment algorithm), Quality of Service (QoS), Packet
Loss Rate (PLR), BER (Bit Error Rate), WDM
(Wavelength Division Multiplexing).
I Introduction
Due to the explosive growth of internet applications in
recent years, data traffic has been exceeded the telephony
traffic and bandwidth demands have been continuously
increasing. It is also expected of the future networks to
transport heterogeneous traffic services including
multimedia and interactive applications necessitating
bandwidth guarantees, minimum delay, less PLR, controlled
jitter and etc. QoS provisioning seems therefore a mandatory
task. Optical networks offer an extremely high traffic
bandwidth capable of providing communication channels
for several hundred nodes. Thus, the network traffic requires
the network to evolve by increasing transmission capacity of
optical fibers as well as switching capability. There are three
switching schemes in optical networks namely optical
circuit switching, optical burst switching (OBS) and optical
packet switching (OPS). In the optical circuit switching, a
dedicated end-to-end light path is established for each
connection. Thus the transmission delay can be guaranteed
and there is no virtually any loss, but less utilization of
wavelength in this technique. In OBS, data is sent in bursts
and a burst control message is sent ahead of each data burst
to reserve a wavelength at each hop based on the expected
arrival time of the data burst. In OPS, messages are
transmitted in packets. At each switching node, the packet
head is processed in the electrical domain for routing
purpose and the packet data is kept in the optical domain.
Wavelengths can be efficiently used in OPS [1]. Thus the
optical packet switching has emerged as one of the most
promising technologies for future telecommunication
networks. OPS utilize very high bandwidth in the optical
fiber using WDM. WDM offers an aggregate throughput of
the order of terabits per second. WDM is widely becoming
accepted as a technology for meeting growing bandwidth
demands, and WDM systems are beginning to be deployed
in both terrestrial and undersea communication links. Thus,
WDM offers an excellent platform for carrying IP traffic.
Consequently, OPS technology has many advantages, attract
more intensive attention than ever. The next generation
telecom infrastructure definitely comprise of optical
networks with improved QoS.
In order to improve the performance of QoS in optical
packet switching network, a detailed study has been made in
this paper. The performance analysis of optical packet
switching consists of two important issues namely packet
loss and delay. Inorder to provide better QoS in optical
packet switching, PLR and delay should be reduced. Packet
loss and delay are not new issues in optical networks;
however minimum loss and delay provide better QoS in
Optical packet switching. In our earlier report the
performance of 8B/10B code, Systematic code and Viterbi
code in optical transmission in terms of Bit Error Rate has
been analyzed. In the physical layer, transmission is done on
bit by bit basis and Bit Error Rate has been reduced in the
physical layer. When the bit errors in the physical layer are
rectified, it will reduce the packet loss rates in the higher
layers [2,3]. We have already reported that PLR has been
238
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, 2010
reduced in non-uniform traffic pattern [4]. We have also
implemented a RB algorithm for minimizing PLR in
buffered non-uniform traffic pattern of optical packet
switching mechanism [5].
In this paper we have analyzed the asynchronous OPS in
terms of delay. Delay rate for fixed length packet and
variable packet length has been studied for various traffic
patterns viz. Non-uniform, Poisson and ON-OFF traffic
models.
This paper has been organized as follows: Section II
describes the architecture of asynchronous OPS. In section
III, the analysis of delay rate is carried out for various traffic
patterns in asynchronous OPS. In section IV, results along
with discussions are presented and Section V deals with the
conclusion and future work.
II. Description of Architecture
The architecture used in this paper is presented in [4,5]. It
has been reproduced for the reference. The size of switch
under consideration is N X N. The switch has F input fibers
and F output fibers. By utilizing WDM, each fiber provides
N wavelengths to transport data with a capacity of C bps.
Buffers with the size of 5 are used in the OPS switch
catering to each of the service classes. The switching
process in OPS can take one of the two main forms. It can
be synchronous (time slotted) with the fixed packet length or
asynchronous (non-slotted) with variable packet lengths. In
synchronous operation mode, all arriving packets have a
fixed size and their arrival on each wavelength is
synchronized on a time-slot basis, where a time slot is the
time needed to transmit a single packet.
The operations of optical packet switching can be briefly
described as follows: When a packet arrives at the switch,
the packet header is extracted and processed electronically
by the control module. While the header is processed, the
packet payload is buffered in the optical domain using FDL
processing buffers. Based on the destination, information
extracted from the packet header and the control module
decides to which output fiber/wavelength the packet is
switched and configures the switch accordingly. Contention
occurs when two or more packets are assigned to the same
output port on the same wavelength at the same time [6 and
7]. The network has d service classes, ranging from service
class to service class d- 1 . We assumed that the output on a
single fiber/wavelength as the tagged fiber/wavelength.
Delay is calculated for class i traffic at the tagged output
fiber. Figl shows a switch for packet arrivals to a tagged
output fiber must originate from one of the FN input
wavelengths.
Input Wavelengths
2
Output
Wavelengths
i
i
i
i
or"'
H
c
N
A
g>
Fig 1. Model of the switch in OPS network under consideration
The following assumptions are made for the simulation:
Let ji denote the number of class i packets that arrive in a
time slot. The total number of packet arrivals at the fiber in a
time slot is k. That is j + j i +. . . +jd-i = k. In order to isolate
the service classes, the parameter li (0 < li < N) is
introduced, which is the number of wavelengths reserved for
Class i traffic in case of contention in a time slot. For a
service class i, if ji < li (the number of incoming packets are
less than the wavelengths assigned), it will result in ji - li
free slots. For a service class i, if ji > li (the number of
incoming packets are greater than the wavelengths
assigned), resulting in 1 ± - j ± overflow of packets.
The RB algorithm [5] is implemented by introducing
buffers. The RB algorithm is used for slotted OPS wherein
buffers are used to avoid overlapping of packets. This
algorithm is designed and used to improve the QoS of the
networks in terms of reduction in PLR. By using the same
algorithm delay rate is found and analyzed in slotted OPS
with fixed packet lengths for various traffic patterns. Hence
QoS is improved in terms of reduction in PLR and delay
[4,5]. In this paper the same algorithm is implemented in
asynchronous OPS for the packets of variable lengths.
III. Operating Principle
Operation of optical packet switch in a synchronous manner
with fixed length of packets is explained in [5]. Here we
assume that optical packet switch operates in an
asynchronous manner with variable length of packets.
Analysis of delay rate for variable packet sizes in
asynchronous OPS using three types of traffic patterns viz.
Non-uniform, Poisson and ON-OFF traffic models is
presented in this paper. In contrast to [8], immaterial of
packet size, allotted time slots remain the same and no
shifting of time slot allocation is done.
In non-uniform traffic pattern all nodes are not to receive
and send similar volumes of traffic [4, 9]. The number of
packets arrived and transmitted of packets is not equal.
There is an incoming packet for every slot. We assume that
239
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, 2010
each packet has equal probability of 1/N being addressed to
any given output port and successive packets that arrive at
the tagged output/wavelength are independent. Also, the
packet arrival time at the input and transmission time to the
output port are not equal. This is due to randomness of
packet arrival. This increases the occupancy of the free
slots. In Poisson traffic model, packet arrivals are random
and are mutually independent. The Poisson distribution can
be obtained for packet arrivals during an infinitesimal short
period of time X t, where X is called the arrival rate. Packet
arrival at the output fiber is a single Poisson process with X.
In Poisson model, the number of arrivals in non-overlapping
intervals is statistically independent. The arrival rate X is
expressed as the average number of arrivals during a unit of
time. The time distance between consecutive packet arrivals
is exponentially distributed [10]. However Poisson arrival
model is not an accurate model considering packet arrivals
at the tagged output fiber/wavelength in OPS. The Poisson
arrival model assumes an infinite number of sources, which
is not the case in a real switch. When using a Poisson arrival
model, there is the possibility of independency between the
service time and the packet inter-arrival time [11].
Bursty traffic is also known as ON-OFF traffic model. Both
the ON and OFF periods are distributed using exponential
distribution with the meanl/// andl//l respectively. It is
an alternating process where an OFF period follows an ON
period, and an ON period follows an OFF period. During
ON periods, series of packets are transmitted from the
source node and the time is called active periods. OFF
periods are called passive periods and no packets are
transmitted from the source. Active periods of a source are
exponentially distributed with one specific mean value, and
passive periods are exponentially distributed with another
mean value. During an active period, packets are generated
at regular periods. The most commonly used VOIP traffic
model is based on a two-state on-off model of a single voice
source. When a voice source is transmitted, it is in the ON
state, and when the voice source is silent, it is in the OFF
state; and the ON and OFF states appear alternatively. ON-
OFF sources, each of which exhibits a phenomenon called
the "Noah Effect" resulting in self similar aggregate traffic.
The Noah Effect for an individual ON-OFF source model
results in ON and OFF periods, i.e. "train lengths" and
"intertrain distances" that can be very large without
negligible probability [12 and 13]. The number of packet
arrivals divides the timeslot and produces the time slice
which is utilized for packet transmission. Hence the packet
arrival rate is random, time slice is determined according to
randomness of the packet arrival. ON and OFF periods are
used with their own mean values and using the same mean
values, various packets arrive and are transmitted during ON
period and idling occurs during the OFF period. During ON
period, packet is transmitted and there is no flow of packet
during OFF period. In this model, FN independent state
model generate packet arrivals at a tagged output fiber. In
this paper, at every node there should be arrival(s) of
different classes of packets. In ON-OFF model, ON periods
and OFF periods occur at each node for different classes.
When one class is in ON period, other classes may be in
OFF period at that particular node and OFF period of one
class is used by the other classes for transmitting packets.
Also, other possibilities are (1) all classes may be in OFF
period, (2) all classes may be in ON period and (3) some of
the classes may be in ON period and rest may be in OFF
period.
We assume that the switch is having buffers. In RB
algorithm, the packets use the wavelength according to their
service classes. If there is flow of packets into the ports
with a specified service class, the incoming packets with
assigned wavelengths occupy the ports as per the assigned
service class. When the assigned wavelengths in one service
class are occupied, it checks for the free wavelengths of
other service classes and the packets will occupy the free
slots in the other service classes. When the assigned
wavelengths are completely occupied in all the ports, the
packets overflow. By introducing buffers, the packets that
overflow are saved. When a free slot is not available for an
incoming packet, instead of dropping that packet, it will be
saved in the buffer which is provided in the switch. Before
entering the buffer, a bit is added in the packet header for
the purpose of reservation with respect to their service
classes. Whenever free slots are available, the packets in the
buffer occupy free slots. Buffered packets will have the
priority over the incoming packets. In the FF algorithm, all
wavelengths are numbered in a certain order, for example
ascending order from to W-l, where W is a number of
wavelengths. When the deciding port attempts to assign a
wavelength, it sequentially searches all wavelengths in an
ascending order and assigns the first available wavelength
[14]. In class based model, each node transmits the packets
according to their classes. Buffers are placed in the port for
every class. In port based model, wavelengths are placed in
a sequential order. Irrespective of the class, the available
wavelength is used by the incoming packets in a sequential
order and buffers are only placed in each port.
FF algorithm is implemented in port based and packets are
transmitted according to their wavelengths, whereas
transmission of packets is class based in RB algorithm and
packets are transmitted according to their service classes and
wavelengths. Thus the drop rates of packets are reduced in
optical networks resulting in improved QoS. The delay rate
is found and analyzed in asynchronous OPS for various
traffic patterns viz. Non-uniform, Poisson and ON-OFF
traffic models for various service classes with packets of
variable lengths. The delay rates for RB algorithm and FF
algorithm are found and compared.
For fixed size packets and variable size packets in OPS,
delay rate is encounted by implementing RB algorithm. The
packet includes payload and header. Fig 2 shows the packet
header. For fixed size packets, size considered is 512 bytes.
For variable packet size, the range of packet size is in the
240
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, 2010
range of 512 bytes and 2k bytes. 20 bytes is included as
header along with the payload invariably for any type of
packets. Variable size packets are used in VoIP applications.
In variable packet size, the size has been controlled by the
application. Packet sizes of the range of 1024 to 2048 bytes
show good efficiency in terms of bandwidth and reliability
in Digital Video Broadcasting [15]. Packet size is measured
using uniform min and max distribution. Packet size is
chosen depending on the application.
Source
IP
address
Desti
-nation
IP
address
Source
Port
Desti
-nation
Port
Packet
Sequence
no
Time
Stamp
Flow
id
4
bytes
4
bytes
2
bytes
2
bytes
2
bytes
4
bytes
2
bytes
Fig. 2 Structure of the packet header
When a packet is send from one node to other, the following
delays occur: (1) transmission delay (time required to send
all bits of packet into the wire), (2) propagation delay (the
time taken by the packet to travel through the wire), (3)
processing delay (the time taken to handle the packet in the
network system), and (4) queuing delay (the time taken is
buffering the packet before it can be sent). In most cases, the
delay (2) and (4) are considered in simulations and
measurements. The transmission delay (1) is usually small
for fast links and small packets and is therefore not
considered. Traditionally, the processing delay (3) has also
been negligible [16]. In our measurements, we consider the
delay (2) and (4). The mean delay for the above said traffic
pattern is found using equation (1).
T =T +T (U
2 D * Propagation ' * Queue v 1 /
Initially, we calculate the propagation delay that occurs
when a packet travels from the source to the destination.
Next, the queue delay experienced by a packet is calculated
using equation (2). This delay is due to waiting period of the
packet in the queue. A packet is in a queue, if a free
wavelength is not available at that particular time slot.
(2)
T =— Y(T)
A Queue AT /^ V i/
where Ti is the transmission time of class i packets at
particular time slice.
Summation of the waiting time in the buffer and the
transmission time between source node and destination node
through the switch is considered as delay and the same is
found for the above said traffic patterns.
IV. Results and Discussions
The delay values for the fixed length packets in slotted
OPS using RB algorithm is studied and is also compared to
the FF algorithm in our earlier paper [5]. In this paper, a
detailed analysis is carried out to find delay in asynchronous
OPS for variable length packet and is compared to FF
algorithm.
We consider 240 packets for the simulation purpose. Delay
for class i packets are calculated in the tagged fiber.
Wavelength assigned is 16 and total time slot chosen is 10,
hence this architecture can transmit 160 packets in a time
slot. Buffers are used along with RB technique, 240 packets
are chosen with an Erlang load of 1.5.
Delay in SC3(Class based)
Cl.iss
Fig 3. Delay rate for various traffic patterns for service class 3 using
reservation bit technique.
Delay in SC3(Port based)
rf
m
Fi
g 4. Delay rate for various traffic patterns for service class 3 using First-
Fit wavelength assignment algorithm
Delay in SC4(Class based)
Fig 5. Delay rate for various traffic patterns for service class 4 using
reservation bit technique
241
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, 2010
Deay in SC4(Port based)
Delay Rate for various Traffic pattern
Fig 6. Delay rate for various traffic patterns for service class 4 using
First-Fit wavelength assignment algorithm
Delay in SC5(Class based)
Fig 7. Delay rate for various traffic patterns for service class 5 using
reservation bit technique
Delay in SC5(Port based)
*
□ NU
■ O-F
DP
Fi
g 8. Delay rate for various traffic patterns for service class 5 using First-
Fit wavelength assignment algorithm
X10
D
e
m
SC5
SC4
Port Based
Class Based
Fig 9. Delay rate comparison
Transmitted Bytes
SC5
Fig 10. Comparison of data transmission
In service class 3, RB technique has 15 buffers, each service
class has 5 buffers, but in FF algorithm, each port has 5
buffers and the total number of buffers is 20. 10.96ms,
10.88ms and 9.83ms are the delay values while employing
FF algorithm and 7.55ms, 7.79ms and 7.91ms are the delay
values while employing RB technique in asynchronous OPS
using Non-Uniform, ON-OFF and Poisson traffic model
respectively for Service class 3 and the same is shown in
figs 3 and 4.
In Service class 4, both the techniques have 20 buffers.
10.94ms, 10.56ms and 10.18ms are the delay values while
employing FF algorithm and 7.77ms, 8.01ms and 7.561ms
are the delay values while employing RB technique in
asynchronous OPS using Non-Uniform, ON-OFF and
Poisson traffic model respectively for Service class 4 and
242
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, 2010
the same is shown in figs 5 and 6. From these it is seen that
both the techniques have same amount of buffer. FF
algorithm exhibits more delay, but delay is less in our
approach.
While in service class 5, RB techniques have 25 buffers and
FF algorithm need 20 buffers. 10.09ms, 10.88ms and
9.50ms are the delay values while employing FF algorithm
and 7.1ms, 6.27ms and 6.48ms are the delay values while
employing RB technique in asynchronous OPS using Non-
Uniform, ON-OFF and Poisson traffic model respectively
for Service class 5 and the same is shown in figs 7 and 8.
For all the service classes under consideration, buffers are
more or less the same for RB technique and FF algorithm,
whereas the delay rate is slightly higher in FF algorithm.
The above statement is true when the class based
transmission is compared with port based transmission.
Simulation results show that for all service classes under
any type of traffic pattern, class based model produces 29%
reduction of delay rate when compared to port based model
and the same is shown in fig 9. In RB technique the buffered
packets with respect to their classes will occupy the free
slots of the corresponding service class wavelengths on First
Come First Serve basis which reduces the waiting time in
the buffer whereas in FF algorithm wavelength utilization is
sequential, so buffered packets wait until they are serviced
with the next sequential order of wavelengths. Thus delay
rate is less in asynchronous OPS when RB technique is
employed. The delays are improved much in our algorithm.
This is due to the class based transmission.
Fig 10 shows the total number of transmitted bytes per slot.
It is shown that more number of packets is transmitted in
class based model for all service classes for all traffic
patterns under consideration when compared to port based
model. Hence class based model produces lesser delay as
well as more number of bytes transmission when RB
technique is employed. It is also shown that non-uniform
traffic pattern produces lesser delay when compared to ON-
OFF and Poisson model.
V. Conclusion
The delay rates for the Non-uniform, Poisson and ON-OFF
traffic models for various service classes are analyzed.
Minimum delay has been achieved in all service classes
under consideration by using reservation bit technique. It
provides lesser delay for all service classes for fixed and
variable length packets when compared to first-fit
wavelength assignment algorithm. It is seen from our
simulation, Poisson arrival model which is assumed in the
analysis approximates a more realistic model wherein all
input wavelengths are modelled as independent on/off
processes with exponential holding time. Also we have
presented a comparative study of reservation bit technique
and first-fit wavelength assignment algorithm for
synchronous and asynchronous OPS. It is concluded that in
Optical packet switching reservation bit technique reduces
delay; hence QoS is improved and at the same time Non-
uniform traffic pattern result in better Quality of service
compared to ON-OFF model and Poisson model. At the
same time ON-OFF traffic pattern have better QoS when
compared to Poisson process if OFF periods of one service
classes are more efficiently utilized by other service classes.
[I] Biao Chen and Jianping , " Hybrid Switching and P-Routing for
Optical Burst Switching Networks", IEEE Journal on Selected areas in
communications, vol 21, no 7,pp. 1071-1080, Sep 2003.
[2] A.Kavitha, V.Rajamani and P.AnandhaKumar, "Performance
analysis of Coding Techniques to find BER in Optical Transmission",
IEEE 1 st International Conference on Advanced Computing- ICAC 2009,
13-15 Dec. 2009 Page(s):21 - 27
[3] A.Kavitha, V.Rajamani and P.AnandhaKumar, "Evaluation of BER
in Optical Packet Switching using Various Coding Schemes" IEEE
Transaction on Optical Communication and Networking (under review).
[4] A.Kavitha, V.Rajamani, "Performance Analysis of Slotted Optical
Switching Scheme using Non-Uniform traffic" - Journal of Optical
Communications, Vol. 29, July 2008, pp. 107-1 11.
[5] A.Kavitha, V.Rajamani and P.AnandhaKumar, "Performance
Analysis of Slotted Optical Packet Switching Scheme in Non-Uniform
Traffic Pattern Using Reservation Bit Technique"- Selected for
publication- INFOCOMP, Journal of Computer Science.
[6] H. Overby and N. Stol, "Evaluating and comparing two different
service differentiation methods for OPS: the wavelength allocation
algorithm and the preemptive drop policy," in Proc. 3rd Int. Conf
Networking, vol. 1, Feb. 2004, pp. 8-15.
[7] S.Bjornstad, N.Stol and D.R.Hjelme, "A highly efficient optical
packet switching node design supporting guaranteed service" in proc. Of
SPIE, Vol 4910, 2002, pp.63-74.
[8] BO Wen, Ramakrishna Shenai, and Krishna Sivalingam," Routing
Wavelength and Time-Slot-Assignment: Algorithms for Wavelength-
Routed Optical WDM/TDM Networks", Journal of Lightwave
Technology, vol 23, no 9, pp.2598-2609,Sep 2005.
[9] Adisak Mekkittikul, Nick McKeown "Scheduling VOQ Switches
under Non-Uniform Traffic", CSL Technical Report, CSL-TR 97-747,
Stanford University, 1997 Stanford University,Standford.
[10] Wuyi Yue, Yutaka takahashi,hideaki tagaki, "Advances in
Queueing theory and network application:, Springer,ISBN: 978-0-387-
09702-2, e-ISBN: 978-0-387-09703-9.
[II] Harald Overby and N.Stol " Quality of Service in synchronous
bufferless optical packet switched networks" Kluwer Telecomm. Sys.
Vol 27, 2004. pp. 151- 179.
[12]Mihails Kulikos and Ernests Petersons"Remarks Regarding
Queueing Model and Packet Loss Probability for the Traffic with self -
Similar Characteristics", Networks", International Journal of Computer
Science 3;2,Spring 2008,pp. 85-90.
[13] Eric W.M. Wong and Moshe Zukerman," Bandwidth and Buffer
Tradeoffs in Optical Packet Switching", Journal of Lightwave
Technology, vol 24, no 12, Dec 2006, pp 4790- 4798.
[14] Xuehong Sun, Yuunhao Li, Ioannis Lambadaris and Yiqiang
Q.Zhao, Performmance Analysis of First - Fit Wavelength Assignment
Algorithm in Optical Networks", IEEE 7th International Conference on
Telecommunications - ConTEL2003, Junell-13, 2003, pp. 403-409.
[15]Vadakital, V.K.M. Hannuksela, M.M. Razaei, M. Gabbouj,
M," Optimal IP Packet Size for Efficient Data Transmission in DVB-H",
Proceedings of the 7th Nordic ,IEEE Signal Processing Symposium, pp.
82-85, 7-9 June 2006.
[16] Ramaswamy Ramaswamy, Ning Weng and Tilman Wolf,
"Characterizing Network Processing Delay", Proc. in Globecom 2004,
IEEE Communication Society, pp.1629- 1634.
AUTHORS PROFILE
243
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, 2010
A.Kavitha received the B.E
degree in Electronics and
Communication Engineering
from Madurai Kamaraj
University, India in 1997. The
Master of Engineering degree
in Computer and
Communication Engineering
from Anna University, Chennai
in 2004 and currently pursuing
Ph.D degree in Optical
Communication Networks in
Anna University, Chennai. At
present she is working as a Senior Lecturer in Chettinad College
of Engineering and Technology, Karur. She has published more
than 10 papers in referred National and International
conferences/Journals. Her area of interest includes Networks,
Computer Architecture, Digital Communication and etc.
V.Rajamani received the B.E
degree in Electronics and
Communication Engineering from
National Engineering College, Anna
University, India in 1990. The
Master of Engineering in Applied
Electronics from Government
College of Technology, Bharathiyar
University Coimbatore in the year
1995 and the Ph.D. degree in
Electronics Engineering from Institute of Technology, Banaras
Hindu University, Varanasi in 1999. He started his carrier as
Lecturer in Mohamed Sathak Engineering College from 1991
onwards. He has held various positions in various Engineering
Colleges. At present he is working as a Principal in Indra
Ganesan College of Engineering, Tiruchirappalli. He has
completed a project under AICTE - RPS scheme successfully. He
has published more than 80 papers in referred National and
International Journals/ conferences. His area of interest includes
Device modeling, VLSI, Image processing, Optical
Communication and system.
P.Anandhakumar received the B.E degree in Electronics and
Communication Engineering from University of Madras, India in
1994. The Master of Engineering degree in Computer Science
and Engineering from Bharathiyar University in 1997 and Ph.D
degree in Computer Science and Engg. from Anna University,
Chennai in the year 2006.At present he is working as a Assistant
Professor in Madras Institute of Technology, Chennai. He
published more than 50 papers in referred National and
International conferences/Journals. His area of interest includes
Digital Communication, Soft Computing, Robotics and etc.
244 http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
A Feedback Design for Rotation Invariant Feature
Extraction in Implementation with Iris Credentials
M. Sankari
Department of Computer Applications,
Nehru Institute of Engineering and Technology,
Coimbatore, INDIA.
R. Bremananth
School of EEE, Information Engg. (Div.),
Nanyang Technological University,
Singapore.
Abstract — Rotation invariant feature extraction is an essential
objective task in computer vision and pattern credentials
problems, that is, recognizing an object must be invariant in
scale, translation and orientation of its patterns. In the iris
recognition, the system should represent the iris patterns, which
is invariant to the size of the iris in the image. This depends upon
the distance from the sensors to subjects' eye positions and the
external illumination of the environments, which in turn make
the changes in the pupil diameter. Another invariant factor is the
translation, the explicit iris features should be a positional
independent even though eye present anywhere in the acquired
image. These two invariants are perfectly achieved by the weight
based localization approaches. However, the iris orientation
estimation is an important problem to avoid in preserving
selective orientation parameters. Multiple source points are used
to estimate the segmented objects orientations. After estimating
the deviation in angle of segmented object that can be rotated to
its principal origin and then the feature extraction process is
applied. A multi resolution approach such as wavelet transform
is employed for feature extraction process that provides efficient
frequency and spatial texture feature deviations present in the
irises. In this paper, we work on a feedback design with Radon
transform with wavelet statistical analysis of iris recognition in
two different ways. In order to check the viability of the proposed
approaches invariant features are directly compared with
weighted distance (WD) measures, in the first phase and second
phase is to train the Hamming neural network to recognize the
known patterns.
Keywords- Iris credentials; Invariant Features; Rotation
estimation; Multiresolution anlysis;
I.
Introduction
In computer vision and pattern recognition, rotation
invariant feature extraction is an essential task, that is,
recognizing an object must be invariant in scale, translation and
orientation of its patterns. This paper emphasizes on invariant
feature extraction and statistical analyses. In the iris
recognition, the system should represent the iris patterns, which
is invariant to the size of the iris in the image. This depends
upon the distance from the sensors to subjects' eye positions
and the external illumination of the environments that make the
changes in the pupil diameter. Another invariant factor is the
translation where iris features should be a positional
independent of iris pattern, it could occur anywhere in the
acquired eye image. However, the iris orientation estimation is
an important problem to avoid in preserving selective
orientation parameters, for example, 7 relative orientations
were maintained for iris best matching process in the literature
[1] and seven rotation angles (-9, -6, -3, 0, 3, 6 and 9 degrees)
used by Li ma et al. [2]. In the real time imaging, due to the
head tilt, mirror angle and sensor positions, iris images are
captured in widely varied angels or divergent positions. We
estimate the rotation angle of iris portion within the acquired
image by using multiple line integral approaches, which
provide better accuracy in the real time capturing. Local binary
patterns, gray-level and auto-correlation features were used to
estimate orientation of the texture patterns. It projected the
angles that are locally invariant to rotation [3]. In [4], texture
rotation-invariant was achieved by autoregressive models. It
used several circle's neighborhood points to project the rotation
angle of the object. Aditya Vailaya et al. [5] had dealt with
Bayesian learning framework with small code features that are
extracted from linear vector quantization. Thus, these features
can be used for automatic image rotation detection. A hidden
Markov model and multichannel sub-band were used for
estimating rotation angles of gray level images in the study [6].
In this work, we propose Radon transform based multipoint
sources to estimate the rotation angle estimation for real-time
objects.
Classification is a final stage of pattern recognition system
where each unknown pattern is classified to a particular
category. In iris recognition system, a person is automatically
recognized based on his / her iris pattern already trained by the
system. This is done in a way of training a brain to teach
certain kind of sample patterns. In the testing process, system
recalls the trained iris patterns as a weighted distance specified
by the system. If threshold is attained then system genuinely
accepts a person, otherwise false alarm sounds. However, the
way to find the statistical level is a tedious work because it
makes decision to evaluate the pattern either genuine or fake.
Hence combinatorics of iris code sequence should be carried
out by means of statistical independence. Moreover, failure of
iris recognition is principally concerned with a test of statistical
independence because it absorbs more degree-of-freedom. The
test is nearly assured to be allowed whenever the extracted iris
code comparing from two different eyes are evaluated. In
addition, the test may exclusively fail when any iris code is
compared with another version of itself. The test of statistical
independence was implemented by the Hamming distance in
245
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
[1] with a set of mask bits to prevent non-iris artifacts. Li ma et
al. [7] proposed a classifier design that was based on exclusive-
OR operation to compute the match between pairs of iris bits.
In [2], authors worked with the nearest centre classifier to
recognize diverse pair of iris patterns. A competitive neural
network with linear vector quantization was reported for both
identification and recognition of iris patterns by Shinyoung
Lim et al. [8]. Our main contribution to this paper is a feedback
design (Fig. 1) to extract an appropriate set of rotation invariant
features based on Radon and wavelet transforms. An iteration
process is repeated until a set of essential invariant features is
extracted from the subject. We have done two different phases
of statistical analyses of rotation invariant iris recognition.
During phase I, wavelet features are directly compared with
weighted distance (WD) measures and in phase II invariant
features were trained and recognized by the Hamming neural
network.
Rotation
estimation
using multiple
sources
Rotation
Correction to
its principal
direction
Wavelet based
Rotation
Invariant
extraction
i
i
No,
Rotat
Fine
ion
i another suitable
invariant Feature
1
s
^ — ^Ts it prov
r
ide best^^
>
identification?,^ — ^
Yes ^^>£
Enroll the
Invariant
Features
Fig. 1. A feedback design of rotation invariant feature extraction.
The remainder of this paper is organized as follows: Section
II emphasizes on invariance and estimation of rotation angle.
Radon and wavelet based rotation invariant is described in
section III. Section IV depicts the results obtained based on the
proposed methodologies while Concluding remarks and future
research direction are accentuated Section V.
II. Invariance in Rotation
A 2D rotation is applied to an object by repositioning it along a
circular path. A rotation angle 6 and pivot point about which
the object to be rotated is specified for generating series of
rotation. In counterclockwise, positive angle values are used for
rotation about the pivot point and in contrast clockwise rotation
requires negative angle values. The rotation transformation is
also described as a rotation about an axis that is perpendicular
to the xy plane and passes through the pivot point. The rotation
transformation equations are determined from position
( Xi , y± ) to position ( x 2 , y 2 ) through an angle B relative to the
coordinate origin. The original displacement of the point from
the x-axis is, angle A. By trigonometric ratios, sin(A) = Vj Ir ,
sin(A + Z?) = y 2 lr , cos(A + B) = x 2 /r and
cos(A) = jci / r • From the compound angle formulae described
as
sin(A + B) = sin(A) • cos(B) + cos(A) • sin(B) . (1)
Substituting trigonometric ratios and obtain the following
(y 2 /r) = (y 1 /r)*cos(B) + (x 1 /r)*sin(B), (2)
y 2 = y\ cos(B) + x\ sin(5) , (3)
y 2 = x\ sin(Z?) + yy cos(B) , (4)
Likewise, substituting trigonometric ratios and derived as
cos(A + B) = cos(A) • cos(B) - sin(A) • sin(2?) , (5)
(jc 2 / r) = (jq / r) • cos(£) - ( y l I r) • sin(£) , (6)
x 2 = xi cos(B) - yi sin(i?) , (7)
Therefore, from Eqs. (5) and (10) we can get counterclockwise
rotation matrix and the new coordinate position can be found
as described in Eq. (8). The basics of rotation and line
integrals are incorporated together to form equations for
projecting the object in single and multi source points.
A. Multipoint source
Based on the basics of rotation, multipoint source method
computes the line integrals along parallel beams in a specific
direction. A projection of image f(x,y) is a set of line integrals
to represent an image. This phase takes multiple parallel-
beams from different angles by rotating the source around the
centre of the image.
y 2
cos(B) -sin(5)
sin(£) cos(B)
yx
(8)
This method is based on Radon transform, which estimates
the angle of rotation using the projection data in different
orientations. A fusion of Radon transform and Fourier
transform had been performed for digital watermarking which
is invariant to the rotation, scale and translation invariant in the
literature [9]. A parallel algorithm for Fast Radon transform
and its inverse was proposed by Mitra et al. [10]. Radon
transform was employed for estimating angle of rotated texture
by Kourosh et al. [11]. Image object recognition based Radon
transform was proposed by Jun Zhang et al. [12], this method is
robust and invariant to rotation, scale and translation of image
object. Fig. 2 shows a multipoint source at a specified angle for
estimating rotation angle of a part of iris. This method projects
the image intensity with a radial line orientation at a specific
angle from the multipoint sources. Multipoint projection
computes any angle 9 by using the Radon transform R(x\6) of
f(x,y), it is the line integral of parallel paths to the y axis. After
applying the function of multipoint sources R( x \@) the
resultant data contain row and column. Column describes
projection data for each angle in 9 and it contains the respective
coordinates along the x' axis. The procedure for applying
multipoint source projection to estimate the angle is as follows:
Image is rotated to a specific angle in counterclockwise by bi-
cubic interpolation method.
246
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
Rotation in counter
clockwise
Fig. 2. Multipoint estimation using multipoint sources.
Assume the rotation angle from 1° to 180° in order to find the
peak area of rotation angles. After applying the multipoint
sources, Radon transform coefficients have been generated for
each angle. The standard deviation of the Radon transform
coefficients is calculated to find the maximum deviation of
rotation angle. This is shown in Fig. 3. Then, using estimated
angle, object rotation is rotated to its original principal angle
using bi-cubic interpolation method. If the estimated angle 6 is
positive then rotate the object as -( + 90° ) in clockwise
direction else if the estimated angle is negative or above
90° then rotate the object as -(#-90°) in clockwise direction.
Multi point source Radon transform R(X',e)
10 20 30 40 50 60 70 fiO 90 100 110 120 130 140 150 160 170 160
Rotation e (degrees)
Fig. 3. Illustration of orientation angle estimation using multipoint.
III. Iris wavelt Feature analysis
In this phase wavelet based feature extraction process has
been employed to extract feature obscured in the iris patterns. It
is an essential task for recognising a pattern from others
because some features may produce same type of responses for
diverse patterns. It causes the hypothesis in pattern recognition
process to differentiate one from another. To overcome the
problem of uncertainty the system needs an efficient way to
extort quality features from the acquired pattern. Iris provides
sufficient amount of interclass variability and minimises intra-
class variability. Thus the characteristics of these patterns are
well efficiently taken out by the sense of using less
computational process. Among various feature extractors,
wavelet series approximate hasty transitions much more
accurately than Fourier series. Consequently, wavelet analysis
perfectly replicates constant measurements. It produces better
approximation for data that exhibit local variation and because
of its basis function each term in a wavelet series has a compact
support within a finite interval. The other sense to employ
wavelet is orthogonal. This means that information carried by
one idiom is independent of information conceded by the other.
Thus, there is no redundancy in the feature extraction. This is
fine when neither computational sequence time nor storage is
wasted as a result of wavelet coefficient computed or stored.
The next sense related with wavelet is multi resolution, which
is like biological sensory system. Many physical systems are
organised into divergent levels or scales of some variables. It
provides an economic structure and positional notion of
arithmetic whose computational complexity is O (N), where N
data points are to be accessed [13]. In the current literature
various computer vision and signal processing applications
have been based on wavelet theory [14] such as detecting self-
similarity, de-noising, compression, analysis and recognition.
This technique has proven the ability to provide high coding
efficiency, spatial and quality features. However, wavelets
features are not rotation invariant due to its directional changes.
Hence this approach initially estimates the extorted pattern
rotation angle and rotates to its principal direction. Afterwards
multi resolution wavelets have been employed to extort
features from the rotation corrected patterns. In the iris
recognition process, this approach has adopted Daubechies (db)
wavelet to decompose the iris patterns into multiple resolution
sub-bands. These sub-bands are employed to transform well-
distributed complex iris patterns into a set of one-dimensional
iris feature code. Decay is a process to divide the given iris
image into four sub-bands such as approximation, horizontal,
vertical, and diagonal coefficients. A 2D Daubechies wavelet
transform of an iris image (I) can be carried by performing two
steps, Initially, it performs ID wavelet transform, on each row
of (I) thereby producing a new image Ii. In second step it takes
Ii as an input image and performs ID transform on each of its
columns. A Level- 1 wavelet transform of an image can be
described as
a 1 ti
v 1 d 1
,a
a 2 h 2
2 j2
v a
(9)
where the sub-images a 1 , h 1 v 1 and d 1 represent level- 1
approximation, horizontal, vertical and diagonal coefficients
a 2 , h 2 v 2 and d 2 level 2 coefficients. The approximation is
created by computing trends along rows of I followed by
computing trends along columns. Trends represent the running
average of the sub-signals in the given image. It produces a
lower frequency of the image I. The other sub-signals such as
horizontal, vertical and diagonal have been created by taking
fluctuation. It is a running difference of sub-signals. Each
coefficient represents a spatial area corresponding to one-
quarter of the segmented iris image size. The low and high
frequencies represent a bandwidth corresponding to
247
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
0<\cD\<nl2 and tt 1 2 <\co\ < n > respectively. Fig. 4 shows
frequency variation of Daubechies wavelets. The wavelet
transform is defined as
W(a,r x ,r v )= j J I{x,y) ¥a ^ t
>x*>y>
(x, y)dxdy
x v y
^a,T x T y
1 X - T
\a\ a
(10)
(ii)
where I(x,y) is a segmented iris image, W(a,x x ,T y ) is a
wavelet transform function,
^a,T x T y
(x,y)
the wavelet basis
function, a is a scaling factor, T x and T are translation factors
of x and y axes, respectively. The properties separability,
scalability, translatability of discrete wavelet transform is
performed as
<p(x, y) = <p(x)<p(y),i// (x, y) = \ff(x)<p(y) ,
y/ v =(p(x)i//(y),i// D =if/(x)ij/(y),
j M N
i M N
W H ¥ (l,m,n) = -j= X £/(*, y)w" m Jx, y) '
J M N
W v ¥ (l,m,n) = -j== X £/(x, y)y/l m>n {x, y) '
yIMN x=ly=l
W D v/ (l,m,n) =
J M N
^g, /(M) ^ (lj) '
(12)
(13)
(14)
(15)
(16)
(17)
where (p(x, y) and W(p{l^ , in, ri) are scaling function and
approximation coefficients of I(x,y) at scale L , respectively.
W H „ x W V si x W D ,i x are coefficients of horizontal,
YY y/(l,m,n) rr i//(l,m,n) rr i//(l,m,n)
vertical and diagonal details for scales / > / respectively.
Normally / = o, and assigning M = N = 2 so that
/ = 0,1,2..., L - land m = n = 0,1,2,...,2 J " - 1 .
The decomposition of signals produces sub-signals such as
low, middle and high frequency of the components, which
play a very important role in the feature extraction process. In
this approach Daubechies wavelet is employed for feature
extraction process. Its frequency distribution for different level
is illustrated in Fig. 5. The DWT (Discrete Wavelet
Transform) consists of log 2N stages if the given signal s is of
length N. Fig. 6 shows the scaling and wavelet functions of
Daubechies wavelets. Initially s produces two sets of
coefficients such as approximation coefficients cA b and detail
coefficients cDi. These coefficients are obtained by
convolving s with the low -pass filter Lo_D for approximation,
and with the high-pass filter Hi_D for detail coefficients. In
the case of images, a similar procedure is possible for 2D
wavelets and scaling functions obtained from one-dimensional
wavelets by tensorial products. This kind of 2D DWT leads to
a decomposition of approximation coefficients at level j in
four components: the approximation at level j + 1 and the
details in three orientations (horizontal, vertical, and diagonal).
Fig. 7 shows the decomposition process.
Daubechies(dbl> wavelet (rends and fluctuations
t
5
te^UJ4* ss *tt
Decomposition high-pass tiler
Reconstruction high-pass filter
Fig. 4. Daubechies (dbl) wavelets frequency variations.
| 0.5
i c
Fig. 5. Frequency distribution of Daubechies wavelets by different iterations.
In the feature extraction process of iris patterns four levels of
decompositions have been performed to obtain fine level of
frequency details from the pattern. The scaling factor is very
important for decomposing the given iris signals. At the first
level it produces 648 signals, second level has 162 signals,
third level 45 signals and finally it generates 15 signals for
each frequency. The MRA produces the frequency signals to
compact approximation of features which aid to generate an
efficient set of distinct features that are provided with less
intra class variability and more interclass variability in the iris
pattern recognition process. Fig. 8 shows four levels of
decomposition process for the given iris images. Low-pass
248
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
filter corresponds to an averaging operation and extract the
coarse information of the signals, whereas high-pass filter
corresponds to a dissimilarity operation that extracts the
detailed information of the signals.
Wavelet fund J on (1)
Si
Wavelet function (2)
Wavelet function (3)
Fig. 6. Scaling and wavelet functions of Daubechies wavelets.
When iris signal passes through these low and high pass
filters, it generates the frequency variation occurring in a
pattern.
Approximation
Lo_D
Rows
Hi_D
Rows Down
sample
- 2 + 1
Columns Down
sample
Lo_D -*- 1+2 -*- ^i+i
Hi_D -+ 1+2 -^ cD h
-* 2+ 1 -
Lo D -^ 1+2 -** cD*
Hi D -*~ 1+2 -+ CD*
Fig. 7. Decomposition of wavelet signals in the feature extraction.
A. Iris feature selection
In this phase frequency variation of iris signals in divergent
levels are quantized into iris features. For that multi resolution
frequencies of low and high pass filters are taken for
quantization process of conversion of real signals into binary.
The mean and standard deviation of approximation and detail
coefficients vary in each level of the decomposition of iris
patterns which raises up to generate an efficient feature sets of
the given patterns. The horizontal, vertical and diagonal
coefficients wavelet features have middle and high frequencies
of the components of iris signals. The histogram analyses of
signals in divergent levels are illustrated in Fig. 9. The
frequency distribution of signals at level 1 ranges from -10 to
10 and from -100 to 100 at level 4 for horizontal coefficients.
Thus this approach quantizes these trends and fluctuation of
sub-signals into iris features. After performing the four level of
decay process the horizontal, vertical and diagonal coefficients,
the iris are used for iris feature encoding process. The
frequency variation occurring in these decomposition
coefficients are employed to extract iris feature codes. In order
to make an efficient set of features and reduce the
computational time of iris matching process the coefficient
values are converted into binary values which senses to create a
compact feature set.
Fig. 8. Four level of decomposition of iris patterns.
s image (72 x i6) analysed at level I witit dl
¥«r«itnl Pi;H*a
Fig. 9. Histogram of divergent levels of iris image.
In the current literature, Haar wavelets are used for iris image
feature extraction by decomposing the signals into four levels
[8]. It uses only high frequency of the components for
representing iris patterns. However, iris patterns are having
middle frequency of the components, which are essential for
recognizing iris patterns in large population. Moreover, in
their approach there is no transformation-invariant analysis.
When there is a rotation between a set of irises from the same
subject, it may produce false positives in the recognition
249
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
because these patterns produce different kinds of features for
diverse rotation and translations. Here, transformation
invariant analysis is performed before extracting features from
the iris patterns. In addition, middle and high frequency of iris
wavelet features were extracted for recognition. Thus, it
reduces less false positives in the recognition process using a
feedback design based on the rotation invariant features.
Though iris patterns are unique in nature, it is a difficult
process to generate identical template for the same subject.
This is mainly due to the changes in imaging position,
distance, illumination conditions, eyelashes / eyelid occlusion
and eyewear reflections. These factors may affect the efficacy
of the system. Thus this approach compensates the
deformation of these factors and recognizes the iris patterns,
which are independent of transformation factors and other
artifacts. The classification results of rotation invariant and
wavelet features are illustrated in Section V.
VI. Rotation Invariant classification
The different pairs of eye images were captured in diverse
distances and illuminations provide more challenges to this
approach. Experimentations were also performed with
different eye images in diverse criteria like normal, outdoors,
contact lens, spectacles, and diseased (Tumours, Tear,
Iridocyclities) eyes. The database of iris images has 2500
images captured from 500 different subjects as each has been
acquired as 5 different images with different real-time
conditions [18, 19]. In the iris matching process, inter and
intra class iris features are efficiently separated and they
prevent impostors from entering into the secure system. To
authenticate any genuine user, iris feature sets are treated as
trained sets and stored in the encrypted file. Verification
subjects' irises are represented as test sets. The same subject
iris feature codes could vary due to external noises, lighting,
illuminations and other factors such as closed eyelashes or
eyelids. This possibly will lead to different iris template for an
eye, even though iris is unique in nature. However, capturing
eye images with advanced biometric camera may solve this
problem. The process by which a user's biometric data is
initially acquired, validated, processed and stored in the form
of a template for ongoing use in a biometric system is called
enrolment. Quality enrolment is a critical factor in the long-
term accuracy of biometric system. Wavelet features of irises
are recognized using the weighted distance (WD). It
recognizes the various classes of iris codes by checking a
minimum distance between two iris codes. This is defined as,
imn4pOFC(x tmined ),iFC(r, .w , where
t))
WD(IFC(x ),IFC(x ))
i J
represents weighted distance in between two iris feature sets
as defined as
WD(iFC(x trained )jFC( Xtest )).
\ IFC ^trained^ IFC ^test)\ ? (1 8 )
TV
where N denotes the number of bits in the iris feature set. The
weighted distance (WD) is used to determine the number of
error bits in between two iris classes. In the experimentation,
the weighted distance of the intra class feature set is
discriminated by the constraint < WD < 0.2 and inter class
iris features is abandoned with the constraint WD > 0.2 . These
distances are also evaluated based on the normal distribution
of mean, standard deviation and degree-of-freedom of the
wavelet iris codes. In addition the same candidate's iris image
may have more artifacts due to various deteriorates as stated
previously. Hence WD needs more discriminability range for
recognising the genuine subject. Conversely, if system
maintained large distance variation to allow the subjects, then
more FAR (False accept rate) might be encountered.
Moreover, if WD is reduced then more FRR (False rejection
rate) may be produced by the system. The system was tested
with normal and abnormal images and their mean of weighted
distance of genuine-class iris codes was u=0.10813, its
standard deviation was a= 0.0392 and degree-of-freedom V
was 62.621991. Impostor-class mean value was u=0.27104
and its standard deviation was g=0. 040730. During the
weighted distance computation, an identical iris pattern was
produced WD = and due to abnormal conditions the same
subject iris was assorted from to 0.19 WD. This is shown in
Fig. 10. If distributions are very large then system allows more
changes for impostors to access the system. This type of
limitation of distributions may be provided with more false
reject rate, but minimum false accept rates. In most of the
applications such as Bank-ATM and biometric voting
machines these type of constrained weighted distance are
essentially desirable in order to agree entire genuine subjects.
In the recognition phase, GAR (Genuine accept rate) was
99.3% and FAR was 0.7% and in confirmation MR (Matching
rate) was 99.94% and FRR was 0.06%.
Normal Distribution Analysis of Ins
Frequency Polygon of Ins codes
ins code Weighted distance
code Weighted distance
Fig. 10. Weighted distance distribution for wavelet iris features and frequency
polygon of the iris codes.
A. Hamming neural network (HNN)
Hamming neural network (HNN) is an alternative way to train
and test the extracted features [15]. This network is employed
to train for both iris and character patterns. Its input layer can
accept wavelet features. That is, it works with bipolar value of
the extracted iris wavelet features. Wavelet based iris feature
codes are fed for recognition process. HNN is used to
recognize iris features from the trained set. The aim of the
250
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
HNN is to decide which trained iris feature set is closest to the
test feature set [16]. HNN consists of two layers; the first layer
is called as a feed forward layer (FFL) that is used to calculate
a maximum score of the input patterns and a recurrent layer
(RL) is used to select the maximum score among the input
patterns. Each neuron in the FFL is set up to give maximum
response to one of the trained patterns. If test set is same as
trained set the maximum score is taken by the recurrent
network. The weights initialization process of the HNN is
described as
W ;;
2 '
Wi0 = n/2 ,
(19)
where Wy and x y are the weights value and input features of f 1
bit of the i th iris feature, Wi is a bias value and n is the number
of bits in the iris features. In order to incorporate HNN with
iris recognition, wavelet features are converted from binary to
its corresponding bipolar form. For example, the i th iris feature
set is {-1, +1, +1, -1, +1...+1}. The weight of i th neuron is set
to {Wio= 67.5, Wii= -0.5, w i2 = 0.5, w i3 = 0.5, w i4 = -0.5, w i5 =
0.5, ... , Wii35= 0.5}. The weighted sum is 135. Each of the
neurons in the FFL gives a maximum response of 135 for
exactly identical iris codes, and a minimum value to other
features. In HNN, the number of neurons in the FFL is same as
the number of neurons in the recurrent layer. When a test
feature is given to the FFL, the output from each of the
neurons in the FFL is measured by the Hamming distance
from the iris in the training set. The Hamming distance
between two iris patterns is a measure of the number of bits
that are different between the two iris patterns. For example, if
an input iris pattern of {+1, +1, +1, -1, -1...+1} is fed then the
output of FFL is 133 which has 2 less than the maximum of
135. This is because the given pattern has 2-bit difference.
Perhaps, if entire bits are changed in the iris patterns, the
neuron that corresponds to that pattern produces an output of
0. The function of RL is to select the neuron with the
maximum output. The final output of the RL contains a large
positive value at the neuron that corresponds to the nearest iris
pattern, and all other neurons produce value. The RL is
trained by setting the weight to 1, if the weight connection
corresponds to the same neuron and all other weight value are
small negative value less than -1/TI. The response of the RL is
described as
Y =
^ w { . x t if 2* w / • x i — A*
otherwise
(20)
where TI is the total iris patterns available in the trained set, X
is a threshold maintained in the iris recognition process. In this
process, the output is fixed to the value of the output of the
FFL. The RL is allowed to iterate, initially the outputs of the
RL is equal to the score produced by the FFL. Then, because
of the less than -1/TI weights, the output is gradually reduced.
After some iteration, all the outputs reach except the
recognized pattern with threshold, for example,^ =81, i.e.
weighted distance for HNN, WD h is 0.6. The testing process
of HNN is stopped if there is no change expected in the
iterations.
V. Experiments Invariant classification
Experiments were carried out on cases like left and right eye
evaluation, twins eye evaluation, eyewear and artifact
evaluation, hypothesis test, segmented iris, normalized iris,
Receiver operating characteristics curve (ROC) evaluations
and feature vector dimension variations. To evaluate these
phases, system was tested based on GAR, FAR and FRR
factors of the recognition.
A. Fusion of left and right eyes
Evaluating both the left and right eye combinations provide
better security in the application domains. However, the
recognition time is directly prepositional with the number of
entries in the iris database. A pair of 120 subjects' eye images
was acquired to test the algorithm, that is, a total of 240 iris
patterns were trained and tested by the ENDM, weighted
distance and HNN. The feature vector size is double the
dimension of normal vector. Thus, 270 wavelet iris features
were computed for each subject to test the weighted distance
and HNN. Table I depicts the recognition rates for evaluating
both left and right eye images. In the recognition process, a
system was set by a matching threshold level. It determines
the error tolerance of the system with respect to the features
for which the network is trained, and is used to determine
whether the final result is accepted / rejected. For any
recognition system that is used for security applications, this
error tolerance should be minimal and therefore the setting of
this matching threshold is a crucial factor. Recognition rates
were reported based HNN and WD. WD was better
recognition rates with minimal FAR. Furthermore, its FRR
was also an acceptable one, hence the system with wavelet and
WD produce good performance than the HNN.
TABLE I.
Comparison of classifiers accuracy rate
Feature
Types of
classifier
Left and
right iris
features
Matching
Threshold
Range
Recognition rate
GAR
%
FAR
%
FRR
%
Wavelet
WD
270
[0.0-0.19)
99.4
0.6
1.2
Wavelet
HNN
270
[0.7-1.0)
99.32
0.68
2.7
B. Recognition of twins
Identical twins' irises were separately verified with different
methods. From 50 twins, 500 eye images were acquired. It
contained both left, right eye images with each subject having
10 eye images. The twins' iris code result was generated by
the classifiers as the same weighted distances as the regular
iris codes available in the database. The mean of WD in the
images acquired from twins is 0.086360 with the standard
deviation 0.044329. A confidence interval is a range of values
that have a chosen probability of containing the true
251
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
hypothesized quantity. The standard deviation of confidence
intervals is in the range 0.0417 to 0.0473. Fig. 11 shows the
normal distribution of twin's iris code weighted distance. In
the checking of twin's iris database WD was changed in the
range to 0.19, GAR was 99.3 and FAR was 0.7.
D. Iris hypothesis test
Hypothesis test plays very important role in the biometric
recognition system, i.e., making a decision is based on the
availability of data during the training or enrolment and testing
or verification processes.
C. Eyewear and Artefacts
Eyewear images are major problematic ones in the iris
recognition because these images may produce more false
localization and FAR or FRR of the system. To evaluate the
recognition rates of eyewear images, 50 subject's eye images
were acquired with white glasses from each of them 5 images
were captured, that is total of 250 images were utilized for the
recognition. As stated previously, an exact identical iris
pattern could be produced WD=0, but due to eyewear noises
and other artifacts its patterns require a certain WD range.
Hence system evaluated the same subjects' iris patterns before
and after wearing the eyewear. This also included soft contact
lens and white glass with different varieties. Table II shows
the WD on the image with and without wearing eyewear. In
that hard contact lens produced more FRR in the recognition.
Thus it was around 32 bits average error bits and WD was
0.237. Moreover, localization system may be disrupted by the
designed frames of eyewear in the hypothesis to locate the
ROI. However, in the recognition, it produced minimal
average error of 22-bits.
Identical twins Normal Distribution Analysis of Iris codes
TABLE II.
Eyewear noises and other artifacts assessment
g6
I
>
1 ** 1
i i
1
1 '
*
-
*
A A A
*
A
A
#
~
£
A
"
■
£.
A
-
+
A
A
*
A
A
Inti a-clasi
-
Inter-class
A
r *
A
■
ji=0.269047
A
]A=O,0S6360
-
o=0.050261
A
f a=0.044329
■
•
A
A
A
0.05 0.1 0.15 0.2 0.25 0.3 0.35 0.4 0.45
Iris code Weighted distance
Fig. 11. Representation of weighted distance of twins' iris code.
The FRR and FAR was high when images were acquired with
eyewear and in diverse illuminations such as sunshine and
twilight conditions, eyewear at twilight the average of 34 bits
were corrupted, therefore WD was 0.251. As a consequence,
the system recommends the application domain while
enrolling a void eyewear because during enrolment iris
patterns could be signed up with minimum amount of error
bits. Therefore, it increases system recognition rates in order
to achieve better rotation invariant feature set.
Types of eye wears
Without eyewear
With eyewear
Average
Error bits
out of 135
Average
error in
WD
Average
Error bits
out of 135
Average
error in
WD
White glass
14
0.104
22
0.165
Soft contact lens
15
0.112
26
0.194
Hard contact lens
15
0.112
32
0.237
Sunshine
18
0.129
27
0.198
Twilight
19
0.138
34
0.251
This test may be neither true nor false. It could be dependent
on the feature extraction and classifier design of the system.
Thus this system makes iris images as transformation invariant
patterns to increase the performance of the system. The
biometric estimation is based on some terms of assumptions
that is, make a system as null hypothesis. The null hypothesis
is the original declaration. In iris recognition the null
hypothesis is specified by the WD range between 0.0 and 0.2
for the genuine subject. The significance level is another term
related to the degree of certainty that the system requires in
order to reject the null hypothesis in favor of the alternative
hypothesis. By taking a small sample the system cannot be
certain about the conclusion. So decide in advance to reject the
null hypothesis if the probability of observing the sampled
result is less than the significance level. A typical significance
level is 0.21. The p-value is the probability of observing the
given sample result under the assumption that the null
hypothesis is true. If the p-value is greater than the WD range,
then system rejects the null hypothesis. For example, if
WD= 0.2 and the p-value is 0.22, then the system rejects the
null hypothesis. The results of biometric for many hypothesis
tests also include confidence intervals. That is, a confidence
interval is a range of values that have a chosen probability of
containing the true hypothesized quantity. An illustrative
example, WD = 0.03 is inside a 97% confidence interval for
the mean. That is equivalent to being unable to reject the null
hypothesis at a significance level of 0.03. Conversely, if the
100(1 -WD) is confidence interval that does not contain
weighted distance range then the system rejects the null
hypothesis at the level of significance.
E. Receiver operating characteristics curve analysis
The ROC analysis of wavelet features with WD and HNN
is illustrated in Fig 12. The both WD and HNN classifiers
produced approximately the same amount of accuracy.
However, WD produced quite better exactness than the HNN
since it requires minimal error tolerance and threshold in
252
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
training and testing processes. Thus HNN loosely allows
impostors more than WD.
ROC of Wavelet WD
;© J ; "
I f ■■■■■-■ ■•■•-|--'
ROC of Wavelet HNN
1
0.98
B 0.96
h*
V
c
1 ° 92
9
0.88
0.6i3
! ; i
i
; j] [ ■
005 0.1
False Accept Rate
0.O5 0.1
False Accept Rate
Normal Conditions : ' Outdoor Conditions Contact lens and spectacles Diseased iris |
Fig. 12. ROC analysis wavelet features with WD and HNN.
F. Performance comparison
In this work, a feedback design for extraction of rotation
invariant iris recognition based on local segmentation of iris
portions was suggested. It prevents misclassifications (FAR)
of iris patterns and limits the overall FRR of the system. As
per research work, 40% of iris images have been obscured by
eyelids / eyelashes and 35% of images hid the top portions of
iris. This system pulls out left, right and bottom local area of
iris for iris code extraction. It provides overall accuracy of
98.3% in the iris localization process. In [2], elastic
deformation has occurred in the iris portion due to
illumination changes. It was compensated to convert the
circular portion of the iris (including eyelids / eyelashes) into a
rectangle strip, which was used for convolution operations in
iris matching. In the present work these types of discrepancies
have been resolved by local segmentation process. In addition,
the previous method influenced by seven rotation angles (-9, -
6, -3, 0, 3, 6 and 9 degrees). But in our proposed system, the
rotation-invariance was achieved by on combination of Radon
transform and wavelet feature sets. In [1], 2048 feature
components were used to classify the diverse iris patterns. It
readily achieved scale and translation invariant pattern
analysis using integrodifferential operator. However, rotation-
invariant might be carried out by shifting of iris phase codes.
So, it inclined sequences of orientation of templates for the
recognition process. In our approach, we employed with
sequences of rotation estimation preprocessing based on the
Radon transform in order to extract the rotation invariant
features, which in turn, influence a distinctive template for
each subject in enrolled of the system. In [3], Shiny oung Lim
et al. suggested an approach based on Haar wavelet with linear
vector quantization method. This method worked with 87 high
pass filter of the wavelet transformation. However, middle
frequencies of the iris patterns are very useful in the
recognition. In our present work both middle and high
frequencies of wavelet component of iris patterns are used.
Additionally, transformation-invariant is efficiently achieved
prior to the feature extraction. Therefore, multiple iris features
or additional shift operation is completely avoided in the
proposed methodology. Thus, this paper provides better
accuracy with compact rotation invariant feature set than
previous methods.
IV. Conclusion and Future work
This paper processes a feedback design for rotation invariant
feature extraction in application with iris patterns using Radon
and wavelet analysis. After correcting rotation angle, rotation
invariant contours are processed by feature extractor
repeatedly until a suitable set was encountered. It increases
more recognition rate and rotation estimation with diverse
artifacts than the other methods since the previous methods
used redundant patterns of iris feature templates for different
angle of capturing or additional shift operation for
compensating the invariants. Suggested methods would be
possibly implemented with other applications of object
rotation estimation and recognition. This paper opens a new
direction of research in the vision and biometric committees.
Acknowledgment
Authors thank their family members and children for their
continuous support and consent encouragement to do this
research work successfully.
References
[1] Daugman J., 'How Iris Recognition Works', IEEE Transactions On
Circuits and Systems For Video Technology, vol. 14, no. 1, pp. 21-30,
2004.
[2] Li ma, Tieniu Tan, Yunhong Wang and Dexin Zhang, 'Personal
Identification Based on Iris Texture Analysis', IEEE Transactions on
Pattern Analysis and Machine Intelligence, vol. 25, no. 12, pp. 1519-
1533,2003.
[3] Pietikainen M., Ojala T. and Xu Z., 'Rotation- invariant texture
classification using feature distributions', Pattern Recognition, vol. 33,
pp. 43-52, 2000.
[4] Mao J. and Jain A. K., 'Texture classification and segmentation using
multiresolution simultaneous autoregressive models', Pattern
Recognition, vol. 25, no. 4, pp. 173-188,1992.
[5] Aditya Vailaya, Hong Jiang Zhang, Changjiang Yang, Feng-I Liu and
Anil K. Jain, 'Automatic Image Orientation Detection', IEEE
Transactions on Image Processing, vol. 1 1, no. 7, pp. 746-755, 2002.
[6] Chen J. L. and Kundu A. A., 'Rotation and Gray scale transformation
Invariant Texture Identification Using Wavelet Decomposition and
Hidden Markov Model', IEEE Transactions on Pattern Analysis and
Machine Intelligence, vol. 16, no. 2, pp. 208-214, 1994.
[7] Li ma, Tieniu Tan Yunhong Wang and Dexin Zhang, 'Efficient Iris
Recognition by Characterizing key Local variations', IEEE Transaction
on Image Processing, vol. 13, no. 6, pp. 739-750, 2004.
[8] Shinyoung Lim, Kwanyong Lee, Okhwan Byeon and Taiyun Kim,
'Efficient Iris Recognition through Improvement of Feature Vector and
Classifier', ETRI J., vol. 23, nNo. 2, pp. 61-70, 2001.
[9] Lian Cai and Sidan Du, 'Rotation, scale and translation invariant image
watermarking using Radon transform and Fourier transform',
Proceedings of the IEEE 6th Circuit and systems Symposium Emerging
253
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
Technologies: Mobile and Wireless Communication, Shanghai, China,
pp. 281-284, 2004.
[10] Mitra Abhishek and Banerjee S., 'A Regular Algorithm For Real Time
Radon and Inverse Radon Transform', Proceedings of IEEE Acoustics,
Speech and Signal Processing (ICASSP), Montreal, Quebec, Canada, pp.
V.105-V. 108, 2004.
[11] Kourosh Jafari-Kkouzani and Hamid Soltaian-Zadeh, 'Rotation-
Invariant Multiresolution Texture analysis using Radon and Wavelet
Transforms', IEEE Transactions on Image Processing, vol. 14, no. 6, pp.
783-795,2005.
[12] Jun Zhang, Xiyuan Zhou and Erke Mao, 'Image Object Recognition
based on Radon Transform', Proc. of IEEE 5th World Congress on
Intelligent Control and Automation, Hangzhou, China, pp. 4070-4074,
2004.
[13] Haward L. Resnikoff and Raymond O. Wells , 'Wavelet Analysis-The
Scalable Structure of Information', Springer- Verlag, New York (ISBN:
81-8128-226-4), 1998.
[14] James S. Walker, 'A Primer on Wavelets and their Scientific
Applications', CRC Press LLC, USA, 1999.
[15] Phil Picton, 'Introduction to Neural Networks', The Macmillan Press
Ltd., First edition, Great Britain (ISBN:0-333-61832-7), 1994.
[16] Bremananth R., and Chitra A., 'A new approach for iris pattern analysis
based on wavelet and HNN' , Journal of CSI, vol. 36, no. 2, pp. 33-41
(ISSN: 0254-7813), 2006.
[17] Bremananth R., Chitra A., 'Real-Time Image Orientation Detection and
Recognition', International Conference on Signal and Image Processing
(ICSIP), Dec. 2006, pp. 460-461.
[18] Bremananth R., and Chitra A, 'Rotation Invariant Recognition of Iris',
Journal of Systems Science and Engineering, Systems Society of India,
vol.17, no.l, pp.69-78, 2008.
[19] Bremananth R., Ph.D. Dissertation, Anna University, Chennai, India,
2008.
AUTHORS PROFILE
recognition, Analysis of algorithms, Data structure, Computer graphics
and multimedia.
Bremananth R received the B.Sc and M.Sc.
f degrees in Computer Science from Madurai
Kamaraj and Bharathidsan University,
respectively. He obtained M.Phil, degree in
^ Computer Science and Engineering from
Government college of Technology, Bharathiar
University. He received his Ph.D. degree from
^^^^ ^^^^ Department of Computer Science and Engineering,
PSG College of Technology, Anna University, Chennai, India.
Presently, he is working as a Post-doctoral Research Fellow, at Nanyang
Technological University, Singapore. He received the M N Saha
Memorial award for the best application oriented paper in 2006 by
Institute of Electronics and Telecommunication Engineers (IETE). His
fields of research are acoustic imaging, pattern recognition, computer
vision, image processing, biometrics, multimedia and soft computing.
Dr. Bremananth is a member of Indian society of technical education
(ISTE), advanced computing society (ACS), International Association of
Computer Science and Information Technology (IACIT) and IETE.
Mrs. M. Sankari received her B.Sc. and M.Sc.
degrees in Computer science from Bharathidasan
University, respectively. She has completed her
Master of Philosophy degree in Computer science
from Regional Engineering College, Trichy.
Presently, she is a Head of the department of MCA
at MET and pursuing her doctorate degree in
computer science at Avinashilingam University,
She has published various technical papers at IEEE
Her field of research includes Computer vision, Pattern
254
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
Empirical Mode Decomposition Analysis of Heart Rate Variability
C.Santhi.M.E., Assistant Professor,
Electronics and Communication Engineering, Government College of Technology, Coimbatore-64 1 013
N.Kumaravel Ph.D
Professor, Head of the Department,
Electronics and Communication Engineering, Anna University,Chennai-600 025.
Abstract
The analysis of heart rate variability (HRV)
demands specific capabilities not provided by either
parametric or nonparametric spectral estimation methods.
Empirical mode decomposition (EMD) has the possibility of
dealing with nonstationary and nonlinear embedded
phenomena, for a proper assessment of dynamic and
transient changes in amplitude and time scales of HRV
signal. In this work EMD and a non-linear curve fitting
technique are used to study half an hour HRV signal and its
intrinsic mode function obtained from 20 healthy young
control subjects, 20 healthy old control subjects and 20
subjects with long term ST. The intrinsic oscillations are
measured by means of its meanperiod and variance.
Significant meanperiod reduction is observed in the intrinsic
time scales of healthy old control subjects and subjects with
long term ST, which is used to classify the three groups of
HRV signal with high sensitivity and specificity. The
estimated slope using the non-linear curve fitting technique
represents the flexibility of the cardiovascular system. The
main advantage of this method is it does not make any prior
assumption about the HRV signal being analyzed and no
artificial information is introduced into the filtering method.
Index Terms- Empirical Mode Decomposition, Heart Rate
Variability, Intrinsic Mode Functions, RR intervals,
nonlinear curve fitting.
1. Introduction
Over the last 20 years there has been widespread interest in
the study of variations in the beat-to-beat interval of heart
known as heart rate variability (HRV) or RR interval
variations. HRV has been used as a measure of mortality
primarily with patients who have undergone cardiac
surgery. Clinical depression strongly associated with
mortality with such patients may be seen through a decrease
in HRV [1]. HRV is a non invasive measure of autonomic
nervous system balance. Heart rate is influenced by both
sympathetic and parasympathetic (vagal) activities of ANS.
The influence of both branches of the autonomic nervous
system (ANS) is known as sympathovagal balance reflected
in the RR interval changes. A low frequency (LF)
component provides a measure of sympathetic effects on the
heart and generally occurs in a band between 0.04 Hz and
0.15 Hz. A measurement of the influence of the vagus nerve
in modulating the sinoatrial node can be made in the high
frequency band (HF) loosely defined between 0.15 and 0.4
Hz known as respiratory sinus arrhythmia (RSA), and is a
measure of cardiac parasympathetic activity. The ratio of
power in the LF and HF bands (LF/HF) provides the
measure of cardiac sympathovagal balance. Empirical Mode
Decomposition (EMD) retains the intrinsic nonlinear
nonstationary property of the signal. Any intrinsic timescale
derived from the signal is based on the local characteristics
timescale of the signal [2-4]. EMD carries out layer upon
layer sifting and obtains ordered array components from
smallest scale (highest frequency) to largest scale (lowest
frequency) [4]. Empirical mode decomposition has the
possibility of dealing with nonstationary and nonlinear
embedded phenomena, and owing to its suitability for a
proper assessment of the dynamic and transient changes in
amplitude and in frequency of the HRV components [2& 3].
Application of EMD to half an hour HRV data
yields nine intrinsic mode functions (IMFs). The first scale
represents the highest frequency or the shortest period
component of the signal. The second scale represents the
lower frequency or the longer period component of the
signal. Similarly the last IMF represents the lowest time
scale present in the HRV signal. The first two scales contain
more than 85% of total signal power. The meanperiod and
variance of IMFs are computed as time domain measures.
The variance of IMF decreases exponentially with respect to
increasing timescales (meanperiods). Using nonlinear curve
fitting technique the IMFs variations are estimated. The
estimated parameter represents the flexibility of the
cardiovascular system.The methodology is applied to HRV
signal obtained from 20 healthy young control subjects, 20
healthy old control subjects and 20 subjects with long term
ST. The intrinsic time scale of IMF 2 classifies the three
groups HRV signal with high sensitivity and specificity.
2. Empirical Mode Decomposition (EMD)
EMD is a procedure oriented adaptive method for
decomposing non-linear non- stationary signals. The
components resulting from EMD are called Intrinsic Mode
Functions (IMFs) [2]. The IMFs are amplitude frequency
modulated intrinsic signals. The IMF's represents the
oscillatory modes imbedded in the signal. It should satisfies
the following two conditions. 1. In the whole data set the
number of extrema's and the number of zero crossings must
be either equal or differ by at most one. 2. At any point the
mean value of the envelope defined by the local minima and
the envelope defined by the local maxima is zero.
255
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
Fig. 1 .RR interval signal Fig.2. Intrinsic Mode Functions
tot
mM
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
Step 5: Check h(t) for the conditions of an Intrinsic Mode
Functions. [2]
If h(t) is an IMF compute residue r(t)=x(t)-h(t) and again
the process is repeated to extract the next IMF. If h(t) is not
an IMF x(t) is replaced with h(t) and the procedure is
repeated from step 1. Fig.6 shows all IMFs of the signal
x(t).
The process ends when the range of residue is below a
predetermined level or the residue has a monotonic trend. In
order to guarantee that the IMF components retains enough
physical sense in both amplitude and frequency
modulations, the sifting process is stopped by limiting the
size of standard deviation(SD) which is computed from two
consecutive sifting results.
Fig. 3. Reconstructed signal FigADetrended signal
Figs 1-4 explain the efficiency of EMD for RR
interval signal. The ECG data has been collected from the
biomedical website [7] http://www.physionet.org . The RR
intervals are derived from half an hour ECG signal by
identifying the QRS complexes. The signal is manually
edited and only noise free ectopic free segments are used for
the analysis. A real time RR interval signal and its EMD
decomposed IMFs are shown in Fig.l&2. Application of
EMD to real time RR interval signal identifies eight to nine
IMFs. The IMFs are zero mean amplitude frequency
modulated signal. The decomposition is adaptive and
lossless. The original RR interval signal is reconstructed
using decomposed IMFs (Fig. 3). The nonstationary trend is
removed when the residue or monotonic trend (last IMF) is
omitted while reconstructing the signal (Fig.4).
From the RR intervals the HRV signal or ARR
signal (Ri+i-Ri) is obtained by computing successive
difference between consecutive RR intervals. The obtained
HRV signal and its IMFs are shown in Fig. 5 and Fig.6.
Matlab 7.1 tools are used for the analysis.
3. Methodology
SIFTING ALGORITHM:
Fig.5.HRV signal Fig.6. Intrinsic Mode Functions
Step 1 : All the minima and maxima of the HRV Signal x (t),
are located.
Step 2: Spline Interpolate the minima and maxima points to
obtain lower and upper envelopes of the signal.
Step 3: Compute mean envelope
m (t)=(maxima's+minima's)/2.
Step 4: Subtract local mean from the original Signal to
obtain local details h(t)=x(t)- m(t).
sd = £[|V-d(0-M0| 2 /*» Vd(0] (i)
t=0
where k represents number of siftings.
The process of finding an intrinsic mode function
requires number of iterations and the process to find all the
IMFs requires further more iterations. As a result of this
iterative procedure finally yields many IMFs and a residue.
By summing up all the IMF functions and the residue, the
original signal is reconstructed, given by the mathematical
formulae
X(t) = £h,(t)+r(n)
(2)
1=1
Where each hi represents an intrinsic mode function and r(n)
either a mean trend or a constant.
For each IMF the meanperiod and variance are
computed. The meanperiod is the ratio of distance between
the first and last zero-crossings to number of zero-crossings
of IMF.
The obtained RR interval signal using ECG
represents the response of the cardiovascular system to ANS
activities not the ANS activities themselves. The
characteristics of cardiovascular system determine how the
system responds to ANS activity and can alter significantly
the characteristics of the HRV signal. The response
characteristics are often nonlinear in nature. The IMFs
capture the all the variations present in the HRV signal.
Plotting the variance of all IMFs against its meanperiods
gives a nonlinear function. The variance of IMF decreases
with increasing meanperiod and this behavior is
approximated using a geometric function
Y= aX b ~ (3)
where Y represents vector of IMF's variance,. X represents
vector of meanperiods of IMFs, a is constant and b is the
exponential decrease of the function. The IMFs meanperiod
and variance of healthy young control subjects, healthy old
control subjects and long term ST subjects vary
significantly. The variations in the IMF are quantified by
the parameter b. The parameter b represents the flexibility
of cardiovascular system to ANS activities. The parameter b
is estimated using nonlinear curve fitting technique
explained below.
256
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
Taking logarithm of equation (3), gives
lnY = lna + blnX
(4)
putting Y =ln Y , X =ln X , A = In a then the above
nonlinear equation becomes Y = A + bX which is a linear
equation in X. The corresponding normal equations are
2X = wi*+b5X (5 )
_*2
£jry*=A*jx+bjx
(6)
Fig. 7. Curve fitting
N represents number of IMFs. Solving the normal equations
using least mean square method the variables 'a' and 'b' are
estimated [5]. The simulated response function using the
estimated parameter is shown in Fig. 7.
4. Results and Discussion
EMD and curve fitting techniques are applied to half an
hour HRV signal of 20 healthy young control subjects, 20
healthy old control subjects and 20 subjects with long term
ST. Empirical mode decomposition adaptively decomposes
the half an hour HRV signal into number of Intrinsic mode
functions (Fig. 6). The analysis is done with ARR intervals.
ARR (R i+1 -Ri) represents the difference between successive
beat intervals. The IMFs are measured by their absolute
variance, relative variance and meanperiods. The
meanperiod is the ratio of distance between the first and last
zero-crossings to the number of zero-crossings of IMF. First
3 IMFs contains more than 92% of total variance. The IMF1
represents the highest frequency or the shortest period
component of the signal. The IMF2 contains the lower
frequency or the longer period component of the signal.
Since the 1 st and 2 nd IMF contains more than 85% of total
power they are very significant.
Relative powers are computed with respect to total power
considering all IMFs except the residue with zero
meanperiod. In healthy young subjects an increase in
relative power of IMF1 decreases the relative power of
IMF2 (Fig. 8). IMF 1 and IMF 2 are in phase opposition
representing different components of the HRV signal. The
original signal is interpolated to 2 Hz for a meaningful
frequency measure. The Welch periodogram (with window
width 1024 and window overlap of 512 samples) of IMFs of
a healthy young control subject are shown in Fig. (9). Table-
1 gives the peak frequency(Hz) and absolute spectralpower
(ms 2 -miliseconds square) of IMFs The figure shows the
frequency spectrum of the IMFs falls in the recognized
spectral bands of HRV signal: 1 .High frequency band from
0.1 5Hz to 0.5Hz; 2. Low frequency band from 0.04Hz to
0.15Hz; 3. Very low frequency band from 0.01Hz to
0.04Hz..
Relative powers of IMF 1 and IMF 2
Relative power
8
6
5
3
2
nlfl
ititfl ww\
■ Relative powers of IMF 1
■ Relative powers of IMF 2
1 3 5 7 9 11 13 15 17 19
Healthy young control records
Fig 8: Relative powers of IMF 1 and IMF 2
-A
:.A
k
Fig 9: Welch periodogram of IMFs
IMFs
Peak
frequency in
Hz
Peak power in
ms 2
IMF1
0.2891
0.01
IMF2
0.13
0.003
IMF3
0.068
0.002
IMF4
0.03
0.00069
IMF5
0.021
0.0007
IMF6
0.01
0.00062
Table- 1 Spectral values of IMFs
The meanperiod of IMF2 of healthy young controls subjects
are significantly higher compared to healthy old controls
subjects and subjects with long term ST. Considering
meanperiod of IMF2 (2.9724 sees) as threshold value, we
classified the healthy young control subjects and subjects
with long term ST with 95% sensitivity and 90% specificity.
The classification is shown in Fig. (10). A threshold value
of 2.8 sees classifies the healthy old controls subjects and
subjects with long term ST with 90% sensitivity and 70%
specificity shown in Fig .(11).
IMF 2 meanperiod of healthy young and subjects with long term
St
3.5
3
2.5
2
1.5
1
0.5
* v/A *S*~*^^/\ ^*-^ /
\ r\v ^^ : /O^/
\J ^r-m-mr-^ ^ ^ ~W \J/
—♦—Healthy young
—■—Sub. With long term st
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20
Fig. 10. Meanperiod comparison of healthy young subjects
and subjects with longterm ST.
257
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
IMF 2 meanperiod of healthy old and sub.with long term st
-Healthy old controls
-sub.with long term st
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20
Fig.l 1. Meanperiod comparison of healthy old subjects and
subjects with longterm ST.
The parameter b of IMFs of the three groups HRV signal
are estimated, average plots are shown in Fig. (12). The
estimated parameter b of healthy young control subjects,
healthy old control subjects and long term ST subjects are -
1.49, -1.43 and -1.39 (average values only). The more
negative value represents the flexibility of the system. The
healthy young control subject's cardiovascular system is
more flexible than healthy old subjects and longterm ST
subjects. The flexibility of the system decreases in healthy
old control subjects and longterm ST subjects. The absolute
powers of healthy young control subjects are significantly
higher compared to healthy old subjects and long term ST
subjects as shown in Fig. (13) (average values only). The
higher values of absolute power represent more fluctuating
power in the signal. The results show the HRV of healthy
young control subjects contains higher power, longer time
scales and more adaptive to ANS activities compared to
healthy old control subjects and subjects with long term ST.
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
rate variability analysis", Med.Bio.Eng.Comput., 2001, 39,
471-479.
[3]E.P.Souza Neto, M.A.Custaud, J.C.Cejka,P.Abry,
J.Frutoso, C.Gharib, P.Flandrin, "Assessment of
Cardiovascular Autonomic Control by the Empirical Mode
Decomposition", Methods Inf Med 2004;43:60-5.
[4] N.E.Huang, Z.Shen, S.R.Long, M.C.Wu, H.H.Shih,
etal.1998. "The empirical mode decomposition and the
Hilbert spectrum for nonlinear and nonstationary time series
analysis" Proc.R.Soc.A, vol 454, pp.903-995.
[5] B.V.Ramana, "Higher Engineering Mathematics", Tata
McGraw-Hill Publishing Company Limited,New Delhi.
[6] HRV Analysis Software 1.1, developed by The
Biomedical Signal Analysis Group, Department of Applied
Physics, University of Kuopio, Finland.
http://venda.uku.fi/research/biosignal
[7] www.physionet.org.
[8] Jan W.Kantelhard, Stephan A, Armin Bunde, 2002,
Multifractal Detrended Fluctuation Analysis of
Nonstationary Time Series, Physica A 316, 87-1 14.
Absolute powerslMFlfi IMF2
0.004
Absolute 1
powers 1
0.9)1- '
-lOp[k
tlZoHMR
'
Young 2. Old 3. Long termST
Fig. 12 Correlation graphs Fig. 13. Absolute powers
5. Conclusion
In order to cope up nonlinearity and nonstationarity
issue of HRV signal EMD and nonlinear curve fitting
techniques are used in this work. The IMFs of HRV signal
are negatively correlated. The frequency spectrum of first
two IMFs falls in the recognized HF and LF spectral bands
of HRV signal. The meanperiod of IMF2 classifies half an
hour HRV signal of healthy young control subjects, healthy
old control subjects and subjects with long term ST with
high sensitivity and specificity. The nonlinear curve fitting
technique estimates the flexibility of cardiovascular system.
The method is simple, adaptive and no artificial information
is introduced in the analysis.
6. References
[1] R. M. Carney, J. A. Blumenthal, P. K. Stein, L.
Watkins, D. Catellier, L. F. Berkman, S. M. Czajkowski, C.
O'Connor, P. H. Stone, K. E.Freedland, "Depression, Heart
Rate Variability, and Acute Myocardial Infarction,"
Circulation, vol. 104, no. 17, pp. 2024 - 2028, 2001.
[2] J.C.Echeverria, J.A.Crowe, M.S.Woolfson, B.R.Hayes-
Gill, "Application of empirical mode decomposition to heart
258
http://sites.google.com/site/ijcsis/
ISSN 1947-5500
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
IJCSIS REVIEWERS' LIST
Assist Prof (Dr.) M. Emre Celebi, Louisiana State University in Shreveport, USA
Dr. Lam Hong Lee, Universiti Tunku Abdul Rahman, Malaysia
Dr. Shimon K. Modi, Director of Research BSPA Labs, Purdue University, USA
Dr. J ianguo Ding, Norwegian University of Science and Technology (NTNU), Norway
Assoc. Prof. N. Jaisankar, VIT University, Vellore,Tamilnadu, India
Dr. Amogh Kavimandan, The Mathworks Inc., USA
Dr. Ramasamy Mariappan, Vinayaka Missions University, India
Dr. Yong Li, School of Electronic and Information Engineering, Beijing J iaotong University, P.R. China
Assist. Prof. Sugam Sharma, NIET, India / Iowa State University, USA
Dr. Jorge A. Ruiz-Vanoye, Universidad Autonoma del Estado de Morelos, Mexico
Dr. Neeraj Kumar, SMVD University, Katra (J&K), India
DrGenge Bela, "Petru Maior" University of Targu Mures, Romania
Dr. J unjie Peng, Shanghai University, P. R. China
Dr. Ilhem LENGLIZ, HANA Group - CRISTAL Laboratory, Tunisia
Prof. Dr. Durgesh Kumar Mishra, Acropolis Institute of Technology and Research, Indore, MP, India
Jorge L. Hernandez-Ardieta, University Carlos III of Madrid, Spain
Prof. Dr.C.Suresh Gnana Dhas, Anna University, India
Mrs Li Fang, Nanyang Technological University, Singapore
Prof. Pijush Biswas, RCC Institute of Information Technology, India
Dr. Siddhivinayak Kulkarni, University of Ballarat, Ballarat, Victoria, Australia
Dr. A. Ami Lawrence, Royal College of Engineering & Technology, India
Mr. Wongyos Keardsri, Chulalongkorn University, Bangkok, Thailand
Mr. Somesh Kumar Dewangan, CSVTU Bhilai (C.G.)/ Dimat Raipur, India
Mr. Hayder N. Jasem, University Putra Malaysia, Malaysia
Mr. A.V.Senthil Kumar, C. M. S. College of Science and Commerce, India
Mr. R. S. Karthik, C. M. S. College of Science and Commerce, India
Mr. P. Vasant, University Technology Petronas, Malaysia
Mr. Wong Kok Seng, Soongsil University, Seoul, South Korea
Mr. Praveen Ranjan Srivastava, BITS PI LAN I, India
Mr. Kong Sang Kelvin, Leong, The Hong Kong Polytechnic University, Hong Kong
Mr. Mohd Nazri Ismail, Universiti Kuala Lumpur, Malaysia
Dr. Rami J. Matarneh, Al-isra Private University, Amman, Jordan
Dr Ojesanmi Olusegun Ayodeji, Ajayi Crowther University, Oyo, Nigeria
Dr. Riktesh Srivastava, Skyline University, UAE
Dr. Oras F. Baker, UCSI University - Kuala Lumpur, Malaysia
Dr. Ahmed S. Ghiduk, Faculty of Science, Beni-Suef University, Egypt
and Department of Computer science, Taif University, Saudi Arabia
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
Mr. Tirthankar Gayen, NT Kharagpur, India
Ms. Huei-Ru Tseng, National Chiao Tung University, Taiwan
Prof. Ning Xu, Wuhan University of Technology, China
Mr Mohammed Salem Binwahlan, Hadhramout University of Science and Technology, Yemen
& Universiti Teknologi Malaysia, Malaysia.
Dr. Aruna Ranganath, Bhoj Reddy Engineering College for Women, India
Mr. Hafeezullah Amin, Institute of Information Technology, KUST, Kohat, Pakistan
Prof. Syed S. Rizvi, University of Bridgeport, USA
Mr. Shahbaz Pervez Chattha, University of Engineering and Technology Taxila, Pakistan
Dr. Shishir Kumar, Jaypee University of Information Technology, Wakanaghat (HP), India
Mr. Shahid Mumtaz, Portugal Telecommunication, Instituto de Telecomunicagoes (IT) , Aveiro, Portugal
Mr. Rajesh K Shukla, Corporate Institute of Science & Technology Bhopal M P
Dr. Poonam Garg, Institute of Management Technology, India
Mr. S. Mehta, Inha University, Korea
Mr. Dilip Kumar S.M, University Visvesvaraya College of Engineering (UVCE), Bangalore University,
Bangalore
Prof. Malik Sikander Hayat Khiyal, Fatima J innah Women University, Rawalpindi, Pakistan
Dr. Virendra Gomase , Department of Bioinformatics, Padmashree Dr. D.Y. Patil University
Dr. Irraivan Elamvazuthi, University Technology PETRONAS, Malaysia
Mr. Saqib Saeed, University of Siegen, Germany
Mr. Pavan Kumar Gorakavi, IPMA-USA [YC]
Dr. Ahmed Nabih Zaki Rashed, Menoufia University, Egypt
Prof. Shishir K. Shandilya, Rukmani Devi I nstitute of Science & Technology, I ndia
Mrs.J .Komala Lakshmi, SNR Sons College, Computer Science, I ndia
Mr. Muhammad Sohail, KUST, Pakistan
Dr. Manjaiah D.H, Mangalore University, India
Dr. S Santhosh Baboo, D.G.Vaishnav College, Chennai, India
Prof. Dr. Mokhtar Beldjehem, Sainte-Anne University, Halifax, NS, Canada
Dr. Deepak Laxmi Narasimha, Faculty of Computer Science and Information Technology, University of
Malaya, Malaysia
Prof. Dr. ArunkumarThangavelu, Vellore Institute Of Technology, India
Mr. M. Azath, Anna University, India
Mr. Md. Rabiul Islam, Rajshahi University of Engineering & Technology (RUET), Bangladesh
Mr. Aos Alaa Zaidan Ansaef, Multimedia University, Malaysia
Dr Suresh Jain, Professor (on leave), Institute of Engineering & Technology, Devi Ahilya University, Indore
(MP) India,
Dr. Mohammed M. Kadhum, Universiti Utara Malaysia
Mr. Hanumanthappa. J . University of Mysore, I ndia
Mr. Syed Ishtiaque Ahmed, Bangladesh University of Engineering and Technology (BUET)
MrAkinola Solomon Olalekan, University of I badan, Ibadan, Nigeria
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
Mr. Santosh K. Pandey, Department of Information Technology, The Institute of Chartered Accountants of
India
Dr. P. Vasant, Power Control Optimization, Malaysia
Dr. Petr Ivankov, Automatika - S, Russian Federation
Dr. Utkarsh Seetha, Data Infosys Limited, India
Mrs. Priti Maheshwary, Maulana Azad National Institute of Technology, Bhopal
Dr. (Mrs) Padmavathi Ganapathi, Avinashilingam University for Women, Coimbatore
Assist. Prof. A. Neela madheswari, Anna university, India
Prof. Ganesan Ramachandra Rao, PSG College of Arts and Science, India
Mr. Kamanashis Biswas, Daffodil International University, Bangladesh
Dr. Atul Gonsai, Saurashtra University, Gujarat, India
Mr. Angkoon Phinyomark, Prince of Songkla University, Thailand
Mrs. G. Nalini Priya, Anna University, Chennai
Dr. P. Subashini, Avinashilingam University for Women, India
Assoc. Prof. Vijay Kumar Chakka, Dhirubhai Ambani IICT, Gandhinagar , Gujarat
Mrjitendra Agrawal, : Rajiv Gandhi Proudyogiki Vishwavidyalaya, Bhopal
Mr. Vishal Goyal, Department of Computer Science, Punjabi University, India
Dr. R. Baskaran, Department of Computer Science and Engineering, Anna University, Chennai
Assist. Prof, Kanwalvir Singh Dhindsa, B.B.S.B.Engg. College, Fatehgarh Sahib (Punjab), India
Dr. Jamal Ahmad Dargham, School of Engineering and Information Technology, Universiti Malaysia Sabah
Mr. Nitin Bhatia, DAV College, India
Dr. Dhavachelvan Ponnurangam, Pondicherry Central University, India
Dr. Mohd Faizal Abdollah, University of Technical Malaysia, Malaysia
Assist. Prof. Sonal Chawla, Panjab University, India
Dr. Abdul Wahid, AKG Engg. College, Ghaziabad, India
Mr. Arash Habibi Lashkari, University of Malaya (UM), Malaysia
Mr. Md. Rajibul Islam, Ibnu Sina Institute, University Technology Malaysia
Professor Dr. Sabu M. Thampi, .B.S Institute of Technology for Women, Kerala University, India
Mr. Noor Muhammed Nayeem, Universite Lumiere Lyon 2, 69007 Lyon, France
Dr. Himanshu Aggarwal, Department of Computer Engineering, Punjabi University, India
Prof R. Naidoo, Dept of Mathematics/Center for Advanced Computer Modelling, Durban University of
Technology, Durban, South Africa
Prof. Mydhili K Nair, M S Ramaiah Institute of Technology(M. S.R.I. T), Affliliated to Visweswaraiah
Technological University, Bangalore, India
M. Prabu, Adhiyamaan College of Engineering/Anna University, India
Mr. Swakkhar Shatabda, Department of Computer Science and Engineering, United International University,
Bangladesh
Dr. Abdur Rashid Khan, ICIT, Gomal University, Dera Ismail Khan, Pakistan
Mr. H. Abdul Shabeer, l-NautixTechnologies,Chennai, India
Dr. M. Aramudhan, Perunthalaivar Kamarajar Institute of Engineering and Technology, India
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
Dr. M. P. Thapliyal, Department of Computer Science, HNB Garhwal University (Central University), I ndia
Prof Ekta Walia Bhullar, Maharishi Markandeshwar University, Mullana (Ambala), India
Dr. Shahaboddin Shamshirband, Islamic Azad University, Iran
Mr. Zeashan Hameed Khan, : Universite de Grenoble, France
Prof. Anil K Ahlawat, Ajay Kumar Garg Engineering College, Ghaziabad, UP Technical University, Lucknow
Mr. Longe Olumide Babatope, University Of Ibadan, Nigeria
Associate Prof. Raman Maini, University College of Engineering, Punjabi University, India
Dr. Maslin Masrom, University Technology Malaysia, Malaysia
Sudipta Chattopadhyay, Jadavpur University, Kolkata, India
Dr. Dang Tuan NGUYEN, University of I nformation Technology, Vietnam National University - Ho Chi Minh
City
Dr. Mary Lourde R., BITS-PI LANI Dubai , UAE
Dr. Abdul Aziz, University of Central Punjab, Pakistan
Mr. Karan Singh, Gautam Budtha University, India
Mr. Avinash Pokhriyal, Uttar Pradesh Technical University, Lucknow, India
Associate Prof Dr Zuraini Ismail, University Technology Malaysia, Malaysia
Assistant Prof. Yasser M. Alginahi, College of Computer Science and Engineering, Taibah University, Madinah
Munawwarrah, KSA
Mr. Dakshina Ranjan Kisku, West Bengal University of Technology, India
Mr. Raman Kumar, Dr B R Ambedkar National I nstitute of Technology, J alandhar, Punjab, I ndia
Associate Prof. Samir B. Patel, Institute of Technology, Nirma University, India
Dr. M.Munir Ahamed Rabbani, B. S. Abdur Rahman University, India
Asst. Prof. Koushik Majumder, West Bengal University of Technology, India
Dr. Alex Pappachen James, Queensland Micro-nanotechnology center, Griffith University, Australia
Assistant Prof. S. Hariharan, B.S. Abdur Rahman University, India
Asst Prof. Jasmine. K. S, R.V. College of Engineering, India
Mr Naushad Ali Mamode Khan, Ministry of Education and Human Resources, Mauritius
Prof. Mahesh Goyani, G H Patel Collge of Engg. &Tech, V.V.N, Anand, Gujarat, India
Dr. Mana Mohammed, University of Tlemcen, Algeria
Prof. J atinder Singh, Universal I nstitutiion of Engg. & Tech. CHD, I ndia
Mrs. M. Anandhavalli Gauthaman, Sikkim Manipal I nstitute of Technology, Majitar, East Sikkim
Dr. Bin Guo, Institute Telecom SudParis, France
Mrs. Maleika Mehr Nigar Mohamed Heenaye-Mamode Khan, University of Mauritius
Prof. Pijush Biswas, RCC Institute of Information Technology, India
Mr. V. Bala Dhandayuthapani, Mekelle University, Ethiopia
Mr. Irfan Syamsuddin, State Polytechnic of Ujung Pandang, Indonesia
Mr. Kavi Kumar Khedo, University of Mauritius, Mauritius
Mr. Ravi Chandiran, Zagro Singapore Pte Ltd. Singapore
Mr. Milindkumar V. Sarode, J awaharlal Darda I nstitute of Engineering and Technology, I ndia
Dr. Shamimul Qamar, KSJ I nstitute of Engineering & Technology, I ndia
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
Dr. C. Arun, Anna University, India
Assist. Prof. M.N.Birje, Basaveshwar Engineering College, India
Prof. Hamid Reza Naji, Department of Computer Enigneering, Shahid Beheshti University, Tehran, Iran
Assist. Prof. Debasis Giri, Department of Computer Science and Engineering, Haldia Institute of Technology
Subhabrata Barman, Haldia Institute of Technology, West Bengal
Mr. M. I. Lali, COMSATS Institute of Information Technology, Islamabad, Pakistan
Dr. Feroz Khan, Central Institute of Medicinal and Aromatic Plants, Lucknow, India
Mr. R. Nagendran, Institute of Technology, Coimbatore, Tamilnadu, India
Mr. Amnach Khawne, King Mongkut's Institute of Technology Ladkrabang, Ladkrabang, Bangkok, Thailand
Dr. P. Chakrabarti, Sir Padampat Singhania University, Udaipur, India
Mr. Nafiz Imtiaz Bin Hamid, Islamic University of Technology (I UT), Bangladesh.
Shahab-A. Shamshirband, Islamic Azad University, Chalous, Iran
Prof. B. Priestly Shan, Anna Univeristy, Tamilnadu, India
Venkatramreddy Velma, Dept. of Bioinformatics, University of Mississippi Medical Center, Jackson MS USA
Akshi Kumar, Dept. of Computer Engineering, Delhi Technological University, India
Dr. Umesh Kumar Singh, Vikram University, Ujjain, India
Mr. Serguei A. Mokhov, Concordia University, Canada
Mr. Lai Khin Wee, Universiti Teknologi Malaysia, Malaysia
Dr. Awadhesh Kumar Sharma, Madan Mohan Malviya Engineering College, India
Mr. Syed R. Rizvi, Analytical Services & Materials, Inc., USA
Dr. S. Karthik, SNS Collegeof Technology, India
Mr. Syed Qasim Bukhari, CI MET (Universidad de Granada), Spain
Mr. A.D.Potgantwar, Pune University, India
Dr. Himanshu Aggarwal, Punjabi University, India
Mr. Rajesh Ramachandran, Naipunya Institute of Management and Information Technology, India
Dr. K.L. Shunmuganathan, R.M.K Engg College , Kavaraipettai ,Chennai
Dr. Prasant Kumar Pattnaik, KIST, India.
Dr. Ch. Aswani Kumar, VIT University, India
Mr. Ijaz Ali Shoukat, King Saud University, Riyadh KSA
Mr. Arun Kumar, Sir Padam Pat Singhania University, Udaipur, Rajasthan
Mr. Muhammad Imran Khan, Universiti Teknologi PETRONAS, Malaysia
Dr. Natarajan Meghanathan, Jackson State University, Jackson, MS, USA
Mr. Mohd Zaki Bin Mas'ud, Universiti Teknikal Malaysia Melaka (UTeM), Malaysia
Prof. Dr. R. Geetharamani, Dept. of Computer Science and Eng., Rajalakshmi Engineering College, India
Dr. Smita Rajpal, Institute of Technology and Management, Gurgaon, India
Dr. S. Abdul Khader J ilani, University of Tabuk, Tabuk, Saudi Arabia
Mr. Syedjamal Haider Zaidi, Bahria University, Pakistan
Dr. N. Devarajan, Government College of Technology,Coimbatore, Tamilnadu, INDIA
Mr. R. Jagadeesh Kannan, RMK Engineering College, India
Mr. Deo Prakash, Shri Mata Vaishno Devi University, India
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
Mr. Mohammad Abu Naser, Dept. of EEE, IUT, Gazipur, Bangladesh
Assist. Prof. Prasun Ghosal, Bengal Engineering and Science University, India
Mr. Md. Golam Kaosar, School of Engineering and Science, Victoria University, Melbourne City, Australia
Mr. R. Mahammad Shafi, Madanapalle I nstitute of Technology & Science, I ndia
Dr. F.Sagayaraj Francis, Pondicherry Engineering College,! ndia
Dr. Ajay Goel, HI ET , Kaithal, I ndia
Mr. NayakSunil Kashibarao, Bahirji Smarak Mahavidyalaya, India
Mr. SuhasJ Manangi, Microsoft India
Dr. Kalyankar N. V., Yeshwant Mahavidyalaya, Nanded , India
Dr. K.D. Verma, S.V. College of Post graduate studies & Research, India
Dr. Amjad Rehman, University Technology Malaysia, Malaysia
Mr. Rachit Garg, L K College, J alandhar, Punjab
Mr. J . William, M.A.M college of Engineering, Trichy, Tamilnadujndia
Prof. J ue-Sam Chou, Nanhua University, College of Science and Technology, Taiwan
Dr. Thorat S.B., I nstitute of Technology and Management, I ndia
Mr. Ajay Prasad, Sir Padampat Singhania University, Udaipur, India
Dr. Kamaljit I. Lakhtaria, Atmiya Institute of Technology & Science, India
Mr. Syed Rafiul Hussain, Ahsanullah University of Science and Technology, Bangladesh
Mrs Fazeela Tunnisa, Najran University, Kingdom of Saudi Arabia
Mrs Kavita Taneja, Maharishi Markandeshwar University, Haryana, India
Mr. Maniyar Shiraz Ahmed, Najran University, Najran, KSA
Mr. Anand Kumar, AMC Engineering College, Bangalore
Dr. Rakesh Chandra Gangwar, Beant College of Engg. & Tech., Gurdaspur (Punjab) I ndia
Dr. VV Rama Prasad, Sree Vidyanikethan Engineering College, India
Assist. Prof. Neetesh Kumar Gupta, Technocrats Institute of Technology, Bhopal (M.P.), India
Mr. Ashish Seth, Uttar Pradesh Technical University, Lucknow ,UP India
Dr. VV S S S Balaram, Sreenidhi I nstitute of Science and Technology, I ndia
Mr Rahul Bhatia, Lingaya's I nstitute of Management and Technology, I ndia
Prof. Niranjan Reddy. P, KITS , Warangal, India
Prof. Rakesh. Lingappa, Vijetha Institute of Technology, Bangalore, India
Dr. Mohammed Ali Hussain, Nimra College of Engineering & Technology, Vijayawada, A. P., I ndia
Dr. A.Srinivasan, MNM Jain Engineering College, Rajiv Gandhi Salai, Thorapakkam, Chennai
Mr. Rakesh Kumar, M.M. University, Mullana, Ambala, India
Dr. Lena Khaled, Zarqa Private University, Aman, Jordon
Ms. Supriya Kapoor, Patni/Lingaya's Institute of Management and Tech., India
Dr. Tossapon Boongoen , Aberystwyth University, UK
Dr . Bilal Alatas, Firat University, Turkey
Assist. Prof. Jyoti Praaksh Singh , Academy of Technology, India
Dr. Ritu Soni, GNG College, India
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
Dr . Mahendra Kumar , Sagar I nstitute of Research & Technology, Bhopal, I ndia.
Dr. Binod Kumar, India
Dr. Muzhir Shaban Al-Ani, Amman Arab University Amman - J ordan
Dr. T.C. Manjunath , ATRIA Institute of Tech, India
Mr. Muhammad Zakarya, COMSATS Institute of Information Technology (CUT), Pakistan
Assist. Prof. Harmunish Taneja, M. M. University, India
Dr. Chitra Dhawale , SICSR, Model Colony, Pune, India
Mrs Sankari Muthukaruppan, Nehru Institute of Engineering and Technology, Anna University, India
Mr. Aaqif Afzaal Abbasi, National University Of Sciences And Technology, Islamabad
Prof. Ashutosh Kumar Dubey, Trinity Institute of Technology and Research Bhopal, India
Mr. G. Appasami, Dr. Pauls Engineering College, India
Mr. M Yasin, National University of Science and Tech, karachi (NUST), Pakistan
Mr. Yaser Miaji, University Utara Malaysia, Malaysia
Mr. Shah Ahsanul Haque, International Islamic University Chittagong (IIUC), Bangladesh
Prof. (Dr) Syed Abdul Sattar, Royal Institute of Technology & Science, India
Dr. S. Sasikumar, Roever Engineering College
Assist. Prof. Monit Kapoor, Maharishi Markandeshwar University, India
Mr. Nwaocha Vivian 0, National Open University of Nigeria
Dr. M. S. Vijaya, GR Govindarajulu School of Applied Computer Technology, India
Assist. Prof. Chakresh Kumar, Manav Rachna International University, India
Mr. Kunal Chadha , R&D Software Engineer, Gemalto, Singapore
Mr. Pawan J indal, J aypee University of Engineering and Technology, I ndia
Mr. Mueen Uddin, Universiti Teknologi Malaysia, UTM , Malaysia
Dr. Dhuha Basheer abdullah, Mosul university, Iraq
Mr. S. Audithan, Annamalai University, India
Prof. Vijay K Chaudhari, Technocrats Institute of Technology , India
Associate Prof. Mohd llyas Khan, Technocrats Institute of Technology , India
Dr. Vu Thanh Nguyen, University of Information Technology, HoChiMinh City, VietNam
Assist. Prof. Anand Sharma, MITS, Lakshmangarh, Sikar, Rajasthan, India
Prof. T V Narayana Rao, HITAM Engineering college, Hyderabad
Mr. Deepak Gour, Sir Padampat Singhania University, India
Assist. Prof. Amutharaj Joyson, Kalasalingam University, India
Mr. Ali Balador, Islamic Azad University, Iran
Mr. Mohitjain, Maharaja Surajmal Institute of Technology, India
Mr. Dilip Kumar Sharma, GLA I nstitute of Technology & Management, I ndia
Dr. Debojyoti Mitra, Sir padampat Singhania University, India
Dr. Ali Dehghantanha, Asia-Pacific University College of Technology and I nnovation, Malaysia
Mr. Zhao Zhang, City University of Hong Kong, China
Prof. S.P. Setty, A.U. College of Engineering, India
Prof. Patel Rakeshkumar Kantilal, Sankalchand Patel College of Engineering, India
(IJCSIS) International Journal of Computer Science and Information Security,
Vol. 8, No. 6, September 2010
Mr. Biswajit Bhowmik, Bengal College of Engineering & Technology, India
Mr. Manoj Gupta, Apex Institute of Engineering & Technology, India
Assist. Prof. Ajay Sharma, Raj Kumar Goel Institute Of Technology, India
Assist. Prof. Ramveer Singh, Raj Kumar Goel Institute of Technology, India
Dr. Hanan Elazhary, Electronics Research Institute, Egypt
Dr. Hosam I. Faiq, USM, Malaysia
Prof. Dipti D. Patil, MAEER's MIT College of Engg. & Tech, Pune, I ndia
Assist. Prof. Devendra Chack, BCT Kumaon engineering College Dwarahat Almora, India
Prof. Manpreet Singh, M. M. Engg. College, M. M. University, India
Assist. Prof. M. Sadiq ali Khan, University of Karachi, Pakistan
Mr. Prasad S. Halgaonkar, MIT - College of Engineering, Pune, India.
CALL FOR PAPERS
International Journal of Computer Science and Information Security
IJCSIS 2010
ISSN: 1947-5500
http://sites.google.com/site/ijcsis/
International Journal Computer Science and Information Security, now at its sixth edition, is the premier
scholarly venue in the areas of computer science and security issues. IJCSIS 2010 will provide a high
profile, leading edge platform for researchers and engineers alike to publish state-of-the-art research in the
respective fields of information technology and communication security. The journal will feature a diverse
mixture of publication articles including core and applied computer science related topics.
Authors are solicited to contribute to the special issue by submitting articles that illustrate research results,
projects, surveying works and industrial experiences that describe significant advances in the following
areas, but are not limited to. Submissions may span a broad range of topics, e.g.:
Track A: Security
Access control, Anonymity, Audit and audit reduction & Authentication and authorization, Applied
cryptography, Cryptanalysis, Digital Signatures, Biometric security, Boundary control devices,
Certification and accreditation, Cross-layer design for security, Security & Network Management, Data and
system integrity, Database security, Defensive information warfare, Denial of service protection, Intrusion
Detection, Anti-malware, Distributed systems security, Electronic commerce, E-mail security, Spam,
Phishing, E-mail fraud, Virus, worms, Trojan Protection, Grid security, Information hiding and
watermarking & Information survivability, Insider threat protection, Integrity
Intellectual property protection, Internet/Intranet Security, Key management and key recovery, Language-
based security, Mobile and wireless security, Mobile, Ad Hoc and Sensor Network Security, Monitoring
and surveillance, Multimedia security , Operating system security, Peer-to-peer security, Performance
Evaluations of Protocols & Security Application, Privacy and data protection, Product evaluation criteria
and compliance, Risk evaluation and security certification, Risk/vulnerability assessment, Security &
Network Management, Security Models & protocols, Security threats & countermeasures (DDoS, MiM,
Session Hijacking, Replay attack etc,), Trusted computing, Ubiquitous Computing Security, Virtualization
security, VoIP security, Web 2.0 security, Submission Procedures, Active Defense Systems, Adaptive
Defense Systems, Benchmark, Analysis and Evaluation of Security Systems, Distributed Access Control
and Trust Management, Distributed Attack Systems and Mechanisms, Distributed Intrusion
Detection/Prevention Systems, Denial-of- Service Attacks and Countermeasures, High Performance
Security Systems, Identity Management and Authentication, Implementation, Deployment and
Management of Security Systems, Intelligent Defense Systems, Internet and Network Forensics, Large-
scale Attacks and Defense, RFID Security and Privacy, Security Architectures in Distributed Network
Systems, Security for Critical Infrastructures, Security for P2P systems and Grid Systems, Security in E-
Commerce, Security and Privacy in Wireless Networks, Secure Mobile Agents and Mobile Code, Security
Protocols, Security Simulation and Tools, Security Theory and Tools, Standards and Assurance Methods,
Trusted Computing, Viruses, Worms, and Other Malicious Code, World Wide Web Security, Novel and
emerging secure architecture, Study of attack strategies, attack modeling, Case studies and analysis of
actual attacks, Continuity of Operations during an attack, Key management, Trust management, Intrusion
detection techniques, Intrusion response, alarm management, and correlation analysis, Study of tradeoffs
between security and system performance, Intrusion tolerance systems, Secure protocols, Security in
wireless networks (e.g. mesh networks, sensor networks, etc.), Cryptography and Secure Communications,
Computer Forensics, Recovery and Healing, Security Visualization, Formal Methods in Security, Principles
for Designing a Secure Computing System, Autonomic Security, Internet Security, Security in Health Care
Systems, Security Solutions Using Reconfigurable Computing, Adaptive and Intelligent Defense Systems,
Authentication and Access control, Denial of service attacks and countermeasures, Identity, Route and
Location Anonymity schemes, Intrusion detection and prevention techniques, Cryptography, encryption
algorithms and Key management schemes, Secure routing schemes, Secure neighbor discovery and
localization, Trust establishment and maintenance, Confidentiality and data integrity, Security architectures,
deployments and solutions, Emerging threats to cloud-based services, Security model for new services,
Cloud-aware web service security, Information hiding in Cloud Computing, Securing distributed data
storage in cloud, Security, privacy and trust in mobile computing systems and applications, Middleware
security & Security features: middleware software is an asset on
its own and has to be protected, interaction between security-specific and other middleware features, e.g.,
context-awareness, Middleware-level security monitoring and measurement: metrics and mechanisms
for quantification and evaluation of security enforced by the middleware, Security co-design: trade-off and
co-design between application-based and middleware-based security, Policy-based management:
innovative support for policy-based definition and enforcement of security concerns, Identification and
authentication mechanisms: Means to capture application specific constraints in defining and enforcing
access control rules, Middleware-oriented security patterns: identification of patterns for sound, reusable
security, Security in aspect-based middleware: mechanisms for isolating and enforcing security aspects,
Security in agent-based platforms: protection for mobile code and platforms, Smart Devices: Biometrics,
National ID cards, Embedded Systems Security and TPMs, RFID Systems Security, Smart Card Security,
Pervasive Systems: Digital Rights Management (DRM) in pervasive environments, Intrusion Detection and
Information Filtering, Localization Systems Security (Tracking of People and Goods), Mobile Commerce
Security, Privacy Enhancing Technologies, Security Protocols (for Identification and Authentication,
Confidentiality and Privacy, and Integrity), Ubiquitous Networks: Ad Hoc Networks Security, Delay-
Tolerant Network Security, Domestic Network Security, Peer-to-Peer Networks Security, Security Issues
in Mobile and Ubiquitous Networks, Security of GSM/GPRS/UMTS Systems, Sensor Networks Security,
Vehicular Network Security, Wireless Communication Security: Bluetooth, NFC, WiFi, WiMAX,
WiMedia, others
This Track will emphasize the design, implementation, management and applications of computer
communications, networks and services. Topics of mostly theoretical nature are also welcome, provided
there is clear practical potential in applying the results of such work.
Track B: Computer Science
Broadband wireless technologies: LTE, WiMAX, WiRAN, HSDPA, HSUPA, Resource allocation and
interference management, Quality of service and scheduling methods, Capacity planning and dimensioning,
Cross-layer design and Physical layer based issue, Interworking architecture and interoperability, Relay
assisted and cooperative communications, Location and provisioning and mobility management, Call
admission and flow/congestion control, Performance optimization, Channel capacity modeling and analysis,
Middleware Issues: Event-based, publish/subscribe, and message-oriented middleware, Reconfigurable,
adaptable, and reflective middleware approaches, Middleware solutions for reliability, fault tolerance, and
quality-of-service, Scalability of middleware, Context-aware middleware, Autonomic and self-managing
middleware, Evaluation techniques for middleware solutions, Formal methods and tools for designing,
verifying, and evaluating, middleware, Software engineering techniques for middleware, Service oriented
middleware, Agent-based middleware, Security middleware, Network Applications: Network-based
automation, Cloud applications, Ubiquitous and pervasive applications, Collaborative applications, RFID
and sensor network applications, Mobile applications, Smart home applications, Infrastructure monitoring
and control applications, Remote health monitoring, GPS and location-based applications, Networked
vehicles applications, Alert applications, Embeded Computer System, Advanced Control Systems, and
Intelligent Control : Advanced control and measurement, computer and microprocessor-based control,
signal processing, estimation and identification techniques, application specific IC's, nonlinear and
adaptive control, optimal and robot control, intelligent control, evolutionary computing, and intelligent
systems, instrumentation subject to critical conditions, automotive, marine and aero-space control and all
other control applications, Intelligent Control System, Wiring/Wireless Sensor, Signal Control System.
Sensors, Actuators and Systems Integration : Intelligent sensors and actuators, multisensor fusion, sensor
array and multi-channel processing, micro/nano technology, microsensors and microactuators,
instrumentation electronics, MEMS and system integration, wireless sensor, Network Sensor, Hybrid
Sensor, Distributed Sensor Networks. Signal and Image Processing : Digital signal processing theory,
methods, DSP implementation, speech processing, image and multidimensional signal processing, Image
analysis and processing, Image and Multimedia applications, Real-time multimedia signal processing,
Computer vision, Emerging signal processing areas, Remote Sensing, Signal processing in education.
Industrial Informatics: Industrial applications of neural networks, fuzzy algorithms, Neuro-Fuzzy
application, biolnformatics, real-time computer control, real-time information systems, human-machine
interfaces, CAD/CAM/CAT/CIM, virtual reality, industrial communications, flexible manufacturing
systems, industrial automated process, Data Storage Management, Harddisk control, Supply Chain
Management, Logistics applications, Power plant automation, Drives automation. Information Technology,
Management of Information System : Management information systems, Information Management,
Nursing information management, Information System, Information Technology and their application, Data
retrieval, Data Base Management, Decision analysis methods, Information processing, Operations research,
E-Business, E-Commerce, E-Government, Computer Business, Security and risk management, Medical
imaging, Biotechnology, Bio-Medicine, Computer-based information systems in health care, Changing
Access to Patient Information, Healthcare Management Information Technology.
Communication/Computer Network, Transportation Application : On-board diagnostics, Active safety
systems, Communication systems, Wireless technology, Communication application, Navigation and
Guidance, Vision-based applications, Speech interface, Sensor fusion, Networking theory and technologies,
Transportation information, Autonomous vehicle, Vehicle application of affective computing, Advance
Computing technology and their application : Broadband and intelligent networks, Data Mining, Data
fusion, Computational intelligence, Information and data security, Information indexing and retrieval,
Information processing, Information systems and applications, Internet applications and performances,
Knowledge based systems, Knowledge management, Software Engineering, Decision making, Mobile
networks and services, Network management and services, Neural Network, Fuzzy logics, Neuro-Fuzzy,
Expert approaches, Innovation Technology and Management : Innovation and product development,
Emerging advances in business and its applications, Creativity in Internet management and retailing, B2B
and B2C management, Electronic transceiver device for Retail Marketing Industries, Facilities planning
and management, Innovative pervasive computing applications, Programming paradigms for pervasive
systems, Software evolution and maintenance in pervasive systems, Middleware services and agent
technologies, Adaptive, autonomic and context-aware computing, Mobile/Wireless computing systems and
services in pervasive computing, Energy-efficient and green pervasive computing, Communication
architectures for pervasive computing, Ad hoc networks for pervasive communications, Pervasive
opportunistic communications and applications, Enabling technologies for pervasive systems (e.g., wireless
BAN, PAN), Positioning and tracking technologies, Sensors and RFID in pervasive systems, Multimodal
sensing and context for pervasive applications, Pervasive sensing, perception and semantic interpretation,
Smart devices and intelligent environments, Trust, security and privacy issues in pervasive systems, User
interfaces and interaction models, Virtual immersive communications, Wearable computers, Standards and
interfaces for pervasive computing environments, Social and economic models for pervasive systems,
Active and Programmable Networks, Ad Hoc & Sensor Network, Congestion and/or Flow Control, Content
Distribution, Grid Networking, High-speed Network Architectures, Internet Services and Applications,
Optical Networks, Mobile and Wireless Networks, Network Modeling and Simulation, Multicast,
Multimedia Communications, Network Control and Management, Network Protocols, Network
Performance, Network Measurement, Peer to Peer and Overlay Networks, Quality of Service and Quality
of Experience, Ubiquitous Networks, Crosscutting Themes - Internet Technologies, Infrastructure,
Services and Applications; Open Source Tools, Open Models and Architectures; Security, Privacy and
Trust; Navigation Systems, Location Based Services; Social Networks and Online Communities; ICT
Convergence, Digital Economy and Digital Divide, Neural Networks, Pattern Recognition, Computer
Vision, Advanced Computing Architectures and New Programming Models, Visualization and Virtual
Reality as Applied to Computational Science, Computer Architecture and Embedded Systems, Technology
in Education, Theoretical Computer Science, Computing Ethics, Computing Practices & Applications
Authors are invited to submit papers through e-mail ij csiseditorfo) gmail. com . Submissions must be original
and should not have been published previously or be under consideration for publication while being
evaluated by IJCSIS. Before submission authors should carefully read over the journal's Author Guidelines,
which are located at http://sites.google.com/site/ijcsis/authors-notes .
© IJCSIS PUBLICATION 2010
ISSN 1947 5500